Introduction to Spark


Details
This talk is for data engineers, analysts, software engineers, and technical managers interested in brief overview of Apache Spark. Spark is a cluster computing framework in analyzing large scale data, with data parallelism and fault tolerance. As Spark achieves better results for Machine learning algorithms, Data scientists are interested to perform data analysis at scale using Spark. Here, we will discuss core APIs for using Spark, basic elements of the framework, like Spark core, RDDs, Dataframes, Datasets, Spark SQL.
Sandeep is an engineer over a decade long experience with various programming languages and databases. He is certified Spark Developer. Currently working with Comcast Cable as a Principal Data Engineer. Dinner will be provided by Datalere Consulting. Thank you, Datalere!
Agenda:
6:00-6:15 Introductions
6:15-7:15 Presentation, Sandeep Kumar, Principal Data Engineer
7:15-7:30 Questions & Networking
When you arrive at Galvanize, use the elevators to go up to the 3rd level. We are typically located in one of the classrooms. From there, look for signs for 'Denver All Things Data'.

Introduction to Spark