Apache Spark is a General-purpose computing engine that has in-memory computing capabilities. It can be used for a variety of workloads like Batch processing, Iterative problems, stream processing, etc. It is designed to be highly scalable and provides various APIs like Scala, Python, R, Java, and SQL. It can be easily integrated with other BIG Data tools as well.
In this workshop, we will cover the below topics.
• RDD: Resilient Distributed Dataset
• Q & A
Please note the Time Zone for this event.