Learn the fundamentals of Spark. Gain hands-on experience through online labs using Hadoop, Mesos, and more.
Apache Spark equips individuals to make informed data-driven decisions. It is used in data-intensive industries such as retail, healthcare, financial services, and manufacturing. It is important to know Apache Spark if you are keen on pursuing a career in the field of data science.
This course will teach you the fundamentals of Spark and how to leverage its universe of hands-on tools. During this course, you will explore the fundamentals of Spark and become familiar with various core Spark tools. You will discover why and when Spark is used. You will explore the components of the Spark unified stack. You will learn the fundamentals of Spark's principal data abstraction, the Resilient Distributed Dataset. You will learn how to download and install Spark standalone. Plus, you will be introduced to Scala and Python.
This course comprises well-designed modules that take you on a carefully defined learning journey.
This self-paced course does not run to a fixed schedule with regard to completing modules or submitting assignments. However, it is anticipated that if you work 2-3 hours per week, you will complete the course within 2-3 weeks. You can work at your own pace as long as the course is completed before the deadline.
The materials for every module are accessible from the start of the course and will remain available for the duration of your enrollment.
As part of our mentoring service, you will have access to guidance and support throughout the course. We provide a dedicated discussion space where you can ask questions, chat with your peers, and resolve issues.
Once you have successfully completed the course, you will get your IBM Certificate.
After completing this course, you will be able to:
You should have a basic understanding of:
0 Comments