IBM - Scalable Machine Learning on Big Data using Apache Spark
- Offered byCoursera
Scalable Machine Learning on Big Data using Apache Spark at Coursera Overview
Duration | 7 hours |
Start from | Start Now |
Total fee | Free |
Mode of learning | Online |
Difficulty level | Intermediate |
Official Website | Explore Free Course |
Credential | Certificate |
Scalable Machine Learning on Big Data using Apache Spark at Coursera Highlights
- Earn a shareable certificate upon completion.
- Flexible deadlines according to your schedule.
Scalable Machine Learning on Big Data using Apache Spark at Coursera Course details
- This course will empower you with the skills to scale data science and machine learning (ML) tasks on Big Data sets using Apache Spark. Most real world machine learning work involves very large data sets that go beyond the CPU, memory and storage limitations of a single computer.
- Apache Spark is an open source framework that leverages cluster computing and distributed storage to process extremely large data sets in an efficient and cost effective manner. Therefore an applied knowledge of working with Apache Spark is a great asset and potential differentiator for a Machine Learning engineer.
- After completing this course, you will be able to:
- - gain a practical understanding of Apache Spark, and apply it to solve machine learning problems involving both small and big data
- - understand how parallel code is written, capable of running on thousands of CPUs.
- - make use of large scale compute clusters to apply machine learning algorithms on Petabytes of data using Apache SparkML Pipelines.
- - eliminate out-of-memory errors generated by traditional machine learning frameworks when data doesn?t fit in a computer's main memory
- - test thousands of different ML models in parallel to find the best performing one ? a technique used by many successful Kagglers
- - (Optional) run SQL statements on very large data sets using Apache SparkSQL and the Apache Spark DataFrame API.
- Enrol now to learn the machine learning techniques for working with Big Data that have been successfully applied by companies like Alibaba, Apple, Amazon, Baidu, eBay, IBM, NASA, Samsung, SAP, TripAdvisor, Yahoo!, Zalando and many others.
- NOTE: You will practice running machine learning tasks hands-on on an Apache Spark cluster provided by IBM at no charge during the course which you can continue to use afterwards.
- Prerequisites:
- - basic python programming
- - basic machine learning (optional introduction videos are provided in this course as well)
- - basic SQL skills for optional content
- The following courses are recommended before taking this class (unless you already have the skills)
- https://www.coursera.org/learn/python-for-applied-data-science or similar
- https://www.coursera.org/learn/machine-learning-with-python or similar
- https://www.coursera.org/learn/sql-data-science for optional lectures
Scalable Machine Learning on Big Data using Apache Spark at Coursera Curriculum
Week 1: Introduction
Introduction to Apache Spark for Machine Learning on BigData
What is Big Data?
Data storage solutions
Parallel data processing strategies of Apache Spark
Functional programming basics
Resilient Distributed Dataset and DataFrames - ApacheSparkSQL
Course Syllabus
Setup of the grading and exercise environment
Exercise 1 - working with RDD
Exercise 2 - functional programming basics with RDDs
Exercise 3 - working with DataFrames
Programming Lanuage Options for Apache Spark (optional)
Practice Quiz (Ungraded) - Apache Spark concepts
Apache Spark and parallel data processing
Week 2: Scaling Math for Statistics on Apache Spark
Averages
Standard deviation
Skewness
Kurtosis
Covariance, Covariance matrices, correlation
Plotting with ApacheSpark and python's matplotlib
Dimensionality reduction
PCA
Exercise 1 - statistics and transfomrations using DataFrames
Exercise on Plotting
Exercise on PCA
Practice Quiz (Ungraded) - Statistics and API usage on Spark
Parallelism in Apache Spark
Questions on Plotting
Questions on PCA
Week 3: Introduction to Apache SparkML
How ML Pipelines work
Introduction to SparkML
Extract - Transform - Load
Introduction to Clustering: k-Means
Using K-Means in Apache SparkML
Exercise 1: Modifying a Apache SparkML Feature Engineering Pipeline
Exercise 2 - Working with Clustering and Apache SparkML
Practice Quiz (Ungraded) - ML Pipelines
SparkML concepts
Practice Quiz (Ungraded) - SparkML Algorithms
Week 4: Supervised and Unsupervised learning with SparkML
Linear Regression
LinearRegression with Apache SparkML
Logistic Regression
LogisticRegression with Apache SparkML
Exercise 1 - Improving Classification performance
Course Project
Practice Quiz (Ungraded) - SparkML Algorithms (2)
Course Project Quiz