Big Data Crash Course
This course is the perfect blend of exploring the expedition of various Big Data Technologies such as Hadoop, Spark, Nifi, Apache Kafka, etc, and hands-on experience using Google Cloud Platform.
Overview
This course includes:
- 9+ hours of on-demand video
- 12 modules
- Intermediate level
- Direct access/chat with the instructor
- 100% self-paced online
- Many downloadable resources
- Shareable certificate of completion
Skills You Will Gain
Learning Outcomes (At The End Of This Program, You Will Be Able To...)
- Develop the Big Data Pipelines
- Understand various Big Data Technologies such as Hadoop, Apache Spark, Apache Nifi, Apache Kafka, Sqoop, Hive, Impala, HBase, and many more
- Handle a large amount of data easily
- Develop a Big Data Framework VM Instance in Google Cloud Platform using DataProc
- Understand Key Architectures of Big Data
- Gain a holistic understanding of the Big Data Ecosystem
- Work with various File Formats within Big Data Frameworks such as Avro, JSON, Parquet, and many more
- Create a Real-Time Data Analytics Pipeline for fetching the data from Twitter and performing analysis within Apache Spark using Apache NiFi.
Prerequisites
- Basics of SQL and RDBMS
- Unix/Linux Basic Commands like mkdir, ls, cat etc.
- Python/Java (not used extensively in the course)
- Credit card for setting up GCP account (no charges will be deducted if using GCP trial version). You can perform all exercises of this course without incurring charges. Please refer “GCP Account Best Practices” Section for more details.
- Twitter Account
Who Should Attend
- Engineers who are aiming to get a job in Big Data
- Engineers who would like to transition their roles into Big Data Technologies
- Big Data Engineers planning to appear for Certifications like CCA175, CCA159
- Big Data Engineers who are looking for Promotion