Have a question?
Message sent Close
4.68 out of 5
4.68
11 reviews on Udemy

Spark SQL & Hadoop (For Data Scientists & Big Data Analysts)

Learn HDFS commands, Hadoop, Spark SQL, SQL Queries, ETL & Data Analysis| Spark Hadoop Cluster VM | Fully Solved Qs
Instructor:
Matthew Barr
3,271 students enrolled
English [Auto]
Students will get hands-on experience working in a Spark Hadoop environment that’s free and downloadable as part of this course.
Students will have opportunities solve Data Engineering and Data Analysis Problems using Spark on a Hadoop cluster in the sandbox environment that comes as part
Issuing HDFS commands.
Converting a set of data values in a given format stored in HDFS into new data values or a new data format and writing them into HDFS.
Loading data from HDFS for use in Spark applications & writing the results back into HDFS using Spark.
Reading and writing files in a variety of file formats.
Performing standard extract, transform, load (ETL) processes on data using the Spark API.
Using metastore tables as an input source or an output sink for Spark applications.
Applying the understanding of the fundamentals of querying datasets in Spark.
Filtering data using Spark.
Writing queries that calculate aggregate statistics.
Joining disparate datasets using Spark.
Producing ranked or sorted data.

Apache Spark is currently one of the most popular systems for processing big data.

 

Apache Hadoop continues to be used by many organizations that look to store data locally on premises. Hadoop allows these organisations to efficiently store big datasets ranging in size from gigabytes to petabytes.

 

As the number of vacancies for data science, big data analysis and data engineering roles continue to grow, so too will the demand for individuals that possess knowledge of Spark and Hadoop technologies to fill these vacancies.

 

This course has been designed specifically for data scientists, big data analysts and data engineers looking to leverage the power of Hadoop and Apache Spark to make sense of big data.

 

This course will help those individuals that are looking to interactively analyse big data or to begin writing production applications to prepare data for further analysis using Spark SQL in a Hadoop environment.

 

The course is also well suited for university students and recent graduates that are keen to gain exposure to Spark & Hadoop or anyone who simply wants to apply their SQL skills in a big data environment using Spark-SQL.

 

This course has been designed to be concise and to provide students with a necessary and sufficient amount of theory, enough for them to be able to use Hadoop & Spark without getting bogged down in too much theory about older low-level APIs such as RDDs.

 

On solving the questions contained in this course students will begin to develop those skills & the confidence needed to handle real world scenarios that come their way in a production environment.

 

(a) There are just under 30 problems in this course. These cover hdfs commands, basic data engineering tasks and data analysis.

(b) Fully worked out solutions to all the problems.

(c) Also included is the Verulam Blue virtual machine which is an environment that has a spark Hadoop cluster already installed so that you can practice working on the problems.

 

  • The VM contains a Spark Hadoop environment which allows students to read and write data to & from the Hadoop file system as well as to store metastore tables on the Hive metastore.
  • All the datasets students will need for the problems are already loaded onto HDFS, so there is no need for students to do any extra work.
  • The VM also has Apache Zeppelin installed. This is a notebook specific to Spark and is similar to Python’s Jupyter notebook.

     

This course will allow students to get hands-on experience working in a Spark Hadoop environment as they practice:

 

  • Converting a set of data values in a given format stored in HDFS into new data values or a new data format and writing them into HDFS.
  • Loading data from HDFS for use in Spark applications & writing the results back into HDFS using Spark.
  • Reading and writing files in a variety of file formats.
  • Performing standard extract, transform, load (ETL) processes on data using the Spark API.
  • Using metastore tables as an input source or an output sink for Spark applications.
  • Applying the understanding of the fundamentals of querying datasets in Spark.
  • Filtering data using Spark.
  • Writing queries that calculate aggregate statistics.
  • Joining disparate datasets using Spark.
  • Producing ranked or sorted data.
You can view and review the lecture materials indefinitely, like an on-demand channel.
Definitely! If you have an internet connection, courses on Udemy are available on any device at any time. If you don't have an internet connection, some instructors also let their students download course lectures. That's up to the instructor though, so make sure you get on their good side!
4.7
4.7 out of 5
11 Ratings

Detailed Rating

Stars 5
9
Stars 4
0
Stars 3
1
Stars 2
0
Stars 1
0
a517a47374c799160104f7fc87d9908a
30-Day Money-Back Guarantee

Includes

6 hours on-demand video
6 articles
Full lifetime access
Access on mobile and TV
Certificate of Completion

External Links May Contain Affiliate Links read more

Join our Telegram Channel To Get Latest Notification & Course Updates!
Join Our Telegram For FREE Courses & Canva PremiumJOIN NOW