Taming Big Data with Apache Spark 3 and Python - Hands On!
Ditulis pada: September 13, 2021
Taming Big Data with Apache Spark 3 and Python - Hands On!, Dive right in with 15+ hands-on examples of analyzing large data sets with Apache Spark, on your desktop or on Hadoop!
- Created by Sundog Education by Frank Kane, Frank Kane
- English
- English, French [Auto-generated], 3 more
PREVIEW THIS COURSE - GET COUPON CODE
What you'll learn
- Use DataFrames and Structured Streaming in Spark 3
- Frame big data analysis problems as Spark problems
- Use Amazon's Elastic MapReduce service to run your job on a cluster with Hadoop YARN
- Install and run Apache Spark on a desktop computer or on a cluster
- Use Spark's Resilient Distributed Datasets to process and analyze large data sets across many CPU's
- Implement iterative algorithms such as breadth-first-search using Spark
- Use the MLLib machine learning library to answer common data mining questions
- Understand how Spark SQL lets you work with structured data
- Understand how Spark Streaming lets your process continuous streams of data in real time
- Tune and troubleshoot large jobs running on a cluster
- Share information between nodes on a Spark cluster using broadcast variables and accumulators
- Understand how the GraphX library helps with network analysis problems
Learn how to create state of the art neural networks for deep learning with Facebook's PyTorch Deep Learning library!
Build with modern libraries like Tensorflow, Theano, Keras, PyTorch, CNTK, MXNet. Train faster with GPU on AWS