Apache Spark Training Courses

Apache Spark Training Courses

Local, instructor-led live Apache Spark training courses demonstrate through hands-on practice how Spark fits into the Big Data ecosystem, and how to use Spark for data analysis. Apache Spark training is available as "onsite live training" or "remote live training". Onsite live training can be carried out locally on customer premises in Argentina or in NobleProg corporate training centers in Argentina. Remote live training is carried out by way of an interactive, remote desktop. NobleProg -- Your Local Training Provider

Testimonials

★★★★★
★★★★★

Apache Spark Subcategories

Apache Spark Course Outlines

CodeNameDurationOverview
sparkdevSpark for Developers21 hoursOBJECTIVE:

This course will introduce Apache Spark. The students will learn how Spark fits into the Big Data ecosystem, and how to use Spark for data analysis. The course covers Spark shell for interactive data analysis, Spark internals, Spark APIs, Spark SQL, Spark streaming, and machine learning and graphX.

AUDIENCE :

Developers / Data Analysts
hdpHortonworks Data Platform (HDP) for Administrators21 hoursHortonworks Data Platform is an open-source Apache Hadoop support platform that provides a stable foundation for developing big data solutions on the Apache Hadoop ecosystem.

This instructor-led live training introduces Hortonworks and walks participants through the deployment of Spark + Hadoop solution.

By the end of this training, participants will be able to:

- Use Hortonworks to reliably run Hadoop at a large scale
- Unify Hadoop's security, governance, and operations capabilities with Spark's agile analytic workflows.
- Use Hortonworks to investigate, validate, certify and support each of the components in a Spark project
- Process different types of data, including structured, unstructured, in-motion, and at-rest.

Audience

- Hadoop administrators

Format of the course

- Part lecture, part discussion, exercises and heavy hands-on practice
magellanMagellan: Geospatial Analytics on Spark14 hoursMagellan is an open-source distributed execution engine for geospatial analytics on big data. Implemented on top of Apache Spark, it extends Spark SQL and provides a relational abstraction for geospatial analytics.

This instructor-led, live training introduces the concepts and approaches for implementing geospacial analytics and walks participants through the creation of a predictive analysis application using Magellan on Spark.

By the end of this training, participants will be able to:

- Efficiently query, parse and join geospatial datasets at scale
- Implement geospatial data in business intelligence and predictive analytics applications
- Use spatial context to extend the capabilities of mobile devices, sensors, logs, and wearables

Audience

- Application developers

Format of the course

- Part lecture, part discussion, exercises and heavy hands-on practice
alluxioAlluxio: Unifying Disparate Storage Systems7 hoursAlluxio is an open-source virtual distributed storage system that unifies disparate storage systems and enables applications to interact with data at memory speed. It is used by companies such as Intel, Baidu and Alibaba.

In this instructor-led, live training, participants will learn how to use Alluxio to bridge different computation frameworks with storage systems and efficiently manage multi-petabyte scale data as they step through the creation of an application with Alluxio.

By the end of this training, participants will be able to:

- Develop an application with Alluxio
- Connect big data systems and applications while preserving one namespace
- Efficiently extract value from big data in any storage format
- Improve workload performance
- Deploy and manage Alluxio standalone or clustered

Audience

- Data scientist
- Developer
- System administrator

Format of the course

- Part lecture, part discussion, exercises and heavy hands-on practice
sparkpythonPython and Spark for Big Data (PySpark)21 hoursPython is a high-level programming language famous for its clear syntax and code readibility. Spark is a data processing engine used in querying, analyzing, and transforming big data. PySpark allows users to interface Spark with Python.

In this instructor-led, live training, participants will learn how to use Python and Spark together to analyze big data as they work on hands-on exercises.

By the end of this training, participants will be able to:

- Learn how to use Spark with Python to analyze Big Data
- Work on exercises that mimic real world circumstances
- Use different tools and techniques for big data analysis using PySpark

Audience

- Developers
- IT Professionals
- Data Scientists

Format of the course

- Part lecture, part discussion, exercises and heavy hands-on practice
sparksqlApache Spark SQL7 hoursSpark SQL is Apache Spark's module for working with structured and unstructured data. Spark SQL provides information about the structure of the data as well as the computation being performed. This information can be used to perform optimizations. Two common uses for Spark SQL are:
- to execute SQL queries.
- to read data from an existing Hive installation.

In this instructor-led, live training (onsite or remote), participants will learn how to analyze various types of data sets using Spark SQL.

By the end of this training, participants will be able to:

- Install and configure Spark SQL.
- Perform data analysis using Spark SQL.
- Query data sets in different formats.
- Visualize data and query results.

Audience

- Data analysts
- Data scientists
- Data engineers

Format of the Course

- Part lecture, part discussion, exercises and heavy hands-on practice

Notes

- To request a customized training for this course, please contact us to arrange.
introtostreamprocessingA Practical Introduction to Stream Processing21 hoursStream Processing refers to the real-time processing of "data in motion", that is, performing computations on data as it is being received. Such data is read as continuous streams from data sources such as sensor events, website user activity, financial trades, credit card swipes, click streams, etc. Stream Processing frameworks are able to read large volumes of incoming data and provide valuable insights almost instantaneously.

In this instructor-led, live training (onsite or remote), participants will learn how to set up and integrate different Stream Processing frameworks with existing big data storage systems and related software applications and microservices.

By the end of this training, participants will be able to:

- Install and configure different Stream Processing frameworks, such as Spark Streaming and Kafka Streaming
- Understand and select the most appropriate framework for the job
- Process of data continuously, concurrently, and in a record-by-record fashion
- Integrate Stream Processing solutions with existing databases, data warehouses, data lakes, etc.
- Integrating the most appropriate stream processing library with enterprise applications and microservices

Audience

- Developers
- Software architects

Format of the Course

- Part lecture, part discussion, exercises and heavy hands-on practice

Notes

- To request a customized training for this course, please contact us to arrange.
bigdataanahealthBig Data Analytics in Health21 hoursBig data analytics involves the process of examining large amounts of varied data sets in order to uncover correlations, hidden patterns, and other useful insights.

The health industry has massive amounts of complex heterogeneous medical and clinical data. Applying big data analytics on health data presents huge potential in deriving insights for improving delivery of healthcare. However, the enormity of these datasets poses great challenges in analyses and practical applications to a clinical environment.

In this instructor-led, live training (remote), participants will learn how to perform big data analytics in health as they step through a series of hands-on live-lab exercises.

By the end of this training, participants will be able to:

- Install and configure big data analytics tools such as Hadoop MapReduce and Spark
- Understand the characteristics of medical data
- Apply big data techniques to deal with medical data
- Study big data systems and algorithms in the context of health applications

Audience

- Developers
- Data Scientists

Format of the Course

- Part lecture, part discussion, exercises and heavy hands-on practice.

Note

- To request a customized training for this course, please contact us to arrange.
spmllibApache Spark MLlib35 hoursMLlib is Spark’s machine learning (ML) library. Its goal is to make practical machine learning scalable and easy. It consists of common learning algorithms and utilities, including classification, regression, clustering, collaborative filtering, dimensionality reduction, as well as lower-level optimization primitives and higher-level pipeline APIs.

It divides into two packages:

-

spark.mllib contains the original API built on top of RDDs.

-

spark.ml provides higher-level API built on top of DataFrames for constructing ML pipelines.

Audience

This course is directed at engineers and developers seeking to utilize a built in Machine Library for Apache Spark
aitechArtificial Intelligence - the most applied stuff - Data Analysis + Distributed AI + NLP21 hoursThis course is aimed at developers and data scientists who wish to understand and implement AI within their applications. Special focus is given to Data Analysis, Distributed AI and NLP.
graphcomputingIntroduction to Graph Computing28 hoursMany real world problems can be described in terms of graphs. For example, the Web graph, the social network graph, the train network graph and the language graph. These graphs tend to be extremely large; processing them requires a specialized set of tools and processes -- these tools and processes can be referred to as Graph Computing (also known as Graph Analytics).

In this instructor-led, live training, participants will learn about the technology offerings and implementation approaches for processing graph data. The aim is to identify real-world objects, their characteristics and relationships, then model these relationships and process them as data using a graph computing approach. We start with a broad overview and narrow in on specific tools as we step through a series of case studies, hands-on exercises and live deployments.

By the end of this training, participants will be able to:

- Understand how graph data is persisted and traversed
- Select the best framework for a given task (from graph databases to batch processing frameworks)
- Implement Hadoop, Spark, GraphX and Pregel to carry out graph computing across many machines in parallel
- View real-world big data problems in terms of graphs, processes and traversals

Audience

- Developers

Format of the course

- Part lecture, part discussion, exercises and heavy hands-on practice

Upcoming Apache Spark Courses

CourseCourse DateCourse Price [Remote / Classroom]
Spark for Developers - Buenos Aires - Laminar CatalinasWed, 2018-11-07 09:305,259USD / 6,344USD
Weekend Apache Spark courses, Evening Apache Spark training, Apache Spark boot camp, Apache Spark instructor-led, Weekend Apache Spark training, Evening Apache Spark courses, Apache Spark coaching, Apache Spark instructor, Apache Spark trainer, Apache Spark training courses, Apache Spark classes, Apache Spark on-site, Apache Spark private courses, Apache Spark one on one training

Course Discounts

Course Venue Course Date Course Price [Remote / Classroom]
Neural Network in R Buenos Aires - Laminar Catalinas Thu, 2018-10-25 09:30 3,293USD / 4,367USD
Forecasting with R Buenos Aires - Laminar Catalinas Mon, 2018-11-12 09:30 2,743USD / 3,817USD
Business Plan building with Business Motivation Model Buenos Aires - Laminar Catalinas Tue, 2019-01-29 09:30 2,743USD / 3,817USD
One Day Workshop for PEAP Authentication of Windows 7 Supplicant using a Cisco Switch as Authenticator and Windows 2008 R2 Server Buenos Aires - Laminar Catalinas Wed, 2019-02-06 09:30 1,398USD / 2,461USD
Robotics in business - AI/Robotics Buenos Aires - Laminar Catalinas Wed, 2019-03-06 09:30 2,743USD / 3,817USD

Course Discounts Newsletter

We respect the privacy of your email address. We will not pass on or sell your address to others.
You can always change your preferences or unsubscribe completely.

Some of our clients

is growing fast!

We are looking to expand our presence in your region!

As a Business Development Manager you will:

  • expand business in the region
  • recruit local talent (sales, agents, trainers, consultants)
  • recruit local trainers and consultants

We offer:

  • Artificial Intelligence and Big Data systems to support your local operation
  • high-tech automation
  • continuously upgraded course catalogue and content
  • good fun in international team

If you are interested in running a high-tech, high-quality training and consulting business.

contact us right away!