
Online or onsite, instructor-led live Big Data training courses start with an introduction to elemental concepts of Big Data, then progress into the programmeming languages and methodologies used to perform Data Analysis. Tools and infrastructure for enabling Big Data storage, Distributed Processing, and Scalability are discussed, compared and implemented in demo practice sessions.
Big Data training is available as "online live training" or "onsite live training". Online live training (aka "remote live training") is carried out by way of an interactive, remote desktop. Onsite live Big Data training can be carried out locally on customer premises in the UK or in NobleProg corporate training centres in the UK.
NobleProg -- Your Local Training Provider
Testimonials
I like the exercises done.
Nour Assaf
Course: Data Mining and Analysis
The hands-on exercise and the trainer capacity to explain complex topics in simple terms.
youssef chamoun
Course: Data Mining and Analysis
The information given was interesting and the best part was towards the end when we were provided with Data from Durex and worked on Data we are familiar with and perform operations to get results.
Jessica Chaar
Course: Data Mining and Analysis
We know a lot more about the whole environment.
John Kidd
Course: Spark for Developers
Richard's training style kept it interesting, the real world examples used helped to drive the concepts home.
Jamie Martin-Royle - NBrown Group
Course: From Data to Decision with Big Data and Predictive Analytics
The content, as I found it very interesting and think it would help me in my final year at University.
Krishan Mistry - NBrown Group
Course: From Data to Decision with Big Data and Predictive Analytics
Richard was very willing to digress when we wanted to ask semi-related questions about things not on the syllabus. Explanations were clear and he was up front about caveats in any advice he gave us.
ARM Limited
Course: Spark for Developers
The trainer was very concern about individual understanding.
Muhammad Surajo Sanusi - Birmingham City University
Course: Foundation R
Excellent presentation and it gives me confidence to build on knowledge gained.
Birmingham City University
Course: Foundation R
Background knowledge and 'provenance' of trainer.
Francis McGonigal - Birmingham City University
Course: Foundation R
Resources
Hafiz Rana - Birmingham City University
Course: Foundation R
Good explanations on how we do things
Birmingham City University
Course: Foundation R
I feel more confident with coding now. I've never done it before but now I understand that it's not rocket science and I can do it when necessary.
Anna Yartseva - Birmingham City University
Course: Foundation R
The fact that all the data and software was ready to use on an already prepared VM, provided by the trainer in external disks.
vyzVoice
Course: Hadoop for Developers and Administrators
The trainer was so knowledgeable and included areas I was interested in.
Mohamed Salama
Course: Data Mining & Machine Learning with R
Very tailored to needs.
Yashan Wang
Course: Data Mining with R
Richard is very calm and methodical, with an analytic insight - exactly the qualities needed to present this sort of course.
Kieran Mac Kenna
Course: Spark for Developers
I mostly liked the trainer giving real live Examples.
Simon Hahn
Course: Administrator Training for Apache Hadoop
I genuinely enjoyed the big competences of Trainer.
Grzegorz Gorski
Course: Administrator Training for Apache Hadoop
I genuinely enjoyed the many hands-on sessions.
Jacek Pieczątka
Course: Administrator Training for Apache Hadoop
I thought that the information was interesting.
Allison May
Course: Data Visualization
I really appreciated that Jeff utilised data and examples that were applicable to education data. He made it interesting and interactive.
Carol Wells Bazzichi
Course: Data Visualization
Learning about all the chart types and what they are used for. Learning the value of cluttering. Learning about the methods to show time data.
Susan Williams
Course: Data Visualization
Trainer was enthusiastic.
Diane Lucas
Course: Data Visualization
I really liked the content / Instructor.
Craig Roberson
Course: Data Visualization
I am a hands-on learner and this was something that he did a lot of.
Lisa Comfort
Course: Data Visualization
I liked the examples.
Peter Coleman
Course: Data Visualization
I liked the examples.
Peter Coleman
Course: Data Visualization
I enjoyed the good real world examples, reviews of existing reports.
Ronald Parrish
Course: Data Visualization
I really was benefit from the willingness of the trainer to share more.
Balaram Chandra Paul
Course: A practical introduction to Data Analysis and Big Data
The trainer made the class interesting and entertaining which helps quite a bit with all day training.
Ryan Speelman
Course: Spark for Developers
I think the trainer had an excellent style of combining humor and real life stories to make the subjects at hand very approachable. I would highly recommend this professor in the future.
Course: Spark for Developers
Liked very much the interactive way of learning.
Luigi Loiacono
Course: Data Analysis with Hive/HiveQL
It was a very practical training, I liked the hands-on exercises.
Proximus
Course: Data Analysis with Hive/HiveQL
I was benefit from the good overview, good balance between theory and exercises.
Proximus
Course: Data Analysis with Hive/HiveQL
I enjoyed the dynamic interaction and “hands-on” the subject, thanks to the Virtual Machine, very stimulating!.
Philippe Job
Course: Data Analysis with Hive/HiveQL
Ernesto did a great job explaining the high level concepts of using Spark and its various modules.
Michael Nemerouf
Course: Spark for Developers
I think the trainer had an excellent style of combining humor and real life stories to make the subjects at hand very approachable. I would highly recommend this professor in the future.
Course: Spark for Developers
Big Data Course Outlines
This instructor-led, live courses covers the working principles behind Accumulo and walks participants through the development of a sample application on Apache Accumulo.
Format of the Course
- Part lecture, part discussion, hands-on development and implementation, occasional tests to gauge understanding
In this instructor-led, live course, we introduce the processes involved in KDD and carry out a series of exercises to practice the implementation of those processes.
Audience
- Data analysts or anyone interested in learning how to interpret data to solve problems
Format of the Course
- After a theoretical discussion of KDD, the instructor will present real-life cases which call for the application of KDD to solve a problem. Participants will prepare, select and cleanse sample data sets and use their prior knowledge about the data to propose solutions based on the results of their observations.
By the end of this training, participants will be able to:
- Install and configure Apache Spark.
- Understand how .NET implements Spark APIs so that they can be accessed from a .NET application.
- Develop data processing applications using C# or F#, capable of handling data sets whose size is measured in terabytes and pedabytes.
- Develop machine learning features for a .NET application using Apache Spark capabilities.
- Carry out exploratory analysis using SQL queries on big data sets.
In this instructor-led, live training, participants will learn how to use MonetDB and how to get the most value out of it.
By the end of this training, participants will be able to:
- Understand MonetDB and its features
- Install and get started with MonetDB
- Explore and perform different functions and tasks in MonetDB
- Accelerate the delivery of their project by maximizing MonetDB capabilities
Audience
- Developers
- Technical experts
Format of the course
- Part lecture, part discussion, exercises and heavy hands-on practice
By the end of this training, participants will:
- Understand the evolution and trends for machine learning.
- Know how machine learning is being used across different industries.
- Become familiar with the tools, skills and services available to implement machine learning within an organization.
- Understand how machine learning can be used to enhance data mining and analysis.
- Learn what a data middle backend is, and how it is being used by businesses.
- Understand the role that big data and intelligent applications are playing across industries.
In this instructor-led, live training, participants will learn the essentials of MemSQL for development and administration.
By the end of this training, participants will be able to:
- Understand the key concepts and characteristics of MemSQL
- Install, design, maintain, and operate MemSQL
- Optimize schemas in MemSQL
- Improve queries in MemSQL
- Benchmark performance in MemSQL
- Build real-time data applications using MemSQL
Audience
- Developers
- Administrators
- Operation Engineers
Format of the course
- Part lecture, part discussion, exercises and heavy hands-on practice
By the end of this training, participants will be able to build producer and consumer applications for real-time stream data procesing.
Audience
- Developers
- Administrators
Format of the course
- Part lecture, part discussion, exercises and heavy hands-on practice
Note
- To request a customized training for this course, please contact us to arrange.
This instructor-led, live training introduces the concepts and approaches for implementing geospacial analytics and walks participants through the creation of a predictive analysis application using Magellan on Spark.
By the end of this training, participants will be able to:
- Efficiently query, parse and join geospatial datasets at scale
- Implement geospatial data in business intelligence and predictive analytics applications
- Use spatial context to extend the capabilities of mobile devices, sensors, logs, and wearables
Format of the Course
- Interactive lecture and discussion.
- Lots of exercises and practice.
- Hands-on implementation in a live-lab environment.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.
In this instructor-led live training, participants will learn how to use Apache Kylin to set up a real-time data warehouse.
By the end of this training, participants will be able to:
- Consume real-time streaming data using Kylin
- Utilize Apache Kylin's powerful features, rich SQL interface, spark cubing and subsecond query latency
Note
- We use the latest version of Kylin (as of this writing, Apache Kylin v2.0)
Audience
- Big data engineers
- Big Data analysts
Format of the course
- Part lecture, part discussion, exercises and heavy hands-on practice
By the end of this training, participants will be able to:
- Install and configure Confluent KSQL.
- Set up a stream processing pipeline using only SQL commands (no Java or Python coding).
- Carry out data filtering, transformations, aggregations, joins, windowing, and sessionization entirely in SQL.
- Design and deploy interactive, continuous queries for streaming ETL and real-time analytics.
This course for KNIME Analytics Platform is an ideal opportunity for beginners, advanced users and KNIME experts to be introduced to KNIME, to learn how to use it more effectively, and how to create clear, comprehensive reports based on KNIME workflows
In this instructor-led, live training, participants will learn how to integrate Kafka Streams into a set of sample Java applications that pass data to and from Apache Kafka for stream processing.
By the end of this training, participants will be able to:
- Understand Kafka Streams features and advantages over other stream processing frameworks
- Process stream data directly within a Kafka cluster
- Write a Java or Scala application or microservice that integrates with Kafka and Kafka Streams
- Write concise code that transforms input Kafka topics into output Kafka topics
- Build, package and deploy the application
Audience
- Developers
Format of the course
- Part lecture, part discussion, exercises and heavy hands-on practice
Notes
- To request a customized training for this course, please contact us to arrange
By the end of this training, participants will be able to:
- Understand NiFi's architecture and dataflow concepts.
- Develop extensions using NiFi and third-party APIs.
- Custom develop their own Apache Nifi processor.
- Ingest and process real-time data from disparate and uncommon file formats and data sources.
In this instructor-led, live training (onsite or remote), participants will learn how to set up and integrate different Stream Processing frameworks with existing big data storage systems and related software applications and microservices.
By the end of this training, participants will be able to:
- Install and configure different Stream Processing frameworks, such as Spark Streaming and Kafka Streaming.
- Understand and select the most appropriate framework for the job.
- Process of data continuously, concurrently, and in a record-by-record fashion.
- Integrate Stream Processing solutions with existing databases, data warehouses, data lakes, etc.
- Integrate the most appropriate stream processing library with enterprise applications and microservices.
Audience
- Developers
- Software architects
Format of the Course
- Part lecture, part discussion, exercises and heavy hands-on practice
Notes
- To request a customized training for this course, please contact us to arrange.
- Developers
Format of the Course
- Lectures, hands-on practice, small tests along the way to gauge understanding
Impala enables users to issue low-latency SQL queries to data stored in Hadoop Distributed File System and Apache Hbase without requiring data movement or transformation.
Audience
This course is aimed at analysts and data scientists performing analysis on data stored in Hadoop via Business Intelligence or SQL tools.
After this course delegates will be able to
- Extract meaningful information from Hadoop clusters with Impala.
- Write specific programs to facilitate Business Intelligence in Impala SQL Dialect.
- Troubleshoot Impala.
By the end of this training, participants will be able to:
- Use Hortonworks to reliably run Hadoop at a large scale.
- Unify Hadoop's security, governance, and operations capabilities with Spark's agile analytic workflows.
- Use Hortonworks to investigate, validate, certify and support each of the components in a Spark project.
- Process different types of data, including structured, unstructured, in-motion, and at-rest.
We will walk a developer through HBase architecture and data modelling and application development on HBase. It will also discuss using MapReduce with HBase, and some administration topics, related to performance optimization. The course is very hands-on with lots of lab exercises.
Duration : 3 days
Audience : Developers & Administrators
In this instructor-led, live training, participants will learn how to work with Hadoop, MapReduce, Pig, and Spark using Python as they step through multiple examples and use cases.
By the end of this training, participants will be able to:
- Understand the basic concepts behind Hadoop, MapReduce, Pig, and Spark
- Use Python with Hadoop Distributed File System (HDFS), MapReduce, Pig, and Spark
- Use Snakebite to programmatically access HDFS within Python
- Use mrjob to write MapReduce jobs in Python
- Write Spark programs with Python
- Extend the functionality of pig using Python UDFs
- Manage MapReduce jobs and Pig scripts using Luigi
Audience
- Developers
- IT Professionals
Format of the course
- Part lecture, part discussion, exercises and heavy hands-on practice
This course is intended to demystify big data/hadoop technology and to show it is not difficult to understand.
This course introduces Project Managers to the most popular Big Data processing framework: Hadoop.
In this instructor-led training, participants will learn the core components of the Hadoop ecosystem and how these technologies can be used to solve large-scale problems. In learning these foundations, participants will also improve their ability to communicate with the developers and implementers of these systems as well as the data scientists and analysts that many IT projects involve.
Audience
- Project Managers wishing to implement Hadoop into their existing development or IT infrastructure
- Project Managers needing to communicate with cross-functional teams that include big data engineers, data scientists and business analysts
Format of the course
- Part lecture, part discussion, exercises and heavy hands-on practice
By the end of this training, participants will be able to:
- Install and configure Apachi NiFi.
- Source, transform and manage data from disparate, distributed data sources, including databases and big data lakes.
- Automate dataflows.
- Enable streaming analytics.
- Apply various approaches for data ingestion.
- Transform Big Data and into business insights.
By the end of this training, participants will be able to:
- Manage Teradata space.
- Protect and distribute data in Teradata.
- Read Explain Plan.
- Improve SQL proficiency.
- Use main utilities of Teradata.
It divides into two packages:
-
spark.mllib contains the original API built on top of RDDs.
-
spark.ml provides higher-level API built on top of DataFrames for constructing ML pipelines.
Audience
This course is directed at engineers and developers seeking to utilize a built in Machine Library for Apache Spark
By the end of this training, participants will be able to:
- Install and configure Zeppelin
- Develop, organize, execute and share data in a browser-based interface
- Visualize results without referring to the command line or cluster details
- Execute and collaborate on long workflows
- Work with any of a number of plug-in language/data-processing-backends, such as Scala (with Apache Spark), Python (with Apache Spark), Spark SQL, JDBC, Markdown and Shell.
- Integrate Zeppelin with Spark, Flink and Map Reduce
- Secure multi-user instances of Zeppelin with Apache Shiro
This instructor-led, live training introduces the challenges of serving large-scale data and walks participants through the creation of an application that can compute responses to user requests, over large datasets in real-time.
By the end of this training, participants will be able to:
- Use Vespa to quickly compute data (store, search, rank, organize) at serving time while a user waits
- Implement Vespa into existing applications involving feature search, recommendations, and personalization
- Integrate and deploy Vespa with existing big data systems such as Hadoop and Storm.
Audience
- Developers
Format of the course
- Part lecture, part discussion, exercises and heavy hands-on practice