Big Data Diploma : Data Management&Analytics CLS Learn
Interested in increasing your knowledge of the Big Data Systems?
This course is for those who are new to big data and drive your through hands-on practical experience to learn Apache Hadoop.
You will gain an experience with Ingesting data into Hadoop file system, working with data in batch and Stream processing, Deliver visualization and insights of Data and too many technical skills.
Diploma Duration : 104 Hours
Audience Profile:
This course targeted for
Who are interested in big data management, data engineering and data analysis using Big Data technologies
Database Administrators, Developers & System Administrators which aims to get hands on training about big data and NoSQL
Course Outcomes:
Students will be trained as big data Engineers with real world hands-on experience in Hadoop Administration, ETL (Batch/Stream), Sql-On-Hadoop and many other technologies related to Big Data.
Instructor Bio:
Having more than 7 years building data Integration & Data Warehouse project at different governmental project. Leading many project in Big Data Platform “Design, Implementation & Development”.
Working for National project at “Administrative Control Authority” building data warehouse and data integration.
Working for National project at “Traffic Information Technology Department – Ministry of Interior”.
Course Outline in details:
Beginner Level: 3 Days / 24 Hours
Introduction to Big Data
Intro to Distributed Systems & HDFS
Exploring Big Data Ecosystem & Distributions iv. Basic Intro to NoSQL
Hadoop Fundamentals
Intro to Hadoop
Introduction to MapReduce
Intro to YARN
Basic Hadoop Cluster Implementation
SQL-On-Hadoop
Intro to SQL-On-Hadoop
Intro to Hive iii. Basic Implementation of Hive & Hive Server 2
Ingesting Data into Hadoop
Intro to Sqoop
Concept of Stream Processing
Into to Kafka
Advanced Level: 5 Days / 40 Hours
MapReduce In Depth
Hadoop Architecture In Depth
Intro to Apache Zookeeper
Advanced Cluster Implementation
SQL-On-Hadoop
Advanced Hive Architecture
Ingesting Data Into Hive Using Sqoop
HiveQL
Intro to Apache HBase
Ingesting Data into Hadoop
Intro to ETL Concepts
Intro to Data Flow using Apache Nifi
Implementing Apache Nifi Cluster
Simple Integration Project with Apache Nifi
Intro to Python Programming
Ingesting Data into Hadoop
Implementing Kafka Cluster
Building Simple Kafka Producer & Consumer Using Python
Intro to Apache Spark
Spark Implementation
Simple Data Analysis with Apache Spark
Big Data Workshop: 5 Days / 40 Hours
Project 1: Data Ingestion from RDBMS into HIVE (ORC File)
Using Sqoop to Connect to RDBMS(Oracle / SQL Server)
Creating Hive Table on ORC File
Hive Table Performance Tuning
Project 2: Data Ingestion Streaming Data Using Kafka into Hadoop/Hive
Using Kafka to Connect to RDBMS(Oracle / SQL Server)
Kafka Advanced Cluster Configuration
Ingesting Data into Hadoop/Hive
Project 3: Data Analysis & Stream Processing using Spark 1
Spark Cluster Implementation
Using PySpark
Processing Data With Spark
Stream Processing With Spark
Project 4: Data Analysis & Stream Processing using Spark & Hive 2
Integrating Spark with Hive
Ingesting Data with Spark into Hive
Using Spark SQL
Project 5: Data Visualization Using Apache
Implementing Apache Zeppelin
Integrating Apache Zeppeling With Spark
Visualize Data With Zeppeling
This course is no longer available.
- Other Location: Dokki
- Duration: 104 Hours
- Timings: Part Time, Weekday evening classes
Course details
Course Overview:Interested in increasing your knowledge of the Big Data Systems?
This course is for those who are new to big data and drive your through hands-on practical experience to learn Apache Hadoop.
You will gain an experience with Ingesting data into Hadoop file system, working with data in batch and Stream processing, Deliver visualization and insights of Data and too many technical skills.
Diploma Duration : 104 Hours
Audience Profile:
This course targeted for
Who are interested in big data management, data engineering and data analysis using Big Data technologies
Database Administrators, Developers & System Administrators which aims to get hands on training about big data and NoSQL
Course Outcomes:
Students will be trained as big data Engineers with real world hands-on experience in Hadoop Administration, ETL (Batch/Stream), Sql-On-Hadoop and many other technologies related to Big Data.
Instructor Bio:
Having more than 7 years building data Integration & Data Warehouse project at different governmental project. Leading many project in Big Data Platform “Design, Implementation & Development”.
Working for National project at “Administrative Control Authority” building data warehouse and data integration.
Working for National project at “Traffic Information Technology Department – Ministry of Interior”.
Course Outline in details:
Beginner Level: 3 Days / 24 Hours
Introduction to Big Data
Intro to Distributed Systems & HDFS
Exploring Big Data Ecosystem & Distributions iv. Basic Intro to NoSQL
Hadoop Fundamentals
Intro to Hadoop
Introduction to MapReduce
Intro to YARN
Basic Hadoop Cluster Implementation
SQL-On-Hadoop
Intro to SQL-On-Hadoop
Intro to Hive iii. Basic Implementation of Hive & Hive Server 2
Ingesting Data into Hadoop
Intro to Sqoop
Concept of Stream Processing
Into to Kafka
Advanced Level: 5 Days / 40 Hours
MapReduce In Depth
Hadoop Architecture In Depth
Intro to Apache Zookeeper
Advanced Cluster Implementation
SQL-On-Hadoop
Advanced Hive Architecture
Ingesting Data Into Hive Using Sqoop
HiveQL
Intro to Apache HBase
Ingesting Data into Hadoop
Intro to ETL Concepts
Intro to Data Flow using Apache Nifi
Implementing Apache Nifi Cluster
Simple Integration Project with Apache Nifi
Intro to Python Programming
Ingesting Data into Hadoop
Implementing Kafka Cluster
Building Simple Kafka Producer & Consumer Using Python
Intro to Apache Spark
Spark Implementation
Simple Data Analysis with Apache Spark
Big Data Workshop: 5 Days / 40 Hours
Project 1: Data Ingestion from RDBMS into HIVE (ORC File)
Using Sqoop to Connect to RDBMS(Oracle / SQL Server)
Creating Hive Table on ORC File
Hive Table Performance Tuning
Project 2: Data Ingestion Streaming Data Using Kafka into Hadoop/Hive
Using Kafka to Connect to RDBMS(Oracle / SQL Server)
Kafka Advanced Cluster Configuration
Ingesting Data into Hadoop/Hive
Project 3: Data Analysis & Stream Processing using Spark 1
Spark Cluster Implementation
Using PySpark
Processing Data With Spark
Stream Processing With Spark
Project 4: Data Analysis & Stream Processing using Spark & Hive 2
Integrating Spark with Hive
Ingesting Data with Spark into Hive
Using Spark SQL
Project 5: Data Visualization Using Apache
Implementing Apache Zeppelin
Integrating Apache Zeppeling With Spark
Visualize Data With Zeppeling Updated on 13 November, 2019
Eligibility / Requirements
Prerequisites:
To gain the most from the workshop, Prior knowledge of Apache Hadoop is not required, and the following is required:
Knowledge of Programming, Basic of “Java, Scala, Or Python”
Knowledge of Relational Database Management Systems.
Basic knowledge of operating systems & Network.
Job roles this course is suitable for:
Senior Data Analyst , Data Analyst , Big Data Engineer , Big Data DeveloperCourse Location
About CLS Learn
Since 1995, CLS Learning solutions is leading the technology learning market in Egypt, the Middle East, and Africa. With our wide network of international partners, trainers, instructors, and technology leaders; we are able to deliver top notch training programs to our students and technology professionals.
25 Years in the market.
We delivered over 4,200 courses to 63,500 professionals in our centers.
We delivered 1,200 courses to 18,240 corporate employees on Site.
See all CLS Learn courses- JavaScript Full stack web developer virtual internship Virtual Bootcamp + Internship at LaimoonAED 1,449Duration: Upto 30 Hours
- Mastering Big Data. Global EdulinkEGP 2,936
EGP 20,884Duration: Upto 12 Hours - Machine Learning Beginner eduCBAEGP 997Duration: Upto 3 Hours