
Cloudera Developer for Apache Hadoop (CCDH) Certification
Description
Cloudera offers enterprise and express versions of its Cloudera Distribution including Apache Hadoop. Cloudera’s view of the importance of qualified big data talent shines through the elements of its certification. It includes the following:
Cloudera Certified Hadoop Developer (CCDH) – This certification is for Developers who are responsible for coding, maintaining, and optimizing Apache Hadoop projects. The CCHD exam features questions that deliver a consistent exam experience
Cloudera Certified Developer for Apache Hadoop (CCDH)
Individuals who gain Cloudera Developer Certification for Apache Hadoop (CCDH) have exhibited their technical knowledge, skill, and ability to write, maintain, and optimize Apache Hadoop development projects. This certification establishes you as a trusted and invaluable resource for those looking for an Apache Hadoop expert. Cloudera Certification undeniably proves your ability to solve problems using Hadoop.
The Motivation for Hadoop
Problems with traditional large-scale systems
Introducing Hadoop
Hadoopable problems
Hadoop: Basic Concepts and HDFS
The Hadoop project and Hadoop components
The Hadoop Distributed File System
Introduction to MapReduce
MapReduce overview
Example: WordCount
Mappers
Reducers
Hadoop Clusters and the Hadoop Ecosystem
Hadoop cluster overview
Hadoop jobs and tasks
Other Hadoop ecosystem components
Writing a MapReduce Program in Java
Basic MapReduce API Concepts
Writing MapReduce Drivers, Mappers, and Reducers in Java
Speeding up Hadoop development by using eclipse
Differences between the old and new MapReduce APIs
Writing a MapReduce Program Using Streaming
Writing Mappers and Reducers with the streaming API
Unit Testing MapReduce Programs
Unit testing
The JUnit and MRUnit testing frameworks
Writing unit tests with MRUnit
Running unit tests
Delving Deeper into the Hadoop API
Using the ToolRunner class
Setting up and tearing down Mappers and Reducers
Decreasing the amount of intermediate data with combiners
Accessing HDFS programmatically
Using the distributed cache
Using the Hadoop API’s Library of Mappers, Reducers, and Partitioners
Practical Development Tips and Techniques
Strategies for debugging MapReduce code
Testing MapReduce code locally by using LocalJobRunner
Writing and viewing log files
Retrieving job information with counters
Reusing objects
Creating map-only MapReduce jobs
Partitioners and Reducers
How partitioners and Reducers work together
Determining the optimal number of Reducers for a job
Writing customer partitioners
Data Input and Output
Creating custom writable and WritableComparable implementations
Saving binary data using sequenceFile and Avro data files
Issues to consider when using file compression
Implementing custom InputFormats and OutputFormats
Common MapReduce Algorithms
Sorting and searching large data sets
Indexing data
Computing term frequency — Inverse Document Frequency
Calculating word co-occurrence
Performing Secondary Sort
Joining Data Sets in MapReduce Jobs
Writing a Map-Side Join
Writing a Reduce-Side Join
Integrating Hadoop into the Enterprise Workflow
Integrating Hadoop into an existing enterprise
Loading data from an RDBMS into HDFS by using Sqoop
Managing real-time data using Flume
Accessing HDFS from legacy systems with FuseDFS and HttpFS
An Introduction to Hive, Imapala, and Pig
The motivation for Hive, Impala, and Pig
Hive overview
Impala overview
Pig overview
Choosing Between Hive, Impala, and Pig
An Introduction to Oozie
Introduction to Oozie
Creating Oozie workflows
Who this course is for:
- All Levels
Instructor
Salem Salem is a software engineer with extensive experience and management skills and works for a hi-tech telecommunication company . the company developers and sell solution that enable service providers to efficiently delivery high quality voice and data services over broadband access networks while leveraging investment in their network infrastructure . as a senior software engineer