Follow Datanami:
April 16, 2014

Cloudera Introduces Big Data Training Course

PALO ALTO, Calif., April 16 — Cloudera, the leader in enterprise analytic data management powered by Apache Hadoop, today announced the industry’s first hands-on Big Data training course that teaches developers to build end-to-end applications for a Hadoop-based enterprise data hub (EDH). The new four-day course, called Designing and Building Big Data Applications, prepares data professionals to use an EDH’s full capabilities to build custom, converged applications that enable their organizations to achieve greater value from data and solve real-world problems.

Enterprises are increasingly moving beyond the fundamentals of batch processing to analyze all their data and address larger, more business-relevant challenges. An enterprise data hub offers a much broader set of tools enabled by Hadoop — one place to capture and store all data while keeping the information readily accessible for analytic workloads, connected applications, and existing infrastructure. Creating an enterprise data hub is only the first step; from there, companies can develop custom applications to utilize the full capabilities of the Hadoop stack and drive deeper analytical insights along a consolidated 360-degree view of the organization, customer, and market.

Full-scale Big Data engineering requires developers who can work with multiple tools at once and drive projects that transform their businesses. Cloudera’s new course responds to the demand for developers to create end-to-end applications that drive the full analytics lifecycle from data capture to transformation to visualization. Designing and Building Big Data Applications prepares developers to use the entire toolkit of processing engines, ecosystem projects, and customizable functions to build complete solutions with an enterprise data hub.

Through instructor-led discussion and interactive, hands-on exercises, developers will learn to build applications that integrate the full Big Data value chain, including ingesting data, determining the appropriate file format for storage, processing the stored data, and presenting the results to the end-user in an easily digestible form:

  • Creating a data set with the Kite SDK
  • Developing custom Flume components for data ingestion
  • Managing a multi-stage workflow with Oozie
  • Analyzing data with Crunch
  • Writing user-defined functions for Hive and Impala
  • Transforming data with Morphlines
  • Indexing data with Cloudera Search

“Our goal at Cloudera is to build a proven community of talent — those who are intimately connected to the Hadoop ecosystem and the enterprise data hub. Data professionals are being asked to not just know how the tools work but to deliver real solutions for their organizations, stakeholders, and customers,” said Sarah Sproehnle, vice president, Education Services, Cloudera. “Developers that choose Cloudera for their training stand out among their peers with a true talent advantage. Getting trained not only on the fundamentals of MapReduce, but also on the full scope of tools that enable converged analytics represents a highly valuable strategic component for pushing forward the technical benefits of Hadoop and the business benefits of Big Data. Furthermore, being able to validate those skills with a Cloudera Certified Developer for Apache Hadoop (CCDH) credential is arguably the most valuable step a data professional can take to get noticed by hiring managers today and in the future, particularly as this space matures,” said Sproehnle.

Datanami