Data Lake for Enterprises

More Information
  • Build an enterprise-level data lake using the relevant big data technologies
  • Understand the core of the Lambda architecture and how to apply it in an enterprise
  • Learn the technical details around Sqoop and its functionalities
  • Integrate Kafka with Hadoop components to acquire enterprise data
  • Use flume with streaming technologies for stream-based processing
  • Understand stream- based processing with reference to Apache Spark Streaming
  • Incorporate Hadoop components and know the advantages they provide for enterprise data lakes
  • Build fast, streaming, and high-performance applications using ElasticSearch
  • Make your data ingestion process consistent across various data formats with configurability
  • Process your data to derive intelligence using machine learning algorithms

The term "Data Lake" has recently emerged as a prominent term in the big data industry. Data scientists can make use of it in deriving meaningful insights that can be used by businesses to redefine or transform the way they operate. Lambda architecture is also emerging as one of the very eminent patterns in the big data landscape, as it not only helps to derive useful information from historical data but also correlates real-time data to enable business to take critical decisions. This book tries to bring these two important aspects — data lake and lambda architecture—together.

This book is divided into three main sections. The first introduces you to the concept of data lakes, the importance of data lakes in enterprises, and getting you up-to-speed with the Lambda architecture. The second section delves into the principal components of building a data lake using the Lambda architecture. It introduces you to popular big data technologies such as Apache Hadoop, Spark, Sqoop, Flume, and ElasticSearch. The third section is a highly practical demonstration of putting it all together, and shows you how an enterprise data lake can be implemented, along with several real-world use-cases. It also shows you how other peripheral components can be added to the lake to make it more efficient.

By the end of this book, you will be able to choose the right big data technologies using the lambda architectural patterns to build your enterprise data lake.

  • Build a full-fledged data lake for your organization with popular big data technologies using the Lambda architecture as the base
  • Delve into the big data technologies required to meet modern day business strategies
  • A highly practical guide to implementing enterprise data lakes with lots of examples and real-world use-cases
Page Count 596
Course Length 17 hours 52 minutes
ISBN 9781787281349
Date Of Publication 30 May 2017


Tomcy John

Tomcy John lives in Dubai (United Arab Emirates), hailing from Kerala (India), and is an enterprise Java specialist with a degree in Engineering (B Tech) and over 14 years of experience in several industries. He's currently working as principal architect at Emirates Group IT, in their core architecture team. Prior to this, he worked with Oracle Corporation and Ernst & Young. His main specialization is in building enterprise-grade applications and he acts as chief mentor and evangelist to facilitate incorporating new technologies as corporate standards in the organization. Outside of his work, Tomcy works very closely with young developers and engineers as mentors and speaks at various forums as a technical evangelist on many topics ranging from web and middleware all the way to various persistence stores.

Pankaj Misra

Pankaj Misra has been a technology evangelist, holding a bachelor’s degree in engineering, with over 16 years of experience across multiple business domains and technologies. He has been working with Emirates Group IT since 2015, and has worked with various other organizations in the past. He specializes in architecting and building multi-stack solutions and implementations. He has also been a speaker at technology forums in India and has built products with scale-out architecture that support high-volume, near-real-time data processing and near-real-time analytics.