Search icon
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletters
Free Learning
Arrow right icon
Data Engineering with Google Cloud Platform - Second Edition

You're reading from  Data Engineering with Google Cloud Platform - Second Edition

Product type Book
Published in Apr 2024
Publisher Packt
ISBN-13 9781835080115
Pages 476 pages
Edition 2nd Edition
Languages
Author (1):
Adi Wijaya Adi Wijaya
Profile icon Adi Wijaya

Table of Contents (19) Chapters

Preface 1. Part 1: Getting Started with Data Engineering with GCP
2. Chapter 1: Fundamentals of Data Engineering 3. Chapter 2: Big Data Capabilities on GCP 4. Part 2: Build Solutions with GCP Components
5. Chapter 3: Building a Data Warehouse in BigQuery 6. Chapter 4: Building Workflows for Batch Data Loading Using Cloud Composer 7. Chapter 5: Building a Data Lake Using Dataproc 8. Chapter 6: Processing Streaming Data with Pub/Sub and Dataflow 9. Chapter 7: Visualizing Data to Make Data-Driven Decisions with Looker Studio 10. Chapter 8: Building Machine Learning Solutions on GCP 11. Part 3: Key Strategies for Architecting Top-Notch Solutions
12. Chapter 9: User and Project Management in GCP 13. Chapter 10: Data Governance in GCP 14. Chapter 11: Cost Strategy in GCP 15. Chapter 12: CI/CD on GCP for Data Engineers 16. Chapter 13: Boosting Your Confidence as a Data Engineer 17. Index 18. Other Books You May Enjoy

Summary

In this chapter, we learned about Cloud Composer. Having learned about Cloud Composer, we then needed to know how to work with Airflow. We realized that as an open source tool, Airflow has a wide range of features. We focused on how to use Airflow to help us build a data pipeline for our BigQuery data warehouse. There are a lot more features and capabilities in Airflow that are not covered in this book. You can always expand your skills in this area, but you will already have a good foundation after finishing this chapter.

As a tool, Airflow is fairly simple. You just need to know how to write a Python script to define DAGs. We’ve learned in the Level 1 DAG exercise that you just need to write simple code to build your first DAG, but a complication arises when it comes to best practices, as there are a lot of best practices that you can follow. At the same time, there are also a lot of potential bad practices that Airflow developers can make.

By learning the examples...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at €14.99/month. Cancel anytime}