Search icon
Subscription
0
Cart icon
Close icon
You have no products in your basket yet
Save more on your purchases!
Savings automatically calculated. No voucher code required
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletters
Free Learning
Arrow right icon
Building ETL Pipelines with Python

You're reading from  Building ETL Pipelines with Python

Product type Book
Published in Sep 2023
Publisher Packt
ISBN-13 9781804615256
Pages 246 pages
Edition 1st Edition
Languages
Authors (2):
Brij Kishore Pandey Brij Kishore Pandey
Profile icon Brij Kishore Pandey
Emily Ro Schoof Emily Ro Schoof
Profile icon Emily Ro Schoof
View More author details

Table of Contents (22) Chapters

Preface 1. Part 1:Introduction to ETL, Data Pipelines, and Design Principles
2. Chapter 1: A Primer on Python and the Development Environment 3. Chapter 2: Understanding the ETL Process and Data Pipelines 4. Chapter 3: Design Principles for Creating Scalable and Resilient Pipelines 5. Part 2:Designing ETL Pipelines with Python
6. Chapter 4: Sourcing Insightful Data and Data Extraction Strategies 7. Chapter 5: Data Cleansing and Transformation 8. Chapter 6: Loading Transformed Data 9. Chapter 7: Tutorial – Building an End-to-End ETL Pipeline in Python 10. Chapter 8: Powerful ETL Libraries and Tools in Python 11. Part 3:Creating ETL Pipelines in AWS
12. Chapter 9: A Primer on AWS Tools for ETL Processes 13. Chapter 10: Tutorial – Creating an ETL Pipeline in AWS 14. Chapter 11: Building Robust Deployment Pipelines in AWS 15. Part 4:Automating and Scaling ETL Pipelines
16. Chapter 12: Orchestration and Scaling in ETL Pipelines 17. Chapter 13: Testing Strategies for ETL Pipelines 18. Chapter 14: Best Practices for ETL Pipelines 19. Chapter 15: Use Cases and Further Reading 20. Index 21. Other Books You May Enjoy

Building an ETL pipeline using various AWS services

With AWS CodeCommit, CodeBuild, and CodeDeploy set up in your environment, we can walk through how to build our CI/CD pipeline for ETL jobs.

Setting up a CodeCommit repository

As our first step, we’ll initialize a new repository in AWS CodeCommit that will store the scripts for our ETL jobs.

Given your CodeCommit repository is named my-etl-jobs, clone the repository to your local system using the following command (replace [region] with your AWS Region):

git clone https://git-codecommit.[region].amazonaws.com/v1/repos/my-etl-jobs

Add demo-code in the etl directory to your CodeCommit repository:

git add etl/*git commit -m "Initial commit of ETL script"
git push

Orchestrating with AWS CodePipeline

Now that we’ve created our repository, defined our build, and set up our deployment, we need a way to automate the whole process. This is where AWS CodePipeline comes in. AWS CodePipeline is a...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $15.99/month. Cancel anytime}