Search icon
Subscription
0
Cart icon
Close icon
You have no products in your basket yet
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletters
Free Learning
Arrow right icon
Learning Pentaho Data Integration 8 CE - Third Edition

You're reading from  Learning Pentaho Data Integration 8 CE - Third Edition

Product type Book
Published in Dec 2017
Publisher Packt
ISBN-13 9781788292436
Pages 500 pages
Edition 3rd Edition
Languages

Table of Contents (23) Chapters

Title Page
Credits
About the Author
About the Reviewers
www.PacktPub.com
Customer Feedback
Preface
1. Getting Started with Pentaho Data Integration 2. Getting Started with Transformations 3. Creating Basic Task Flows 4. Reading and Writing Files 5. Manipulating PDI Data and Metadata 6. Controlling the Flow of Data 7. Cleansing, Validating, and Fixing Data 8. Manipulating Data by Coding 9. Transforming the Dataset 10. Performing Basic Operations with Databases 11. Loading Data Marts with PDI 12. Creating Portable and Reusable Transformations 13. Implementing Metadata Injection 14. Creating Advanced Jobs 15. Launching Transformations and Jobs from the Command Line 16. Best Practices for Designing and Deploying a PDI Project

Working with Big Data and cloud sources


While flat files and databases are the most common types of sources that you use from PDI, there are many other types of data sources available. People have started to leverage the capabilities of tools such as Hadoop, NoSQL databases, and cloud services. In this section, you will learn to connect, read data from, and load data into some of these big data sources with PDI.

Reading files from an AWS S3 instance

S3 is a scalable storage space and is a common location for files to be processed. If you have files in S3 and want to read them, you don't have to download them. PDI allows you to read those files directly from the Amazon Web Services (AWS) Simple Storage Service (S3) instance.

The step that you will use for doing this is the S3 CSV Input step. This step works very similarly to the CSV Input step. The biggest difference is that to access the file, you have to provide the bucket name where the file is located and the dual keys to access the bucket...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $15.99/month. Cancel anytime}