Search icon
Subscription
0
Cart icon
Close icon
You have no products in your basket yet
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletters
Free Learning
Arrow right icon
Practical Deep Learning at Scale with MLflow

You're reading from  Practical Deep Learning at Scale with MLflow

Product type Book
Published in Jul 2022
Publisher Packt
ISBN-13 9781803241333
Pages 288 pages
Edition 1st Edition
Languages
Author (1):
Yong Liu Yong Liu
Profile icon Yong Liu

Table of Contents (17) Chapters

Preface 1. Section 1 - Deep Learning Challenges and MLflow Prime
2. Chapter 1: Deep Learning Life Cycle and MLOps Challenges 3. Chapter 2: Getting Started with MLflow for Deep Learning 4. Section 2 –
Tracking a Deep Learning Pipeline at Scale
5. Chapter 3: Tracking Models, Parameters, and Metrics 6. Chapter 4: Tracking Code and Data Versioning 7. Section 3 –
Running Deep Learning Pipelines at Scale
8. Chapter 5: Running DL Pipelines in Different Environments 9. Chapter 6: Running Hyperparameter Tuning at Scale 10. Section 4 –
Deploying a Deep Learning Pipeline at Scale
11. Chapter 7: Multi-Step Deep Learning Inference Pipeline 12. Chapter 8: Deploying a DL Inference Pipeline at Scale 13. Section 5 – Deep Learning Model Explainability at Scale
14. Chapter 9: Fundamentals of Deep Learning Explainability 15. Chapter 10: Implementing DL Explainability with MLflow 16. Other Books You May Enjoy

Running remotely in the cloud with remote code in GitHub

The most reliable way to reproduce a DL pipeline is to point to a specific version of the project code in GitHub and then run it in the cloud without invoking any local resources. This way, we know the exact version of the code as well as using the same running environment defined in the project. Let's see how this works with our DL pipeline.

As a prerequisite and a reminder, the following three environment variables need to be set up before you issue the MLflow run command to complete this section of the learning:

export MLFLOW_TRACKING_URI=databricks
export DATABRICKS_TOKEN=[databricks_token]
export DATABRICKS_HOST='https://[your databricks host name/'

We already know how to set up these environment variables from the last section. There is potentially one more setup needed, which is to allow your Databricks server to access your GitHub repository if it is non-public (see the following GitHub Token...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $15.99/month. Cancel anytime}