Search icon
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletters
Free Learning
Arrow right icon
Python Machine Learning

You're reading from  Python Machine Learning

Product type Book
Published in Sep 2015
Publisher Packt
ISBN-13 9781783555130
Pages 454 pages
Edition 1st Edition
Languages
Author (1):
Sebastian Raschka Sebastian Raschka
Profile icon Sebastian Raschka

Table of Contents (21) Chapters

Python Machine Learning
Credits
Foreword
About the Author
About the Reviewers
www.PacktPub.com
Preface
1. Giving Computers the Ability to Learn from Data 2. Training Machine Learning Algorithms for Classification 3. A Tour of Machine Learning Classifiers Using Scikit-learn 4. Building Good Training Sets – Data Preprocessing 5. Compressing Data via Dimensionality Reduction 6. Learning Best Practices for Model Evaluation and Hyperparameter Tuning 7. Combining Different Models for Ensemble Learning 8. Applying Machine Learning to Sentiment Analysis 9. Embedding a Machine Learning Model into a Web Application 10. Predicting Continuous Target Variables with Regression Analysis 11. Working with Unlabeled Data – Clustering Analysis 12. Training Artificial Neural Networks for Image Recognition 13. Parallelizing Neural Network Training with Theano Index

Selecting meaningful features


If we notice that a model performs much better on a training dataset than on the test dataset, this observation is a strong indicator for overfitting. Overfitting means that model fits the parameters too closely to the particular observations in the training dataset but does not generalize well to real data—we say that the model has a high variance. A reason for overfitting is that our model is too complex for the given training data and common solutions to reduce the generalization error are listed as follows:

  • Collect more training data

  • Introduce a penalty for complexity via regularization

  • Choose a simpler model with fewer parameters

  • Reduce the dimensionality of the data

Collecting more training data is often not applicable. In the next chapter, we will learn about a useful technique to check whether more training data is helpful at all. In the following sections and subsections, we will look at common ways to reduce overfitting by regularization and dimensionality...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at €14.99/month. Cancel anytime}