Ensemble Learning for AI Developers Learn Bagging Stacking and Boosting Methods with Use Cases Alok Kumar Mayank Jain all chapter instant download
Ensemble Learning for AI Developers Learn Bagging Stacking and Boosting Methods with Use Cases Alok Kumar Mayank Jain all chapter instant download
com
https://ebookmeta.com/product/ensemble-learning-for-ai-
developers-learn-bagging-stacking-and-boosting-methods-with-
use-cases-alok-kumar-mayank-jain/
OR CLICK HERE
DOWLOAD NOW
https://ebookmeta.com/product/ensemble-methods-for-machine-learning-
meap-version-6-gautam-kunapuli/
ebookmeta.com
https://ebookmeta.com/product/next-generation-electrified-vehicles-
optimised-for-the-infrastructure-1st-edition-bernhard-brandstatter/
ebookmeta.com
The Guest is a Goner Sedona Spirit Cozy Mystery 1 1st
Edition Carly Winter
https://ebookmeta.com/product/the-guest-is-a-goner-sedona-spirit-cozy-
mystery-1-1st-edition-carly-winter/
ebookmeta.com
https://ebookmeta.com/product/handbook-of-the-anthropocene-humans-
between-heritage-and-future-nathanael-wallenhorst/
ebookmeta.com
https://ebookmeta.com/product/essential-math-for-ai-fifth-early-
release-hala-nelson/
ebookmeta.com
https://ebookmeta.com/product/the-vulnerable-observer-anthropology-
that-breaks-your-heart-25th-anniversary-edition-ruth-behar/
ebookmeta.com
https://ebookmeta.com/product/machine-learning-for-data-science-
handbook-data-mining-and-knowledge-discovery-handbook-3rd-lior-rokach/
ebookmeta.com
Scar High Heat BBW Mountain Man Instalove 1st Edition
Kelsie Calloway
https://ebookmeta.com/product/scar-high-heat-bbw-mountain-man-
instalove-1st-edition-kelsie-calloway-2/
ebookmeta.com
Alok Kumar and Mayank Jain
Mayank Jain
Gurugram, India
Standard Apress
The publisher, the authors and the editors are safe to assume that the
advice and information in this book are believed to be true and accurate
at the date of publication. Neither the publisher nor the authors or the
editors give a warranty, expressed or implied, with respect to the
material contained herein or for any errors or omissions that may have
been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.
Distributed to the book trade worldwide by Springer Science+Business
Media New York, 233 Spring Street, 6th Floor, New York, NY 10013.
Phone 1-800-SPRINGER, fax (201) 348-4505, e-mail orders-
[email protected], or visit www.springeronline.com. Apress Media,
LLC is a California LLC and the sole member (owner) is Springer
Science + Business Media Finance Inc (SSBM Finance Inc). SSBM
Finance Inc is a Delaware corporation.
This book is dedicated to my family for their unfettered support.
—Alok Kumar
To my family and Anand Sir, who unbounded my mind.
—Mayank Jain
Introduction
Ensemble learning is fast becoming a popular choice for machine
learning models in the data science world. Ensemble methods combine
the output of machine learning models in many interesting ways. Even
after years of working on machine learning projects, we were unaware
of the power of ensemble methods, as this topic is usually neglected or
only given a brief overview in most machine learning courses and
books. Like many others, we came to know about the power of
ensemble methods by checking competitive machine learning
scenarios. Competitive machine learning platforms, like Kaggle, offer an
unbiased review of machine learning techniques. For the past few
years, ensemble learning methods have consistently outperformed
competitive metrics. This itself speaks to the benefit of learning
ensembling techniques. The objective of this book is to help you learn
ensembles and apply them effectively in real-world situations.
This book starts by explaining why ensemble learning techniques
are needed. The objective of Chapter 1 is to build a basic understanding
of various ensembling techniques. Chapters 2, 3, and 4, cover various
ensemble techniques and are grouped by how we mix training data,
models, and combinations. In these chapters, you learn about some of
the most important ensemble learning techniques like random forests,
bagging, bootstrap aggregating, stacking, and cross validation methods.
Chapter 5 covers ensemble learning libraries. The libraries help you
accelerate your experimentation and implementation. Chapter 6 covers
techniques for integrating ensembles into real-world machine learning
workflows.
This book presents a condensed, easy-to-understand way to learn
and apply ensemble learning concepts in real-world use-cases, which
otherwise takes a lot of back and forth learning for a budding data
scientist. The code in this book (Python scripts) can be used as an
extension to your projects or for reference.
Acknowledgments
We would really like to thank the team at Apress for all their help and
support in our journey writing this book. It has been a pleasure to write
this book, and the team at Apress are certainly a big part of that. Many
thanks to Acquisitions Editor, Celestin Suresh John, who was a great
help in narrowing the scope of this project and making it more
accessible. Many thanks to our development coordinators and our
editors, Aditee Mirashi and Laura C. Berendson. Their continuous
follow-up helped us tremendously with keeping us on track and
focused. We are also thankful to Ashutosh Parida for his valuable
technical feedback.
As always, we would like to thank our family for all their help and
support. Without their support and assistance, we couldn’t have
completed this book.
Lastly, we would like to thank the open source community, who has
helped democratize the knowledge. All the libraries used in this book
come from open source projects.
Table of Contents
Chapter 1:Why Ensemble Techniques Are Needed
Mixing Training Data
Mixing Combinations
Mixing Models
Summary
Chapter 2:Mixing Training Data
Decision Trees
Sampling a Dataset
Sampling Without Replacement (WOR)
Sampling with Replacement (WR)
Bagging (Bootstrap Aggregating)
k-Fold Cross-Validation
Stratified k-Fold Cross-Validation
Summary
Chapter 3:Mixing Models
Voting Ensembles
Hard Voting
Averaging/Soft Voting
Hyperparameter Tuning Ensembles
Horizontal Voting Ensembles
Snapshot Ensembles
Summary
Chapter 4:Mixing Combinations
Boosting
AdaBoost
Gradient Boosting
XGBoost
Stacking
Summary
Chapter 5:Using Ensemble Learning Libraries
ML-Ensemble
Multilayer Ensembles
Ensemble Model Selection
Scale XGBoost via Dask
Dask Arrays and Dataframes
Dask-ML
Scaling XGBoost
Microsoft LightGBM
AdaNet
Summary
Chapter 6:Tips and Best Practices
Feature Selection Using Random Forests
Feature Transformations with Ensembles of Trees
Building a Preprocessing Pipeline for a Random Forest
Classifier
Isolation Forest for Outlier Detection
Scaling Ensembles with Dask
Preprocessing
Hyperparameter Search
Distributed Ensemble Fitting
Summary
Thank You
Index
About the Authors
Alok Kumar
is an AI practitioner and Innovation Lead
at Publicis Sapient. He has extensive
experience in leading strategic initiatives
and driving cutting-edge, fast-paced
innovations. His work has won several
awards. Alok is passionate about
democratizing AI knowledge. He
manages multiple nonprofit learning and
creative groups in National capital region
of india NCR. Find him on LinkedIn at
https://in.linkedin.com/in/al
oksaan .
Mayank Jain
is a technology manager and AI/ML
expert at the Publicis Sapient Kepler
Innovation Lab. He has more than 10
years of experience working on cutting-
edge projects that involve making
computers see and think by using
techniques like deep learning, machine
learning, and computer vision. He has
written for several international
publications, he has several patents in
his name, and has been awarded
multiple times for his contributions. Find
him on LinkedIn at
https://in.linkedin.com/in/mayankjain7 .
About the Technical Reviewer
Ashutosh Parida
is the head of analytics at Manipal Global
Education Services in Bangalore. He has
a decade of experience in data science,
including vision, natural-language
understanding, recommendation
engines, and forecasting. He has been
site lead for multiple projects and has
launched products that serve millions of
users. He also has open source
contributions to his credit. Ashutosh has
a bachelor’s degree in computers science
and engineering from IIIT-Hyderabad. He
has been in the IT industry for 15 years,
having worked with Oracle, Samsung,
Akamai, and Qualcomm. Find him on
LinkedIn at
https://www.linkedin.com/in/ashutoshparida/ .
© Alok Kumar and Mayank Jain 2020
A. Kumar, M. Jain, Ensemble Learning for AI Developers
https://doi.org/10.1007/978-1-4842-5940-5_1
Mixing Combinations
To better understand what mixing, or varying, combinations mean in
the context of ensemble learning, let’s look at an analogy of the learning
process of students in a class.
Random documents with unrelated
content Scribd suggests to you:
AT THE BOSTON BANQUET.