Tree-based Machine Learning Algorithms
- Decision Trees, Random Forests, and Boosting
- Indbinding:
- Paperback
- Sideantal:
- 110
- Udgivet:
- 9. september 2017
- Størrelse:
- 152x229x6 mm.
- Vægt:
- 159 g.
- 8-11 hverdage.
- 27. november 2024
På lager
Normalpris
Abonnementspris
- Rabat på køb af fysiske bøger
- 1 valgfrit digitalt ugeblad
- 20 timers lytning og læsning
- Adgang til 70.000+ titler
- Ingen binding
Abonnementet koster 75 kr./md.
Ingen binding og kan opsiges når som helst.
- 1 valgfrit digitalt ugeblad
- 20 timers lytning og læsning
- Adgang til 70.000+ titler
- Ingen binding
Abonnementet koster 75 kr./md.
Ingen binding og kan opsiges når som helst.
Beskrivelse af Tree-based Machine Learning Algorithms
Get a hands-on introduction to building and using decision trees and random forests. Tree-based machine learning algorithms are used to categorize data based on known outcomes in order to facilitate predicting outcomes in new situations. You will learn not only how to use decision trees and random forests for classification and regression, and some of their respective limitations, but also how the algorithms that build them work. Each chapter introduces a new data concern and then walks you through modifying the code, thus building the engine just-in-time. Along the way you will gain experience making decision trees and random forests work for you. This book uses Python, an easy to read programming language, as a medium for teaching you how these algorithms work, but it isn't about teaching you Python, or about using pre-built machine learning libraries specific to Python. It is about teaching you how some of the algorithms inside those kinds of libraries work and why we might use them, and gives you hands-on experience that you can take back to your favorite programming environment.
Table of Contents:
A brief introduction to decision treesChapter 1: Branching - uses a greedy algorithm to build a decision tree from data that can be partitioned on a single attribute.Chapter 2: Multiple Branches - examines several ways to partition data in order to generate multi-level decision trees.Chapter 3: Continuous Attributes - adds the ability to partition numeric attributes using greater-than.Chapter 4: Pruning - explore ways of reducing the amount of error encoded in the tree.Chapter 5: Random Forests - introduces ensemble learning and feature engineering.Chapter 6: Regression Trees - investigates numeric predictions, like age, price, and miles per gallon.Chapter 7: Boosting - adjusts the voting power of the randomly selected decision trees in the random forest in order to improve its ability to predict outcomes.
Table of Contents:
A brief introduction to decision treesChapter 1: Branching - uses a greedy algorithm to build a decision tree from data that can be partitioned on a single attribute.Chapter 2: Multiple Branches - examines several ways to partition data in order to generate multi-level decision trees.Chapter 3: Continuous Attributes - adds the ability to partition numeric attributes using greater-than.Chapter 4: Pruning - explore ways of reducing the amount of error encoded in the tree.Chapter 5: Random Forests - introduces ensemble learning and feature engineering.Chapter 6: Regression Trees - investigates numeric predictions, like age, price, and miles per gallon.Chapter 7: Boosting - adjusts the voting power of the randomly selected decision trees in the random forest in order to improve its ability to predict outcomes.
Brugerbedømmelser af Tree-based Machine Learning Algorithms
Giv din bedømmelse
For at bedømme denne bog, skal du være logget ind.Andre købte også..
© 2024 Pling BØGER Registered company number: DK43351621