
- 328 pages
- English
- ePUB (mobile friendly)
- Available on iOS & Android
About this book
AI doesn't have to be a black box. These practical techniques help shine a light on your model's mysterious inner workings. Make your AI more transparent, and you'll improve trust in your results, combat data leakage and bias, and ensure compliance with legal requirements. In Interpretable AI, you will learn: Why AI models are hard to interpret
Interpreting white box models such as linear regression, decision trees, and generalized additive models
Partial dependence plots, LIME, SHAP and Anchors, and other techniques such as saliency mapping, network dissection, and representational learning
What fairness is and how to mitigate bias in AI systems
Implement robust AI systems that are GDPR-compliant Interpretable AI opens up the black box of your AI models. It teaches cutting-edge techniques and best practices that can make even complex AI systems interpretable. Each method is easy to implement with just Python and open source libraries. You'll learn to identify when you can utilize models that are inherently transparent, and how to mitigate opacity when your problem demands the power of a hard-to-interpret deep learning model. About the technology
It's often difficult to explain how deep learning models work, even for the data scientists who create them. Improving transparency and interpretability in machine learning models minimizes errors, reduces unintended bias, and increases trust in the outcomes. This unique book contains techniques for looking inside "black box" models, designing accountable algorithms, and understanding the factors that cause skewed results. About the book
Interpretable AI teaches you to identify the patterns your model has learned and why it produces its results. As you read, you'll pick up algorithm-specific approaches, like interpreting regression and generalized additive models, along with tips to improve performance during training. You'll also explore methods for interpreting complex deep learning models where some processes are not easily observable. AI transparency is a fast-moving field, and this book simplifies cutting-edge research into practical methods you can implement with Python. What's inside Techniques for interpreting AI models
Counteract errors from bias, data leakage, and concept drift
Measuring fairness and mitigating bias
Building GDPR-compliant AI systemsAbout the reader
For data scientists and engineers familiar with Python and machine learning.
About the author
Ajay Thampi is a machine learning engineer focused on responsible AI and fairness.Table of ContentsPART 1 INTERPRETABILITY BASICS
1 Introduction
2 White-box models
PART 2 INTERPRETING MODEL PROCESSING
3 Model-agnostic methods: Global interpretability
4 Model-agnostic methods: Local interpretability
5 Saliency mapping
PART 3 INTERPRETING MODEL REPRESENTATIONS
6 Understanding layers and units
7 Understanding semantic similarity
PART 4 FAIRNESS AND BIAS
8 Fairness and mitigating bias
9 Path to explainable AI
Frequently asked questions
- Essential is ideal for learners and professionals who enjoy exploring a wide range of subjects. Access the Essential Library with 800,000+ trusted titles and best-sellers across business, personal growth, and the humanities. Includes unlimited reading time and Standard Read Aloud voice.
- Complete: Perfect for advanced learners and researchers needing full, unrestricted access. Unlock 1.4M+ books across hundreds of subjects, including academic and specialized titles. The Complete Plan also includes advanced features like Premium Read Aloud and Research Assistant.
Please note we cannot support devices running on iOS 13 and Android 7 or earlier. Learn more about using the app.
Information
Table of contents
- inside front cover
- Interpretable AI
- Copyright
- dedication
- brief content
- contents
- front matter
- Part 1. Interpretability basics
- 1 Introduction
- 2 White-box models
- Part 2. Interpreting model processing
- 3 Model-agnostic methods: Global interpretability
- 4 Model-agnostic methods: Local interpretability
- 5 Saliency mapping
- Part 3. Interpreting model representations
- 6 Understanding layers and units
- 7 Understanding semantic similarity
- Part 4. Fairness and bias
- 8 Fairness and mitigating bias
- 9 Path to explainable AI
- Appendix A. Getting set up
- Appendix B. PyTorch
- index
- inside back cover