Building Responsible AI with Python
Learn to identify and mitigate bias with hands-on code examples

By (author) Serg Masis,Serg Masís,Nicholas Schmidt,Dr. Ali El-Sharif,Nick Schmidt,Muhammad Rehman Zafar

ISBN13: 9781803249919

Imprint: Packt Publishing Limited

Publisher: Packt Publishing Limited

Format: Paperback / softback

Published: 30/01/2026

Availability: Available

Description
Generate different forms of machine learning model explanations to gain insight into the logic of models Learn how to measure bias in machine learning models Key Features Measure group fairness, individual fairness and choose the right metric for different scenarios Explain model’s logic using different explanation techniques Mitigate bias at different stages of the machine learning pipeline Book DescriptionAs we incorporate the next wave of AI-enabled products in high-stakes decisions, we need some level of assurance of the safety that we have come to expect from everyday products. Continuing the progress of using AI in high-stakes decisions requires trusting AI-enabled solutions to deliver their promised benefits while protecting the public from harm. Questions about the security, safety, privacy, and fairness of AI-enabled decisions need to be answered as a condition for deploying AI solutions at scale. This book is a guide that will introduce you to key concepts, use cases, tools, and techniques of the emerging field of Responsible AI. We will cover hands-on coding techniques to identify and measure bias. Measuring bias is not enough: we also need to explain and fix our models. This book outlines how to do this throughout the machine learning pipeline. By the end of this book, you will have mastered Python coding techniques of explaining machine learning models’ logic, measuring their fairness at the individual and group levels and monitor them in production environments to detect degradation in their accuracy or fairness.What you will learn Explain the fundamental concepts of Responsible AI Audit models machine learning models to ascertain their group and individual fairness outcomes Apply explanatory techniques to gain insight into the inner logic of complex machine learning models Alter the development of machine learning models using pre-processing, in-processing, and post-processing techniques to mitigate biased outcomes Monitor machine learning models in production to identify drift and manage adverse impacts drift Describe emerging trends in Responsible AI Apply mitigation techniques to models, so that identified biases in models are remediated Monitor models' post-production launch degradation to ensure accuracy and fairness objectives are maintained over time Who this book is forData Scientists, Machine Learning Developers, and Data Science professionals who want to ensure that their machine learning model predictions are non- biased and accurate. Working knowledge of Python programming and basic concepts of machine learning model training and data validation is good to have.
Table of Contents What is Responsible AI and Why Do We Need it? Responsible AI Concepts Bias and Fairness Concepts Introducing our Datasets Individual Fairness Assessment Choosing the Right Metrics: A Summary Choosing the Right Metrics: A Summary Feature Importance Explanations Visual Explanations Decision Trees & Decision Rules Explanations Contrastive and Counterfactual Explanations Choosing the Right Explanation: A Summary Pre-Processing Methods In-Processing Methods Choosing the Right Mitigation Approaches: A Summary Choosing the Right Mitigation Approaches: A Summary What is Model Drift and Why do We Care? Managing the Adverse Impacts of Model Drift
  • Computer programming / software development
  • Computer science
  • Artificial intelligence
  • General (US: Trade)
Height:
Width:
Spine:
Weight:0.00
List Price: £37.99