Bias
Definition
Systematic errors in AI systems that lead to unfair or discriminatory outcomes, often reflecting biases present in training data or design choices.
AI bias manifests in multiple forms: data bias (training data that underrepresents or misrepresents certain groups), algorithmic bias (model architectures or training procedures that amplify disparities), and deployment bias (using AI in contexts different from its training conditions). High-profile examples include facial recognition systems with higher error rates for dark-skinned faces, hiring algorithms that disadvantaged women, and language models that associate certain professions with specific genders. Addressing bias requires diverse and representative training data, fairness metrics and auditing, inclusive development teams, and ongoing monitoring in production. Regulatory frameworks like the EU AI Act increasingly require bias assessments for high-risk AI applications. Bias mitigation is both a technical challenge and an ethical imperative.
Related Terms
AI Alignment
The research field focused on ensuring AI systems behave in accordance with human values, intentions...
AI Ethics
The branch of applied ethics examining the moral implications and societal impacts of artificial int...
Explainability
The degree to which a human can understand how an AI model arrives at its predictions or decisions, ...
Training Data
The dataset used to teach a machine learning model, consisting of examples from which the model lear...