Bias in AI Systems

Core Contribution

Examination of how bias manifests in AI systems, from training data through deployment. Provides taxonomy of bias types and mitigation strategies.

Bias Types

Historical Bias: Training data reflects past inequities Representation Bias: Underrepresentation of groups in data Measurement Bias: Features proxy for protected attributes Aggregation Bias: Single model for heterogeneous populations Evaluation Bias: Benchmark datasets not representative

Mitigation Approaches

  • Data augmentation and resampling
  • Fairness constraints in training
  • Post-hoc calibration
  • Diverse evaluation sets

Ongoing Challenge

Bias mitigation requires tradeoffs. Optimizing for one fairness metric may worsen another. Context-dependent decisions required.

Related: 05-atom—evaluation-metric-limitations, 04-molecule—data-cascades-concept