External Bias Integration

Version: 1.1.0 Status: Production Ready

Overview

ATHENA External Bias Integration enables you to ingest fairness metrics from industry standard bias detection tools and correlate them with ATHENA human trust calibration data. This combination reveals bias amplification, which occurs when human behavior reinforces algorithmic bias.

The Problem

Statistical fairness tools such as IBM AIF360, Fairlearn, and AWS Clarify detect model bias, identifying when an AI produces unfair outputs for specific demographic groups.

These tools miss a critical layer: how humans interact with biased output.

ATHENA detects two key behavioral patterns:

  • Overtrust: Users following biased AI recommendations without appropriate scrutiny

  • Undertrust: Users rejecting accurate AI recommendations for specific groups

When you combine external model bias data with ATHENA human trust patterns, you can identify bias amplification. This is the most dangerous scenario, where human behavior makes algorithmic bias worse.

Supported External Tools

Tool
Tool ID
Metrics Supported

ibm_aif360

Demographic Parity, Equalized Odds, Disparate Impact

fairlearn

Demographic Parity, Equalized Odds, Equal Opportunity

aws_clarify

SHAP values, Disparate Impact, Statistical Parity

google_vertex

Feature attributions

Custom

custom

Any fairness metric

Quick Start

Step 1: Generate Fairness Metrics

Run your bias detection tool. The following example uses IBM AIF360:

Step 2: Send to ATHENA

Step 3: Automatic Correlation

ATHENA performs the following actions automatically:

  1. Stores the fairness signal

  2. Correlates with human decisions for the same model

  3. Calculates amplification risk

  4. Generates alerts when amplification is detected

  5. Fires bias.amplification webhook when severity is high or above

Amplification Risk Levels

Risk Level
Amplification Score
Description

none

0.0 to 0.09

No amplification detected

low

0.1 to 0.24

Minor amplification. Monitor the situation.

medium

0.25 to 0.49

Moderate amplification. Investigation recommended.

high

0.5 to 0.74

Significant amplification. Action required.

critical

0.75 to 1.0

Severe amplification. Immediate intervention required.

API Endpoints

Fairness Signals

Method
Endpoint
Description

POST

/model-fairness-signals

Ingest external fairness signal

GET

/model-fairness-signals

List signals with filters

GET

/model-fairness-signals/:id

Get single signal

Amplification Alerts

Method
Endpoint
Description

GET

/bias/amplification

List amplification alerts

GET

/bias/amplification/:id

Get single alert

PATCH

/bias/amplification/:id

Update alert status

POST

/bias/amplification/analyze/:signalId

Trigger analysis

Webhook Events

Two webhook events are available for external bias integration.

model.bias.ingested

Fired when a fairness signal is successfully ingested.

bias.amplification

Fired when amplification risk reaches high or above.

SDK Support

JavaScript

Python

Regulatory Mapping

Regulation
Article
How External Bias Integration Helps

EU AI Act

Art. 10

Documents external bias audits

EU AI Act

Art. 14

Detects human oversight failures

Colorado AI Act

Section 1702

Subgroup impact analysis

Texas TRAIGA

Section 2(a)

Algorithmic accountability

Best Practices

  1. Run external tools regularly. Execute bias analysis daily or after model retraining.

  2. Include confidence intervals. This helps ATHENA assess statistical significance.

  3. Use consistent model IDs. Match the identifiers in your ATHENA decision logs.

  4. Set appropriate thresholds. Define acceptable bias levels for your industry.

  5. Subscribe to webhooks. Enable real time amplification alerts.

Next Steps

Last updated