
Top Data→AI News
📞 What If AI Explained Its Decisions? The Sports Analytics Technique That Makes Credit Scoring Transparent

In my fairness research from detecting ChatGPT's occupational bias to exploring GiniMachine's adaptive threshold testing. I have focused on measuring bias: quantifying disparities, tracking metrics, showing which groups get disadvantaged.
But here's what's been bothering me: when we discover a credit model discriminates against certain demographics, we can measure the what (23-point gap for seniors) but we struggle with the why. Which features are driving the discrimination? Is it ZIP codes functioning as racial proxies? Income patterns correlating with age? Occupation types encoding gender?
Without understanding why a model makes discriminatory decisions, we're stuck with surface-level fixes: adjusting thresholds, rebalancing samples, hoping the bias disappears. We can't address root causes we can't see.
At the DataIntell Summit 2025, Joseph Jacob, Data Scientist, demonstrated how SHAP (SHapley Additive exPlanations) a technique he used for football performance analytics transfers directly to fintech, making black-box credit models fully transparent and revealing exactly which features contribute to biased predictions.
Key Highlights:
🔍 SHAP Quantifies Every Feature's Contribution: Instead of vague "the model considers multiple factors," SHAP assigns exact contribution values to each input feature for every prediction. Example: Credit card utilization: +0.28 (risk-increasing), Stable income 3+ years: -0.22 (risk-decreasing), Recent missed payment: +0.15. This transforms "Medium Risk score" from mysterious output to fully traceable calculation showing why this specific customer received this specific score.
⚽ Proven Transfer from Sports to Finance: Joseph originally used SHAP to identify which metrics contributed to successful team performance in football, where Logistic Regression delivered 72.34% accuracy but gave only global insight, not why a specific team won. The same problem exists in credit scoring: models might be accurate overall but can't explain individual rejections. SHAP solved both problems identically different industry, same explainability challenge, same solution.
🚨 Bias Detection Through Geographic Proxies: The Brazil credit default study Joseph presented revealed something critical: CEP-3 (postal code) SHAP values varied significantly by neighborhood, and these neighborhoods correlated strongly with census data on self-declared race. This proves geographically defined features can carry racial information as proxies, making models biased even when they never see race as a direct input. SHAP made the hidden pattern visible.
Why It Matters:
From Black Box to Transparent Decisions
85% of UK financial services firms now use ML (up from 72% in 2022), with data bias and lack of explainability as the biggest risks. These problems are connected: you can't fix bias you can't explain. SHAP breaks this cycle by providing exact feature contributions for every prediction.
When a customer's loan gets rejected, instead of "our algorithm determined you're high risk," lenders can explain: "Your credit was declined because high credit card utilization (+28% risk) and recent missed payment (+15% risk) outweighed your stable income (-22%)." This is actionable feedback customers can use to improve their position.
Regulatory Compliance and Bias Detection
GDPR's "right to explanation" and the EU AI Act demand explainable high-risk decisions. SHAP works with any model architecture generating consistent, human-interpretable explanations that satisfy regulatory requirements while maintaining model sophistication.
Critically, SHAP reveals bias sources, not just symptoms. The Brazil study showed ZIP codes carrying racial information as proxies. SHAP quantified exactly how much each neighborhood contributed to discrimination. When GiniMachine detects seniors facing 56% Disparate Impact, SHAP answers which features drive this disparity, enabling surgical fixes rather than full retraining.
The Trust Question
Joseph posed it perfectly: "If a model is accurate but not explainable, can organizations truly trust it?" Fair lending violations show the cost of unexplainable decisions: investigations, penalties, reputation damage. SHAP transforms this from risk to competitive advantage and a sophisticated models with full transparency.
Paper: Read More