When Improving AI Means Admitting You Caused Harm
Companies ship model updates every day. What they rarely ship is honesty, the courage to say, “We got the last version wrong.” In high-stakes systems such as credit, hiring, and housing, that silence is a strategy.
In 2019, a tech founder said his Apple card limit was far higher than his wife’s despite her stronger finances, prompting New York regulators to investigate whether the underwriting was discriminatory. Months later, the state’s financial regulator reported no fair-lending violations but also flagged transparency gaps and customer-service missteps that fueled public distrust.
The episode showed how a single complaint can surface a complex mix of model opacity and consumer harm, even when a legal finding of discrimination isn’t made.
The Polished Rewrite
Hiring has its version of the quiet fix. In 2021, HireVue dropped its facial-analysis scoring from video interviews after sustained criticism and regulatory pressure. The company framed it as a product evolution and continued using other AI features. Whatever you think of the science, the pattern is the same: change the system but avoid an admission.
Why the careful language? Because an improved Version 2 can become circumstantial evidence that Version 1 produced worse outcomes. Legal departments weigh discovery risk. If you document suspected bias, test thoroughly, and then overhaul the model, adversaries may argue you knew earlier results were inequitable. That risk calculus can slow audits, muffle change logs, and keep fixes narrow.
The Cost of Waiting
Meanwhile, communities bear the delays. Small-business credit remains uneven by race, with Black-owned firms reporting higher denial rates and tougher credit experiences in national surveys. These are structural patterns that predate modern AI, but algorithms trained on biased data can mirror and magnify them unless actively constrained.
Academic work on lending shows how tricky the line is. In mortgages, one large study found significant pricing disparities for Black and Hispanic borrowers across the market; algorithmic lenders reduced but did not eliminate those gaps. This is evidence that data-driven systems can help and still encode inequity without explicit guardrails. That is precisely why plain-English transparency and accountable remediation matter.
What Better Looks Like
So, what would a healthier pattern look like?
- Safe harbours for proactive fixes.
Give institutions limited protection when they disclose and correct disparate impacts, so fixing bias is legally safer than sitting on it.
- Mandatory change logs for high-stakes models.
When a credit or hiring model changes, publish what changed, why, and who might be affected in language a customer can read.
- Retroactive review where impacts surface.
If Version 2 shows Version 1 likely disadvantaged a group, run a look-back and invite affected applicants to be reassessed.
Saying It Out Loud
We don’t need to villainise every team that ships an enhancement. We need rules that reward naming the problem out loud. Otherwise, the most reliable feature of our systems will not be fairness or accuracy. It will be silence.


