Fair Lending & Compliance

Four ways we can ensure that AI-driven lending works for everyone

Teddy Flo
December 9, 2021

The financial services industry is undergoing a dramatic transition caused by the broader adoption of AI and machine learning to automate and improve dozens of core functions. Among the most sensitive parts is consumer lending, where the improved accuracy of AI/ML promises enormous statistical improvements over existing credit scores. As a result, AI-based lending can widen credit access for consumers and small business owners while yielding healthier portfolios for lenders.

Regulators and lawmakers have been closely watching this transition. In 2020, House Financial Services Committee Chairwoman Maxine Waters (D-CA) formed a dedicated AI task force to ensure that AI in financial services lives up to principles of “transparency, enforceability, privacy and security, and fairness and equity.”

The committee asserted that AI must be used “in a way that serves the American public, its consumers, investors, and labor workforce, first and foremost.” The Committee has held two years of hearings that raised the right topics and concerns. Zest AI testified on AI’s potential for both good and harm at the first hearing in June 2019, where we stressed the importance of ensuring AI is fair and unbiased.

We can close the wealth gap measurably with a few minor changes to existing consumer financial services laws and regulations.

That’s why we applaud the recent letter from Chairwoman Waters and AI Task Force Chairman Rep. Bill Foster (D-IL) to the heads of the five financial regulatory agencies advancing and clarifying these principles for the rule writers and enforcement teams. The committee is rightly concerned about poorly designed algorithms that could amplify existing human biases or generate unforeseen discrimination. This risk is most acute in mortgage underwriting, where we don’t want “technological redlining” to perpetuate systemic racist practices.

Existing tools have left us in a place where America’s consumer lending industry still struggles to develop predictive models that minimize disparate impact, despite decades of focused effort by banks, legislators, and regulators.

Tens of millions of Americans lack sufficient credit history to compute a traditional credit score. Millions more have artificially depressed scores due to how credit is scored today. The Black homeownership rate is now back down to where it was in the Civil Rights Era. Why? Only a fifth of Black households has a credit score over 700, compared to half of all White households. As a result, Black applicants are denied mortgages at a rate 80 percent higher than White applicants. The share of Black families with a mortgage would increase nearly 11 percentage points if their credit score distribution were the same for White households.

AI-based lending offers us a once-in-a-generation opportunity to close the racial wealth gap. Technology now exists to modify underwriting models to reduce bias and disparate impact with minimal to no impact on accuracy. Dozens of lenders use Zest AI’s fair lending patented tools to generate underwriting models that are accurate and more inclusive. Zest AI's technology has lifted approval rates by 25 percent for low-income Black and Hispanic auto loan applicants.

AI models, to be sure, are more complex than simple credit scores and scorecards, but thanks to recent innovations by Zest AI and others, ML models are no more difficult to interpret than current scoring methods.

Built correctly and with care, AI models can be even more transparent than legacy scores, which have yet to undergo real public scrutiny of their disparate impact.

The Committee's letter urges regulators to “look under the hood of the algorithms underpinning products and services utilizing AI.” And they should.

But while they’re at it, they should also look under the hood of the most widely-used scoring models in consumer financial markets. Let’s make sure no player is excluding people from access to safe credit. As part of their new equitable housing finance plans, the FHFA, Fannie, and Freddie should test alternative solutions to FICO and generate evidence that no sufficiently performing LDA model exists to standard scores.

Former NCUA Chairman Rodney Hood correctly identified the lack of access to affordable, mainstream credit products as a civil rights issue. Real competition in credit scoring and rigorous fair lending testing for all players will help expand that credit access.

Here’s four policy additions we’d love to see as part of the conversation about closing the racial wealth gap via fairer access to credit:

  • Add language to the commentary to Regulation B, which implements the Equal Credit Opportunity Act (ECOA), to clarify how a lender can explain to a consumer a credit score assigned using ML models. The change would give consumers more accurate information when they are denied credit and allow lenders to use advanced credit scoring techniques that make lending more inclusive.
  • Inject real competition in credit scoring in the mortgage market. The validation requirement to be an “official” score kept everyone out except FICO. That lack of competition shuts millions of Americans out of the mortgage market. By lowering the cost of applying for validation, we’d have real competition in credit scoring that could put millions of Americans into new homes while improving the safety and soundness of Freddie Mac and Fannie Mae.
  • Improve how we estimate race and ethnicity in fair lending analysis. Congress should encourage regulators to study more advanced race estimation methods to make fair lending more accurate. As part of this effort, the Consumer Reporting Agencies should also be encouraged to make race and ethnicity data more readily available for the analysis required to produce fairer models.
  • Require regulators to study and adopt advanced analytical tools for broader use in fair lending analysis, which will spur wider private-sector adoption of better fairness tools. For example, fair lending enforcement has focused on problematic inputs to credit decisions (like the use of attributes like the number of criminal records) instead of focusing on problematic outcomes (like lower approval rates for people of color). Better tools can help shift the analysis to metrics that matter.

We can close the wealth gap measurably with a few minor changes to existing consumer financial services laws and regulations.

April 18, 2024
Redefining financial literacy through innovation and community
April 9, 2024
Learning from nature — you must water and prune a plant for it to grow
March 28, 2024
Innovation In Lending
Looking beyond market pain points to find purpose