No Comments

Financial Industry Must Tackle Gender Bias In Algorithms, According To Global Fintech Leader, Finastra

Finastra, one of the world’s largest fintechs, is calling upon the global finance industry to tackle algorithmic bias which is likely to be impacting millions of people every day.The fintech firm,which supplies vital technology to financial institutions of all sizes, including 90 of the world’s top100 banks, recently commissioned consultancy firm KPMG to look at the issue across banking, lending and insurance. The research considered how decisions coming from this advanced technology have the potential to impact outcomes for certain people and groups. In response to the findings, Finastra has published a five-point plan to identify and tackle algorithmic bias and is urging the financial industry to come together to take action and build a fairer society.

In the past decade, the financial world has been industrialized and digitalized through the introduction of artificial intelligence (AI), particularly forms of machine learning, boosting efficiencies and automating processes, resulting in many parts of banking, lending and insurance decision-making processes now being made by algorithms. The pandemic has accelerated the use of these technologies and whilst it brings clear positives, these vital algorithms can only be as ‘fair’ and unbiased as the datasets that are used to build them. The industry must check if the biases that exist in society are being repeated through the design and deployment of these technologies.

To understand the severity of the problem, Finastracommissioned KPMG to produce a report which reveals the sheer size of consumer lending markets and the potential impact of algorithmic bias. For example,in 2020, consumer lending and transactions across key financial products (credit card, other consumer product lending and mortgage/home lending) were over:

  • $6,110bn in the U.S.
  • HK$1,270bn in Hong Kong
  • £440bn in the United Kingdom
  • €280bnin France
  • SG$110bn in Singapore

Both the provision and costs – e.g. the interest rates charged – to consumers of this credit will be informed in many cases by the algorithms that are used.

Simon Paris, CEO at Finastra, said, “Without this being a priority in the financial industry, AI will become a flywheel that will accelerate the negative impact on human lives.Finastra doesn’t have all the answersbut we believe that the industry must first acknowledge that there is a problem with algorithmic bias – only then can we work together to find a solution. We will work with our partners and ecosystem to drive the change the industry needs to make – collectively and collaboratively we can redefine finance for good and open it up to all.Finastra’s goal is to ensure financial technology is benevolent and fair in every way to give everyone a level playing field when it comes to borrowing money.”

Dr Leanne Allen, Director at KPMG Financial Services Tech Consulting, said, “Consumer and public trust are critical success factors for Financial Services. The findings in our report for Finastra make it clear that providers need to take care when designing, building and implementing these algorithms to ensure innovation can continue to advance in a safe and ethical way. The report brings together recent thinking on algorithmic bias, with specific applications to financial services and the potential for biased decision-making. Mitigating bias is vitally important in our digital and data-led world. Not doing so could run the risk of serious financial harm to the consumers who use these services.”

To show its commitment to tackling this problem, Finastra has published a five-point plan as part of its drive to help redefine finance for good.

1.) Reforming Finastra’s developer agreement: Finastra has updated its developer terms and conditions for FusionFabric.cloud, its open platform and marketplace for developers. This means developers and partners will be expected to account for algorithmic bias and Finastra has the right to inspect for this bias within any new application

2.) Creating new proof of concept technologies: such as FinEqual, a digital tool that enables bias-free lending, to give users the technology to empower them to tackle algorithmic bias within their own businesses. Currently at proof-of-concept stage, Finastra aims to make FinEqualavailable to customers in the next 18 months

3.) Hacking for good: Finastra commits to all future hacks having a focus on inclusion. To support this, Finastra will be launching a global hacking competition as part of its Hack to the Future series to shine a light on female talent in the industry by finding and celebrating balanced, female-led teams pushing the boundaries of AI and machine learning

4.) Workplace equality:Within the organization, Finastra is continuing its journey to 50:50 male to female ratios across all its teams. This includes increasing women amongst our top 200 leaders and engineers from 30% to 40% by 2025 and to 50% by 2030

5.)Work with regulators: Finastra is fully committed to tackling AI bias. The company is working closely with regulators in multiple markets and, as a technology leader, is calling upon the financial services industry to take note of the threat algorithmic bias poses to society

The full report by KPMG can be sent directly upon request or found at the Finastra website.

Join the conversation ahead of International Women’s Day using #OpenByDefault #ChooseToChallenge

#OpenByDefault – Finastra is working hard to create a future where finance is open for all. Starting with International Women’s Day, the ‘Open By Default’ movement will be reflected across the hosted events, branded content and marketing strategies run by Finastra. 

Cyber Gear Webinar Series