BankThink

Banks embracing the AI future need to pay attention to its risks

AI banking benefits need to be offset by risk mitigation
The potential benefits of AI in the banking industry far outweigh the risks, especially when those risks are properly mitigated, writes Zenus Bank's Pedro Martinez.
Poca Wander Stock - stock.adobe.com

The ongoing rise of artificial intelligence (AI) has profoundly impacted the banking and fintech sectors, offering the potential for increased efficiency, improved customer service and enhanced risk management. However, as we ride this wave of technological transformation, important questions arise concerning security vulnerabilities, code ownership and application copyrights associated with AI-generated code. 

Historically, major U.S. banks had slow AI adoption, but this gained momentum in the mid-2010s. Initially, AI technologies were used for chatbots to enhance customer service. Gradually, they evolved to handle complex tasks like fraud detection, credit scoring and risk assessment. Today, financial institutions rely increasingly on AI platforms such as OpenAI's ChatGPT. However, AI adoption poses challenges. AI-generated code can introduce security vulnerabilities, as human oversight becomes harder. This could lead to large-scale data breaches, necessitating advanced code vetting systems and increased regulation protocols.

Code ownership also becomes a question. If AI helps write an application, who owns it? Current patent and copyright laws lack definitive guidance. Lawmakers need to address this issue as AI-generated code blurs the lines. The application of copyright laws to AI-generated code is equally unclear. AI lacks legal personhood, but as AI contributes significantly to content generation, it's uncertain whether works are products of AI or human authorship. This remains unresolved.

As AI permeates the banking industry, it also opens new avenues for malicious actors. The potential security threats range from subtle identity theft to major data breaches.

One major concern is the use of AI by cybercriminals for identity theft, particularly during the onboarding process. By leveraging deep-fake technology — a sophisticated application of AI that can manipulate or fabricate visual and audio content — fraudsters can now convincingly fake identities. For instance, in 2021, a man in New Jersey was charged for allegedly using AI to create synthetic identities and defraud banks out of hundreds of thousands of dollars.

AI-generated code vulnerabilities also pose a significant risk. AI, particularly machine learning models, are vulnerable to adversarial attacks that manipulate input data to deceive the model, leading to erroneous outputs. An adversary could, for example, exploit this vulnerability to trick a fraud detection system to approve fraudulent transactions.

There is also the threat of data poisoning, where an attacker inserts malicious data into the training set to bias the AI system's learning process. In a banking scenario, this could compromise a risk assessment model, leading to financial losses.

As a response, the industry must not only stay abreast of these threats but also invest in developing AI systems that are resilient to these types of attacks. Rigorous testing, ongoing monitoring and using state-of-the-art encryption and cybersecurity measures can help banks safeguard their AI systems against these threats.

The security threats posed by AI have direct and indirect economic and regulatory impacts on the banking industry. They can lead to significant financial losses, reputational damage, regulatory sanctions and increased cost of operations.

In 2022, a sophisticated AI-led cyberattack on a U.S. bank resulted in unauthorized transfers amounting to $1 million. This attack used a form of AI manipulation known as "model evasion," which modified transactional data in a way that evaded the bank's fraud detection system. The attack led to a significant financial loss and increased scrutiny from regulators.

The release of two malicious language models — WormGPT and FraudGPT — demonstrate attackers' evolving capability to harness language models for criminal activities.

August 8
Data And Text On Computer Screens

Another example is a 2023 case where a major credit card company suffered a large-scale data breach due to an undetected vulnerability in its AI-powered customer service chatbot. The breach exposed the personal data of millions of customers, resulting in class-action lawsuits, regulatory penalties and significant reputational damage.

The costs of these incidents extend beyond immediate financial losses. Banks face increased expenses in improving their cybersecurity measures, hiring skilled personnel and ensuring regulatory compliance. Following such incidents, banks typically experience higher customer churn rates and decreased shareholder confidence, both of which can impact long-term profitability.

On the regulatory front, such security breaches typically attract the attention of regulatory bodies like the Federal Reserve and the Office of the Comptroller of the Currency (OCC) in the U.S. These institutions may impose penalties, increase the rigor of regulatory examinations and may mandate more stringent risk management practices.

For instance, in response to the aforementioned credit card company data breach, the OCC increased its scrutiny of AI implementations across the banking sector and issued new guidelines on AI security. This regulatory action has increased the compliance burden on banks and other financial institutions, prompting them to reassess their AI adoption strategies and risk management practices.

Despite these challenges, it is critical to remember that AI is a tool that can be harnessed safely with adequate measures. For instance, to address security vulnerabilities, banking institutions can implement robust code review and testing practices, invest in advanced cybersecurity technologies and regularly update their security protocols. As for code ownership and copyright issues, clear guidelines and agreements prior to AI utilization can provide some measure of protection until formal legal precedents are set.

Looking at the larger picture, the potential benefits of AI in the banking industry far outweigh the risks, especially when those risks are properly mitigated. AI enables faster and more accurate decision-making, drastically reduces operational costs and significantly improves the customer experience. Moreover, AI's capability to learn and adapt can help banks stay ahead of the curve in the face of ever-evolving financial threats.

Taking everything into account, while the integration of AI in the banking and fintech industry brings undeniable benefits, it also ushers in a new set of challenges around security, code ownership and application copyrights. As we move forward, these issues must be addressed thoughtfully, with a balanced approach that recognizes AI's potential while maintaining a keen eye on associated risks. Ultimately, the key lies in navigating these challenges strategically to unleash the full potential of AI, fueling innovation and transforming the future of banking.

For reprint and licensing requests for this article, click here.
Bank technology Artificial intelligence Risk management Fintech
MORE FROM AMERICAN BANKER