But, while these technologies could help big tech gain an advantage in countering the ever evolving threat from cybercriminals, detecting fraud, and automating processes like loan applications and credit checks, AI ethics could also become one of its biggest challenges. Much like deeply flawed police profiling tools, biased AI algorithms can also skew results. For instance, in 2019, the algorithm used by Apple Card was found to be biased against women after determining the “creditworthiness” of applicants. But it’s not just bias, highly sensitive financial data will add a new dimension to the data collection vs user privacy debate already ongoing within the EU. In fact, in June of this year the EU Court of Justice backed a ruling that will give national privacy watchdogs more room to scrutinize big tech across the bloc. Having long faced heavy regulation, what can traditional banks teach these new players about building AI capabilities that are fair, ethical, and regulator approved?
It takes more than lip service to tackle bias in AI
Cybercriminals are continuing to evolve their techniques and their attack methods. While AI is becoming increasingly important in our day-to-day lives, such as detecting transactions that differ from normal customer behavior, there are still challenges. “Bias is one of the new challenges of AI,” says Mark Wiggerman, a data scientist at ABN AMRO’s corporate information security office. The functionality of the decision system is determined by the training data Wiggerman explains, adding that a model will often have to be retrained, which, in turn, could also lead to new functionality and possibly new biases. An example (fictitious) of a negative impact could be a fraud detection system that (unintentionally) puts transactions from young people on a queue for manual inspection. The negative impact is that their transactions are processed a few hours later, and the negative impact is structural for that group. A lot of biased AI systems have been reactive rather than proactive. The key to tackling ethical AI issues is to. not only monitor your systems regularly, but also to have clear guidelines on what actually constitutes ethical AI for your organization. Wiggerman shared: Like other companies, ABN AMRO is busy assessing AI and the potential it can deliver to its customers and staff. At the same time it’s developing policies to ensure fairness in AI applications.
Gathering personal data
Because of the favored position that big tech companies have, they’re able to leverage the data they already have on consumers. Yet, it’s this collection of data that brings into question consumer privacy, particularly when it comes to financial services. Wiggerman questions whether it’s okay for big tech to process the type of data they’re gathering. Only collecting necessary data (which is part of the GDPR by the way), is something that big tech needs to take into account. Bias shouldn’t be the only focus, but it should be part of a wider, responsible AI proposition which takes into account different countries’ regulations. For instance, regulations in the US are different from those in Europe, although according to Wiggerman the US is also introducing more GDPR-like rulings than before. Privacy, ethics, and fairness should be added to the product development life cycle from the beginning. In other words, fairness and privacy by design. This isn’t something that should be added afterwards. Consider security by design. Security measures shouldn’t be added afterwards, but taken into account from the beginning. AI ethics needs to trickle down into all the risk policies of any company.
Working together
Although big tech is starting to offer more financial services, one of the drawbacks is that they don’t have the historical financial data that the traditional banking sector enjoys. They also lack experience with financial processing. Wiggerman suggests that big tech organizations work with banks to learn from their financial history, as well as understand how they comply with GDPR models or audit requirements. When it comes to reconciling security and customer privacy, Wiggerman is of the view that the two can come together. “Definitely,” he says. In March, it was reported that ABN AMBRO was collaborating with Dutch scientific research organisation TNO and Dutch financial services company Rabobank to work on an MPC project designed to share and analyze data to detect financial crime. If you have two companies that have data they can’t share, like banks, MPC lets them share the data in an encrypted way so that the receiver can’t understand the data. However, MPC allows you to do a joint computation on all of the [encrypted] data of the three or four banks that want to collaborate. The outcome of this computation is usable for each bank but the raw data remains secret. In this way it allows you to maintain user privacy. ABN AMRO doesn’t publish any of its anti-crime efforts; however, Wiggerman explains that there is a strong collaboration between banks when it comes to cybersecurity. When something happens, information is shared between banking services rather than treating security as a competitive area. This is where big tech companies may differ. It should come as no surprise then that these same companies are turning to financial services, which is swiftly changing the banking sector. But in order for big tech to truly enter the industry, a collaboration between the two may be the way forward.