Reduce AI risk and promote AI trust | Deloitte US has been saved
By Ryan Hittner, Audit & Assurance Principal, Deloitte & Touche LLP
Thanks to the recently developed large language models, or LLMs, we may be close to realizing a decades-old quest: the creation of intelligent machines. This technology, which seeks to imitate the human brain, has opened a new realm called Generative AI—software that can write coherent text and create images and computer code at a level close to that of humans.
As we ponder the vast business potential of AI, one thing is abundantly clear: AI requires sound human decision-making to curb its potential shortcomings.
So how do we harness the power of AI while managing its risks in a manner that engenders trust and not suspicion?
Regulation and transparency are likely to be part of the answer. For example, the city council of New York City recently passed a law requiring bias audits of automated employment decision tools (AEDT) used by companies for employment decisions (e.g., select job candidates from pools of applicants). Moreover, Local Law 144-21 mandates that companies publish a summary of the audit results and that they alert job applicants or employees to the use of AEDTs as well as to the “categories” and “screens” set in evaluating qualifications and background.
Federal, state, and local governments/agencies as well as international bodies are taking a closer look at how organizations use AI. For example, officials in in several states have proposed AI legislation and regulations to enhance transparency and accountability, while the National Institute of Standards and Technology has released an AI Risk Management Framework. In Europe, the Artificial Intelligence Act seeks to provide a framework to regulate AI based on the level of risk that a system might pose and to harmonize cross-border rules dealing with this technology.
This activity suggests that it’s time for companies to focus on developing or refreshing their own AI strategies as well as usage and governance models—because there’s more to concerns over AI than just bias. It’s a matter of trust—the trust between stakeholders (e.g., shareholders, employees, etc.) and companies.
AI presents new and growing challenges to data privacy, risk mitigation (including reputational risk), cybersecurity, and disruptions of operations and business models. And those may increase as the technology and use cases expand. So, what should be on a company’s AI transparency and trust agenda as it works to reimagine and manage risk mitigation, operational controls, and governance processes?
Expectations around AI continue to grow, and companies are increasingly harnessing AI and expanding its impact. Just think, AI is already capable of writing this post with some human assistance. Next time it just might. That’s a reality worth pondering.
Please read our Wall Street Journal article to learn more, and if you have any questions, please contact me.
Ryan is an Audit & Assurance principal with more than 15 years of management consulting experience, specializing in strategic advisory to global financial institutions focusing on banking and capital markets. Ryan co-leads Deloitte's Artificial Intelligence & Algorithmic practice which is dedicated to advising clients in developing and deploying responsible AI including risk frameworks, governance, and controls related to Artificial Intelligence (“AI”) and advanced algorithms. Ryan also serves as deputy leader of Deloitte's Valuation & Analytics practice, a global network of seasoned industry professionals with experience encompassing a wide range of traded financial instruments, data analytics and modeling. In his role, Ryan leads Deloitte's Omnia DNAV Derivatives technologies, which incorporate automation, machine learning, and large datasets. Ryan previously served as a leader in Deloitte’s Model Risk Management (“MRM”) practice and has extensive experience providing a wide range of model risk management services to financial services institutions, including model development, model validation, technology, and quantitative risk management. He specializes in quantitative advisory focusing on various asset class and risk domains such as AI and algorithmic risk, model risk management, liquidity risk, interest rate risk, market risk and credit risk. He serves his clients as a trusted service provider to the CEO, CFO, and CRO in solving problems related to risk management and financial risk management issues. Additionally, Ryan has worked with several of the top 10 US financial institutions leading quantitative teams that address complex risk management programs, typically involving process reengineering. Ryan also leads Deloitte’s initiatives focusing on ModelOps and cloud-based solutions, driving automation and efficiency within the model / algorithm lifecycle. Ryan received a BA in Computer Science and a BA in Mathematics & Economics from Lafayette College. Media highlights and perspectives First Bias Audit Law Starts to Set Stage for Trustworthy AI, August 11, 2023 – In this article, Ryan was interviewed by the Wall Street Journal, Risk and Compliance Journal about the New York City Law 144-21 that went into effect on July 5, 2023. Perspective on New York City local law 144-21 and preparation for bias audits, June 2023 – In this article, Ryan and other contributors share the new rules that are coming for use of AI and other algorithms for hiring and other employment decisions in New York City. Road to Next, June 13, 2023 – In the June edition, Ryan sat down with Pitchbook to discuss the current state of AI in business and the factors shaping the next wave of workforce innovation.