Featured Product

    Christopher Woolard of FCA on AI and Future of Regulation

    July 16, 2019

    Christopher Woolard of FCA spoke at a conference in London, focusing on the issue of whether decisions that materially affect lives of people can be outsourced to a machine and what this might mean for the future of regulation. He mentioned that, as a regulator, FCA considers the use of artificial intelligence, also commonly referred to as AI, in financial services from the three main perspectives of continuity, public value, and collaboration. He reinforced that regulators, academics, industry, and the public need to work together to develop a shared understanding that will determine the approach to be taken to answer the questions artificial intelligence poses in the years ahead. These developments involve "big and complex questions" that "go beyond the day-to-day operations of FCA and other regulators," with the key question being "how can we ensure the regulatory framework adapts to the changing economic, demographic, and political environment in which it operates?"

    Highlighting the results of a joint survey by FCA and BoE to assess the state of play on artificial intelligence, Mr. Woolard mentioned that the use of artificial intelligence in the firms it regulates is best described as nascent. The technology is employed largely for back-office functions, with customer-facing technology largely in the exploration stage. He also highlighted that the risks presented by artificial intelligence will be different in each of the contexts in which it is deployed. After all, the risks around algorithmic trading will be totally different from those that occur when artificial intelligence is used for credit ratings purposes or to determine the premium on an insurance product. FCA does not have one universal approach to harm across financial services because harm takes different forms in different markets and, therefore, has to be dealt with on a case-by-case basis; it will be the same with artificial intelligence too. He added, if firms are deploying artificial intelligence and machine learning, they need to ensure they have a solid understanding of the technology and the governance around it.

    Next, he highlighted the growing consensus around the idea that algorithmic decision-making needs to be "explainable." For example, if a mortgage or life insurance policy is denied to a consumer, the reasons for denial need to be explained. However, the question is what should be the extent and level of explainability (to a consumer or an informed expert) and what takes precedence—the accuracy of prediction or the abilit to explain it. The challenge is that explanations are not a natural by-product of complex machine learning algorithms. It is possible to "build in" an explanation by using a more interpretable algorithm, but this may dull the predictive edge of the technology. This is why FCA has partnered with The Alan Turing Institute to explore the transparency and explainability of artificial intelligence in the financial sector, said Mr. Woolard. Through this project, FCA wants to move the debate from the high-level discussion of principles (which most now agree on) toward a better understanding of the practical challenges on the ground that machine learning presents.

    With regard to collaboration, he emphasized that these innovations cannot be developed in isolation and the problems that artificial intelligence and machine learning have the potential to solve are cross-border, cross-sector, sometimes cross-agency. He also highlighted that, at an international level, FCA is leading a workstream on machine learning and artificial intelligence for IOSCO, exploring issues around trust and ethics and what a framework for financial services might look like. It is also looking inward and asking whether it can do anything differently as a regulator to ensure that it is ready for the challenges of the future. he added that, at a basic level, firms using this technology must keep one key question in mind, not just "is this legal?’"but "is this morally right?" Regulators have a range of powers and tools to tackle these issues now but, with the increasing use of technology, those tools may need to be updated for a fully digital age. "This is something we will be thinking about in our own work on the ‘Future of Regulation.'" FCA is taking a fundamental look at how it carries out conduct regulation and shapes the regulatory framework going forward, in what it calls its "Future of Regulation" project.

     

    Related Link: Speech

     

    Keywords: Europe, UK, Banking, Insurance, Securities, Artificial Intelligence, Machine Learning, Regtech, Future of Regulation, FCA

    Related Articles
    News

    APRA Publishes Approach to Regulating and Supervising GCRA Risks

    APRA published an information paper that sets out a more intensive regulatory approach to transform governance, culture, remuneration, and accountability (GCRA) practices across the prudentially regulated financial sector.

    November 19, 2019 WebPage Regulatory News
    News

    IAIS Publishes Application Paper on Recovery Planning

    IAIS published the final application paper on recovery planning, along with the resolution of comments on the draft application paper.

    November 18, 2019 WebPage Regulatory News
    News

    FSB Publishes Summary of November Meeting of RCG for MENA Region

    FSB published a summary of the November meeting of the Regional Consultative Group (RCG) for Middle East and North Africa (MENA).

    November 17, 2019 WebPage Regulatory News
    News

    EBA Single Rulebook Q&A: Second Update for November 2019

    EBA updated the Single Rulebook question and answer (Q&A) tool with answers to eight questions that relate to the Bank Resolution and Recovery Directive (BRRD) and the Capital Requirements Regulation and Directive (CRR and CRD).

    November 15, 2019 WebPage Regulatory News
    News

    FSI Examines Use of Red Team Testing to Enhance Cyber Resilience

    The Financial Stability Institute (FSI) of BIS published a paper that examines the contribution of red team testing frameworks toward enhancing cyber resilience.

    November 15, 2019 WebPage Regulatory News
    News

    FASB Delays Effective Dates for CECL, Leases, and Hedging Standards

    FASB issued two Accounting Standards Updates finalizing the delays in effective dates for standards on current expected credit losses (CECL), leases, hedging, and long-duration insurance contracts.

    November 15, 2019 WebPage Regulatory News
    News

    ESMA Updates Q&A on Securitization Regulation in November 2019

    ESMA updated questions and answers (Q&A) on the Securitization Regulation (Regulation 2017/2402).

    November 15, 2019 WebPage Regulatory News
    News

    HKMA Announces Finalization of Banking Liquidity Amendment Rules 2019

    HKMA issued a letter informing all authorized institutions that negative vetting of the Banking (Liquidity) (Amendment) Rules 2019 (BLAR) has now expired. Thus, the BLAR will now come into operation from January 01, 2020.

    November 15, 2019 WebPage Regulatory News
    News

    BCBS Consults on Revised Disclosures for Market Risk Framework

    BCBS launched a consultation on the revised disclosure requirements for the market risk framework for banks.

    November 14, 2019 WebPage Regulatory News
    News

    BCBS Consults on Disclosure Templates of Sovereign Exposures of Banks

    BCBS published a consultation on the voluntary disclosure templates related to sovereign exposures of banks.

    November 14, 2019 WebPage Regulatory News
    RESULTS 1 - 10 OF 4167