EC launched a pilot phase to ensure that ethical guidelines for the development and use of artificial intelligence, or AI, can be implemented in practice. EC is taking a three-step approach, which involves setting out the key requirements for trustworthy artificial intelligence, launching a large scale pilot phase for feedback from stakeholders, and working on international consensus building for human-centric artificial intelligence. EC also presented the next steps for building trust in artificial intelligence by taking forward the work of the High-Level Expert Group, which was appointed in June 2018.
As per the EC approach, trustworthy artificial intelligence should respect all applicable laws and regulations as well as a series of requirements; specific assessment lists aim to help verify the application of each of the key requirements:
- Human agency and oversight. Artificial intelligence systems should enable equitable societies by supporting human agency and fundamental rights and not decrease, limit, or misguide human autonomy.
- Robustness and safety. Trustworthy artificial intelligence requires algorithms to be secure, reliable, and robust enough to deal with errors or inconsistencies during all life cycle phases of artificial intelligence systems.
- Privacy and data governance. Citizens should have full control over their own data, while data concerning them will not be used to harm or discriminate against them.
- Transparency. The traceability of artificial intelligence systems should be ensured.
- Diversity, non-discrimination, and fairness. Artificial intelligence systems should consider the whole range of human abilities, skills, and requirements and ensure accessibility.
- Societal and environmental well-being. Artificial intelligence systems should be used to enhance positive social change and enhance sustainability and ecological responsibility.
- Accountability. Mechanisms should be put in place to ensure responsibility and accountability for artificial intelligence systems and their outcomes.
In terms of building international consensus for human-centric artificial intelligence, EC wants to bring this approach to artificial intelligence ethics to the global stage because technologies, data, and algorithms know no borders. To this end, EC will strengthen cooperation with like-minded partners such as Japan, Canada, or Singapore and continue to play an active role in international discussions and initiatives including the G7 and G20. The pilot phase will also involve companies from other countries and international organizations. EC is inviting industry, research institutes, and public authorities to test the detailed evaluation list drawn up by the High Level Expert Group, which complements the guidelines.
Following the pilot phase, in early 2020, the Artificial Intelligence Expert Group review the assessment lists for the key requirements, building on the feedback received. Building on this review, EC will evaluate the outcome and propose any next steps. Furthermore, to ensure the ethical development of artificial intelligence, EC will, by the Autumn 2019, launch a set of networks of artificial intelligence research excellence centers, begin setting up networks of digital innovation hubs, and—together with member states and stakeholders—start discussions to develop and implement a model for data-sharing and making best use of common data spaces.
Keywords: Europe, EU, Banking, Insurance, Securities, PMI, Regtech, Artificial Intelligence, Guidelines, EC
Previous ArticleRBNZ Consults on Framework for Identifying D-SIBs
EBA issued a revised list of validation rules with respect to the implementing technical standards on supervisory reporting.
EBA published its response to the call for advice of EC on ways to strengthen the EU legal framework on anti-money laundering and countering the financing of terrorism (AML/CFT).
NGFS published a paper on the overview of environmental risk analysis by financial institutions and an occasional paper on the case studies on environmental risk analysis methodologies.
MAS published the guidelines on individual accountability and conduct at financial institutions.
APRA published final versions of the prudential standard APS 220 on credit quality and the reporting standard ARS 923.2 on repayment deferrals.
SRB published two articles, with one article discussing the framework in place to safeguard financial stability amid crisis and the other article outlining the path to a harmonized and predictable liquidation regime.
FSB hosted a virtual workshop as part of the consultation process for its evaluation of the too-big-to-fail reforms.
ECB updated the list of supervised entities in EU, with the number of significant supervised entities being 115.
OSFI published the key findings of a study on third-party risk management.
FSB is extending the implementation timeline, by one year, for the minimum haircut standards for non-centrally cleared securities financing transactions or SFTs.