Welcome to A.I. Lab
Research projects ARRS
1.10.2022 – 30.9.2024
/
Artificial Intelligence (AI) technology has achieved a number of breakthrough successes, particularly in the last 10 years, especially with the highly successful application-relevant machine learning methods. These methods are now successfully used in many application areas such as: image recognition, machine translation, robotics, automatic text and speech generation, medical diagnosis and decision making in healthcare, automated decision making, recommender systems, automatic discovery in science, new drug discovery, etc. However, some aspects of AI technology, such as deep neural networks, have aroused public suspicion. Some of the reasons for distrust in the use of AI include: the incomprehensibility of decisions made by AI; the appearance of bias or discrimination in decisions proposed by AI (e.g. in the judiciary, employment, healthcare); the tendency to manipulate people through social media; threats to democracy – the possibility of election manipulation; recommendation systems in social media that lead to the potentiation of harmful and hateful content across social media; the use of AI in killer autonomous weapons.
The above reservations about the use of AI have led to a number of initiatives focusing on the ethical aspects of AI and concerns about some uses of AI. These initiatives have contributed to the definition of a framework for trustworthy AI and principles that would pave the way for trust in the use of AI through regulations and laws. However, the general public’s attitude towards the use of AI often remains highly critical.
One of the main criticisms in this respect is described in the public domain by the phrase algorithmic bias. The following two examples are probably the most widely criticised in the public domain as examples of questionable applications of AI: (1) the COMPAS (Correctional Offender Management Profiling for Alternative Sanctions) system for assisting in judicial decision-making and (2) the Amazon job applicant assessment system. In both controversial cases, the main complaint is that the AI system suggests biased decisions. The alleged bias is related to discrimination on the basis of race or gender. These two cases and similar cases are publicly regarded as unacceptable uses of AI, often associated with phrases such as “algorithm bias”, “AI bias” and “machine learning bias”.
Another shortcoming of some AI systems, which is very often mentioned by the general public and also in political circles, is the incomprehensibility of the decisions of the AI system, or the lack of a humanly comprehensible explanation for these decisions.
In addition to these two problems, which lead to serious reservations about the acceptability of using AI, there are a number of other reservations, which are listed above. Some of these are also being professionally and intensively studied in the field of AI itself, and the negative public perception does not always seem to be fully justified. In this targeted research project, we will assess what the actual risks associated with the general concerns are and what the public perception of them is, both globally and in Slovenia. We will propose a framework for building trustworthy AI, including guidelines for topical amendments to legislation where we think they are needed. We will also assess the effectiveness of technical innovations in AI methods aimed at consolidating trustworthy AI and addressing concerns. In doing so, we will also contribute some of our own innovations in terms of explainable AI and educating the general public about AI in everyday life.
We will address the following concerns and questions regarding the use of AI:
Project goal
C1: Identify and analyse the impact of different characteristics of an AI system such as technological characteristics (robustness, explainability, transparency, models used), development methodologies used, existence of and compliance with relevant standards, type of licensing (e.g. open source, proprietary software) that are relevant for individual trust in AI;
C2: Identification and analysis of the impact of different ethical principles and fundamental rights on individual trust in AI; analysis of how specific concerns and issues not identified above relate to the public’s perception of the use of AI in the light of these ethical principles.
C3: Identify and analyse the impact of cultural and societal characteristics such as economic stability, social vulnerability, access to education, adequate legislation to ensure legal certainty and predictability on individuals’ trust in AI;
C4: Identification and analysis of individual status characteristics such as education, employment, age, ICT knowledge and skills on individuals’ confidence in AI;
C5: Establish a framework for analysing and monitoring the state of public trust in AI based on the interdependence and influence of the factors identified above, i.e. the characteristics of AI solutions, ethical principles, social and cultural characteristics, and the characteristics of the individual’s status;
C6: Creation of a demonstration version of the Orange Machine Learning System specifically designed for general education on the opportunities and risks of machine learning.
Project funding:
The project has received funding by the Slovenian Research and Innovation Agency.
