Antidote (PCI2020-120717-2) funded by MCIN/AEI /10.13039/501100011033 and by European Union NextGenerationEU/PRTR
(2021 - 2024)
ANTIDOTE fosters an integrated vision of explainable AI, where low level characteristics of the deep learning process are combined with higher level schemas proper of the human argumentation capacity.
Providing high quality explanations for AI predictions based on machine learning requires combining several interrelated aspects, including, among the others: selecting a proper level of generality/specificity of the explanation, considering assumptions about the familiarity of the explanation beneficiary with the AI task under consideration, referring to specific elements that have contributed to the decision, making use of additional knowledge (e.g. metadata) which might not be part of the prediction process, selecting appropriate examples, providing evidences supporting negative hypothesis, and the capacity to formulate the explanation in a clearly interpretable, and possibly convincing, way.
Accordingly, ANTIDOTE will exploit cross-disciplinary competences in three areas, i.e. deep learning, argumentation and interactivity, to support a broader and innovative view of explainable AI. Although we envision a general integrated approach to explainable AI, we will focus on a number of deep learning tasks in the medical domain, where the need for high quality explanations, both to clinicians and to patients, is perhaps more critical than in other domains.
Organization: CHIST-ERA - INT-Acciones de Programación Conjunta Internacional (MINECO) 2020
Main researcher: Rodrigo Agerri
Agustin Martinez-Ibargüen, Elisa Espina, Rodrigo Agerri, Aitziber Atutxa, Joseba Fernandez de Landa, Koldo Gojenola, Maite Oronoz, German Rigau