Social Explainable Artificial Intelligence

Project: Research

Project Details

Description

The recent wave of machine-learning (ML) based artificial-intelligence (AI) technologies is having a huge societal and economic impact, with AI being (often silently) embedded in most of our everyday experiences (such as virtual assistants, tracking devices, social media, recommender systems). The research community (and society in general) has already realised that the current centralised approach to AI, whereby our personal data are centrally collected and processed through opaque ML (“black-boxes”), is not an acceptable and sustainable model in the long run. We posit that the “next wave” of ML-driven AI will be (i) human-centric, (ii) explainable, and (iii) more distributed and decentralised (i.e., not centrally controlled). On the one hand, these principles address the societal and ethical expectations for trustworthy AI, such as those recommended at the European Level (e.g., the AI HLEG report includes human agency, transparency, explainability). On the other hand, they fit a clear trend to develop decentralised ML for strictly technical reasons, e.g., performance, scalability, real-time constraints.

SAI will develop the scientific foundations for novel ML-based AI systems ensuring (i) individuation: each individual is associated with their own “Personal AI Valet” (PAIV), which acts as the individual’s proxy in a complex ecosystem of interacting PAIVs; (ii) personalisation: PAIVs process individuals’ data via explainable AI models tailored to the specific characteristics of their human twins; (iii) purposeful interaction: PAIVs interact with each other to build global AI models and/or come up with collective decisions starting from the local (i.e., individual) models; (iv) human-centricity: the interaction between PAIVs is driven by (quantifiable) models of the individual and social behaviour of their human users; (v) explainability: explainable ML techniques are extended through quantifiable human behavioural models and network science analysis to make the global AI models explainable-by-design. The ultimate goal of SAI is to provide the foundational elements enabling a decentralised collective of explainable PAIVs to evolve local and global AI models whose processes and decisions are transparent, explainable and tailored to the needs and constraints of individual users. We provide a concrete example of a SAI-enabled scenario in §1.1.

To this end, the project will deliver (i) the PAIV, a personal digital platform where every person can privately and safely integrate, store, and extract meaning from their own digital tracks, as well as connect with PAIVs of other users; (ii) human-centric local AI models; (iii) global, decentralised AI models, emerging from human-centric interactions between PAIVs; (iv) personalised explainability at the level of local and global AI models; and (v) concrete use cases to validate the SAI design principles based on real datasets, in the areas of private traffic management and opinion diffusion/fake news detection in social media.
AcronymSAI
StatusFinished
Effective start/end date1/03/2131/10/24

Collaborative partners

  • University of Pisa
  • University of Sheffield
  • Polish Academy of Science
  • University of Tartu
  • Sofia University
  • Consiglio Nazionale Delle Ricerche (lead)

Fingerprint

Explore the research topics touched on by this project. These labels are generated based on the underlying awards/grants. Together they form a unique fingerprint.