Advances in Explainable Artificial Intelligence, 2nd Edition

A special issue of Information (ISSN 2078-2489). This special issue belongs to the section "Artificial Intelligence".

Deadline for manuscript submissions: 10 October 2024 | Viewed by 1413

Special Issue Editors


E-Mail Website
Guest Editor
Department of Computer Science, Università degli Studi di Milano, 20122 Milano, MI, Italy
Interests: machine learning; computational intelligence; game theory applications to machine learning and networking
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Department of Computer Science and Information Technology of INSA Lyon, LIRIS laboratory, 69100 Villeurbanne, France
Interests: machine learning; semantic web; information retrieval
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

Machine Learning (ML)-based Artificial Intelligence (AI) algorithms can learn from known examples of various abstract representations and models that, once applied to unknown examples, can perform classification, regression, or forecasting tasks, to name a few.

Very often, these highly effective ML representations are difficult to understand; this holds true particularly for deep learning models, which can involve millions of parameters. However, for many applications, it is of utmost importance for stakeholders to understand the decisions made by the system, in order to use them better. Furthermore, for decisions that affect an individual, future legislation might even advocate for a “right to an explanation”. Overall, improving the algorithms’ explainability may foster trust and social acceptance of AI.

The need to make ML algorithms more transparent and more explainable has generated several lines of research that form an area known as explainable Artificial Intelligence (XAI).

Among the goals of XAI are adding transparency to ML models by providing detailed information about why the system has reached a particular decision; designing more explainable and transparent ML models, while at the same time maintaining high performance levels; finding a way to evaluate the overall explainability and transparency of the models; and quantifying their effectiveness for different stakeholders.

The objective of this Special Issue is to explore recent advances and techniques in the XAI area.

Research topics of interest include (but are not limited to):

- Devising machine learning models that are transparent by design;

- Planning for transparency, from data collection to training, testing, and production;

- Developing algorithms and user interfaces for explainability;

- Identifying and mitigating biases in data collection;

- Performing black-box model auditing and explanation;

- Detecting data bias and algorithmic bias;

- Learning causal relationships;

- Integrating social and ethical aspects of explainability;

- Integrating explainability into existing AI systems;

- Designing new explanation modalities;

- Exploring theoretical aspects of explanation and interpretability;

- Investigating the use of XAI in application sectors such as healthcare, bioinformatics, multimedia, linguistics, human–computer interaction, machine translation, autonomous vehicles, risk assessment, and justice.

Prof. Dr. Gabriele Gianini
Dr. Pierre-Edouard Portier
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Information is an international peer-reviewed open access monthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1600 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • machine learning
  • deep learning
  • explainability
  • transparency
  • accountability

Related Special Issue

Published Papers (2 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

21 pages, 2251 KiB  
Article
Predicting Individual Well-Being in Teamwork Contexts Based on Speech Features
by Tobias Zeulner, Gerhard Johann Hagerer, Moritz Müller, Ignacio Vazquez and Peter A. Gloor
Information 2024, 15(4), 217; https://doi.org/10.3390/info15040217 - 12 Apr 2024
Viewed by 579
Abstract
Current methods for assessing individual well-being in team collaboration at the workplace often rely on manually collected surveys. This limits continuous real-world data collection and proactive measures to improve team member workplace satisfaction. We propose a method to automatically derive social signals related [...] Read more.
Current methods for assessing individual well-being in team collaboration at the workplace often rely on manually collected surveys. This limits continuous real-world data collection and proactive measures to improve team member workplace satisfaction. We propose a method to automatically derive social signals related to individual well-being in team collaboration from raw audio and video data collected in teamwork contexts. The goal was to develop computational methods and measurements to facilitate the mirroring of individuals’ well-being to themselves. We focus on how speech behavior is perceived by team members to improve their well-being. Our main contribution is the assembly of an integrated toolchain to perform multi-modal extraction of robust speech features in noisy field settings and to explore which features are predictors of self-reported satisfaction scores. We applied the toolchain to a case study, where we collected videos of 20 teams with 56 participants collaborating over a four-day period in a team project in an educational environment. Our audiovisual speaker diarization extracted individual speech features from a noisy environment. As the dependent variable, team members filled out a daily PERMA (positive emotion, engagement, relationships, meaning, and accomplishment) survey. These well-being scores were predicted using speech features extracted from the videos using machine learning. The results suggest that the proposed toolchain was able to automatically predict individual well-being in teams, leading to better teamwork and happier team members. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence, 2nd Edition)
Show Figures

Figure 1

14 pages, 15148 KiB  
Article
Explainable Machine Learning Method for Aesthetic Prediction of Doors and Home Designs
by Jean-Sébastien Dessureault, Félix Clément, Seydou Ba, François Meunier and Daniel Massicotte
Information 2024, 15(4), 203; https://doi.org/10.3390/info15040203 - 05 Apr 2024
Viewed by 521
Abstract
The field of interior home design has witnessed a growing utilization of machine learning. However, the subjective nature of aesthetics poses a significant challenge due to its variability among individuals and cultures. This paper proposes an applied machine learning method to enhance manufactured [...] Read more.
The field of interior home design has witnessed a growing utilization of machine learning. However, the subjective nature of aesthetics poses a significant challenge due to its variability among individuals and cultures. This paper proposes an applied machine learning method to enhance manufactured custom doors in a proper and aesthetic home design environment. Since there are millions of possible custom door models based on door types, wood species, dyeing, paint, and glass types, it is impossible to foresee a home design model fitting every custom door. To generate the classification data, a home design expert has to label thousands of door/home design combinations with the different colors and shades utilized in home designs. These data train a random forest classifier in a supervised learning context. The classifier predicts a home design according to a particular custom door. This method is applied in the following context: A web page displays a choice of doors to a customer. The customer selects the desired door properties, which are sent to a server that returns an aesthetic home design model for this door. This door configuration generates a series of images through the Unity 3D engine module, which are returned to the web client. The customer finally visualizes their door in an aesthetic home design context. The results show the random forest classifier’s good performance, with an accuracy level of 86.8%, in predicting suitable home design, marking the way for future developments requiring subjective evaluations. The results are also explained using a feature importance graphic, a decision tree, a confusion matrix, and text. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence, 2nd Edition)
Show Figures

Figure 1

Back to TopTop