Title
|
|
|
|
AIMLAI : Advances in Interpretable Machine Learning and Artificial Intelligence
| |
Author
|
|
|
|
| |
Abstract
|
|
|
|
Recent technological advances rely on accurate decision support systems that can be perceived as black boxes due to their overwhelming complexity. This lack of transparency can lead to technical, ethical, legal, and trust issues. For example, if the control module of a self-driving car failed at detecting a pedestrian, it becomes crucial to know why the system erred. In some other cases, the decision system may reflect unacceptable biases that can generate distrust. The General Data Protection Regulation (GDPR), approved by the European Parliament in 2018, suggests that individuals should be able to obtain explanations of the decisions made from their data by automated processing, and to challenge those decisions. All these reasons have given rise to the domain of interpretable and explainable AI. AIMLAI aims at gathering researchers, experts and professionals, from inside and outside the domain of AI, interested in the topic of interpretable ML and interpretable AI. The workshop encourages interdisciplinary collaborations, with particular emphasis in knowledge management, Infovis, human computer interaction and psychology. It also welcomes applied research for use cases where interpretability matters. AIMLAI envisions to become a discussion venue for the advent of novel interpretable algorithms and explainability modules that mediate the communication between complex ML/AI systems and users. |
| |
Language
|
|
|
|
English
| |
Source (book)
|
|
|
|
31st ACM International Conference on Information and Knowledge, Management (CIKM), OCT 17-21, 2022, Atlanta, GA
| |
Publication
|
|
|
|
New york
:
Assoc computing machinery
,
2022
| |
ISBN
|
|
|
|
978-1-4503-9236-5
| |
DOI
|
|
|
|
10.1145/3511808.3557491
| |
Volume/pages
|
|
|
|
(2022)
, p. 5160
| |
ISI
|
|
|
|
001074639605047
| |
Full text (Publisher's DOI)
|
|
|
|
| |
Full text (publisher's version - intranet only)
|
|
|
|
| |
|