The 4th World Conference on Explainable Artificial Intelligence.

54 views
Skip to first unread message

Sylvio Barbon

unread,
Nov 6, 2025, 7:31:10 AM (4 days ago) Nov 6
to Machine Learning News
Dear colleagues
 
this is to invite you to submit research articles to the 4th World Conference on Explainable Artificial Intelligence.
 
_____________________________________________________________________
Artificial intelligence has undergone a significant shift in focus, with a growing emphasis on designing and developing intelligent systems that are both interpretable and explainable. This is due to the complexity of the models, built from data, and the legal requirements imposed by various national and international parliaments. This has been echoed both in the research literature and the press, attracting scholars worldwide and a lay audience. An emerging field in AI is Explainable Artificial Intelligence (xAI), which is devoted to producing intelligent systems that enable humans to understand their inferences, assessments, predictions, recommendations, and decisions. Initially devoted to designing post-hoc methods for explainability, eXplainable Artificial Intelligence (xAI) is rapidly expanding its boundaries to neuro-symbolic methods for producing self-interpretable models. Research has also shifted the focus to the structure of explanations and human-centred Artificial Intelligence, as the ultimate users of interactive technologies are humans.
 
Important dates 
Authors/title registration on submission platform – (easy-chair)* (it remains open until the paper submission deadline below):

January, 15, 2026

Article upload deadline on submission platform (easy-chair)*:

February 1, 2026

Paper bidding for reviewers

February 2-4, 2026

Review submission deadlines for reviewers

February 20, 2026

Notification of acceptance*:

February 22, 2026

Registration (payment) and camera-ready* (upload to easy-chair):

February 28, 2026

Article presentation instructions notification

June, 2026

Publication (Springer CCIS series)

September/October, 2026
 
The World Conference on Explainable Artificial Intelligence is an annual event that aims to bring together researchers, academics, and professionals, promoting the sharing and discussion of knowledge, new perspectives, experiences, and innovations in the field of Explainable Artificial Intelligence (xAI). This event is multidisciplinary and interdisciplinary, bringing together academics and scholars of different disciplines, including Computer Science, Psychology, Philosophy, Law and Social Science, to mention a few, and industry practitioners interested in the practical, social and ethical aspects of the explanation of the models emerging from the discipline of Artificial Intelligence (AI).
 
The conference organisation encourages submissions related to explainable AI and contributions from academia, industry, and other organisations discussing open challenges or novel research approaches related to the explainability and interpretability of AI systems. Topics include, and are not limited to:
 
Technical methods for XAI
Action Influence Graphs

Agent-based explainable systems

Ante-hoc approaches for interpretability

Argumentative-based approaches for xAI

Argumentation theory for xAI

Attention mechanisms for xAI

Automata for explaining RNN models

Auto-encoders & latent spaces explainability

Bayesian modelling for interpretability

Black-boxes vs white-boxes

Case-based explanations for AI systems

Causal inference & explanations

Constraints-based explanations

Decomposition of NNET-models for XAI

Deep learning & XAI methods

Defeasible reasoning for explainability

Evaluation approaches for XAI-based systems

Explainable methods for edge computing

Expert systems for explainability

Sample-centric and dataset-centric explanations

Explainability of signal processing methods

Finite state machines for explainability

Fuzzy systems & logic for explainability

Graph neural networks for explainability

Hybrid & transparent black box modelling

Interpreting & explaining CNN Networks

Interpretable representational learning

Explainability & the Semantic Web

Model-specific vs model-agnostic methods

Neuro-symbolic reasoning for XAI

Natural language processing for explanations

Ontologies & taxonomies for supporting XAI

Pruning methods with XAI

Post-hoc methods for explainability

Reinforcement learning for enhancing XAI

Reasoning under uncertainty for explanations

Rule-based XAI systems

Robotics & explainability

Sample-centric & Dataset-centric explanations

Self-explainable methods for XAI

Sentence embeddings to xAI semantic features

Transparent & explainable learning methods

User interfaces for explainability

Visual methods for representational learning

XAI Benchmarking

XAI methods for neuroimaging & neural signals

XAI & reservoir computing

 
Ethical Considerations for XAI
Accountability & responsibility in XAI

Addressing user-centric requirements for XAI

Trade-off model accuracy & interpretability

Explainable Bias & fairness of XAI systems

Explainability for discovering, improving, controlling & justifying

Moral Principles & dilemma for XAI

Explainability & data fusion

Explainability/responsibility in policy guidelines

Explainability pitfalls & dark patterns in XAI

Historical foundations of XAI

Moral principles & dilemma for XAI

Multimodal XAI approaches

Philosophical consideration of synthetic explanations

Prevention/detection of deceptive AI explanations

Social implications of synthetic explanations

Theoretical foundations of XAI

Trust & explainable AI

The logic of scientific explanation for/in AI

Expected epistemic & moral goods for XAI

XAI for fairness checking

XAI for time series-based approaches
Psychological Notions & concepts for XAI
Algorithmic transparency & actionability

Cognitive approaches for explanations

Cognitive relief in explanations

Contrastive nature of explanations

Comprehensibility vs interpretability

Counterfactual explanations

Designing new explanation styles

Explanations for correctability

Faithfulness & intelligibility of explanations

Interpretability vs traceability

explanations Interestingness & informativeness

Irrelevance of probabilities to explanations

Iterative dialogue explanations

Local vs. global interpretability & explainability

Local vs global interpretability & explainability

Methods for assessing explanations quality

Non-technical explanations in AI systems

Notions and metrics of/for explainability

Persuasiveness & robustness of explanations

Psychometrics of human explanations

Qualitative approaches for explainability

Questionnaires & surveys for explainability

Scrutability & diagnosis of XAI methods

Soundness & stability of XAI methods
Social examinations of XAI
Adaptive explainable systems

Backwards & forward-looking responsibility forms to XAI

Data provenance & explainability

Explainability for reputation

Epistemic and non-epistemic values for XAI

Human-centric explainable AI

Person-specific XAI systems

Presentation & personalization of AI explanations for target groups

Social nature of explanations
Legal & administrative considerations of/for XAI
Black-box model auditing & explanation

Explainability in regulatory compliance

Human rights for explanations in AI systems

Policy-based systems of explanations

The potential harm of explainability in AI

Trustworthiness of XAI for clinicians/patients

XAI methods for model governance

XAI in policy development

XAI for situational awareness/compliance behavior
Safety & security approaches for XAI
Adversarial attacks explanations

Explanations for risk assessment

Explainability of federated learning

Explainable IoT malware detection

Privacy & agency of explanations

XAI for Privacy-Preserving Systems

XAI techniques of stealing attack & defence

XAI for human-AI cooperation

XAI & models output confidence estimation
Applications of XAI-based systems
Application of XAI in cognitive computing

Dialogue systems for enhancing explainability

 

Explainable methods for medical diagnosis

Business & Marketing

XAI systems for healthcare

Explainable methods for HCI

Explainability in decision-support systems

Explainable recommender systems

Explainable methods for finance & automatic trading systems

Explainability in agricultural AI-based methods

Explainability in transportation systems

Explainability for unmanned aerial vehicles

Explainability in brain-computer interfaces

Interactive applications for XAI

Manufacturing chains & application of XAI

Models of explanations in criminology, cybersecurity & defence

XAI approaches in Industry 4.0

XAI systems for health-care

XAI technologies for autonomous driving

XAI methods for bioinformatics

XAI methods for linguistics/machine translation

XAI methods for neuroscience

XAI models & applications for IoT

XAI methods for XAI for terrestrial, atmospheric, & ocean remote sensing

XAI in sustainable finance & climate finance

XAI in bio-signals analysis
Submitted manuscripts must be novel and not substantially duplicate existing work. Manuscripts must be written using Springer’s Lecture Notes in Computer Science (LNCS) in the format provided here. Latex and word files are admitted; however, the former is preferred. All submissions and reviews will be handled electronically. The conference has a no dual submission policy, so submitted manuscripts must not be currently under review at another publication venue.

Articles must be submitted using the easy-chair platform here.

 
While registering on the platform, the contact author must provide the following information: paper title, all author names, affiliations, postal address, e-mail address, and at least three keywords.
 
The conference will not require a strict page number, as we believe authors have different writing styles and would like to produce scientific material differently. However, the following types of articles are admitted:
 

full articles

between 14 and 24 pages (including references)

short articles

between 10 and 14 pages (including references)
 

 

Full articles should report on original and substantial contributions of lasting value, and the work should concern the theory and/or practice of Explainable Artificial Intelligence (xAI). Moreover, manuscripts showcasing the innovative use of xAI methods, techniques, and approaches and exploring the benefits and challenges of applying xAI-based technology in real-life applications and contexts are welcome. Evaluations of proposed solutions and applications should be commensurate with the claims made in the article. Full articles should reflect more complex innovations or studies and have a more thorough discussion of related work. Research procedures and technical methods should be presented sufficiently to ensure scrutiny and reproducibility. We recognise that user data may be proprietary or confidential; therefore, we encourage sharing (anonymized, cleaned) data sets, data collection procedures, and code. Results and findings should be communicated clearly, and implications of the contributions for xAI as a field and beyond should be explicitly discussed.

 

Shorter articles should generally report on advances that can be described, set into context, and evaluated concisely. These articles are not ‘work-in-progress’ reports but complete studies focused on smaller but complete research work, simple to describe. For these articles, the discussion of related work and contextualisation in the wider body of knowledge can be smaller than that of full articles.
 
 
Appendixes and supplemental material
Appendices and supplemental material must be placed within the article and the maximum number of pages mentioned above.
 
Special session articles
The article submitted to the special sessions follows the submission procedure of the main track and must be submitted via EasyChair, as mentioned above. The types of articles admitted are full and shorter, as described above. The authors of an article to be associated with a special session must select the name of such special session in the list of topics in EasyChair, along with other relevant topics.
 
Authors commit to reviewing
By submitting to the conference, each senior manuscript author (holding at least a PhD) volunteers to be added to the pool of potential PC members/reviewers for the conference and may be asked to review manuscripts. This does not apply to authors who have already agreed to contribute to the conference in some capacity (e.g., as PC/SPC members of the main conference or special tracks, area chairs, or members of the organising committee) and authors who are not qualified to be on the programme committee.
 
Ethical & Human Subjects Considerations
The conference organisers expect authors to discuss the ethical considerations and the impact of the presented work and/or its intended application, where appropriate. Additionally, all authors must comply with the ethical standards and regulatory guidelines associated with human subjects research, including the use of personally identifiable data and research involving human participants. Manuscripts reporting on human subjects research must include a statement identifying any regulatory review the research is subject to (and identifying the form of approval provided) or explaining the lack of required review.
 
Submission and publication of multiple articles
 
Each author is limited to a combined maximum of 4 submissions to the main conference track, and authors may not be added or deleted from papers following submission.
 
Use of Generative AI
 
Generative AI models such as LLMs, including ChatGPT, BARD, LLaMA, and similar, are against the criteria for authorship of scientific manuscripts submitted and published in the conference. If authors use any of these tools while writing their manuscript, they assume full responsibility for all content. This includes verifying its correctness and assessing plagiarism of any part of their work. Suppose the text generated by the above generative AI models is the subject of scientific inquiry as part of the manuscript’s methodology or analysis. In that case, it must be adequately described, documented and made explicit in the paper.
 
 
Important dates
*All dates are Anywhere on Earth time (AoE)
 
Special session proposals
 

Proposal submission (contact):

November 15, 2025

Notification of acceptance & final instructions

November 31, 2025
Articles (main track & special tracks)
 

Authors/title registration on submission platform – (easy-chair)* (it remains open until the paper submission deadline below):

January, 15, 2026

Article upload deadline on submission platform (easy-chair)*:

February 1, 2026

Paper bidding for reviewers

February 2-4, 2026

Review submission deadlines for reviewers

February 20, 2026

Notification of acceptance*:

February 22, 2026

Registration (payment) and camera-ready* (upload to easy-chair):

February 28, 2026

Article presentation instructions notification

June, 2026

Publication (Springer CCIS series)

September/October, 2026
*full, short and special track articles Late-breaking work & demos
 

Late-breaking work & demo author/title/abstract registration opens on submission platform (easy-chair):

March 01, 2026

Late-breaking work & demo article upload deadline on submission platform (easy-chair):

March 07, 2026

Notification of acceptance:

March 31, 2026

Registration (payment) & Late-breaking work & demo camera-ready (upload to easy-chair):

April, 10, 2026

Late-breaking work & demo presentation instructions notification

June, 2026

Publication (planned with CEUR-WS.org*)

September/October, 2026
*Proceedings shall be submitted to CEUR-WS.org for online publication Doctoral consortium (DC) proposals
 

DC Proposal author/title registration deadline opens on submission platform (easy-chair):

March 01, 2026

DC Proposal uploads deadline on the submission platform (easy-chair):

March 07, 2026

Notification of acceptance:

March 31, 2026

Registration (payment)

April 10, 2026

DC presentation and meeting instructions notification

June, 2026

Publication (planned with CEUR-WS.org*)

September/October, 2026
*Proceedings shall be submitted to CEUR-WS.org for online publication
 
 
 
 
Review process
 
The Peer-Review process
 
All articles submitted within the deadlines and per the guidelines will be subjected to a single-blind review. Authors can also opt-out to disclose their names. However, authors will not know the names of their reviewers. Papers that are out of scope, incomplete, or lack sufficient evidence to support the basic claims may be rejected without full review. Manuscripts that do not conform to the specified formatting style will be desk-rejected. A non-dual policy submission exists, and articles submitted parallel to another conference/journal will be desk-rejected. Furthermore, reviewers will be asked to comment on whether the length is appropriate for the contribution. Each submitted article will be reviewed by at least two appropriate committee members (main/special track programme committee, late-breaking work/demo/DC committee).
 
After completion of the review process, the authors will be informed about the acceptance or rejection of the submitted work. The reviewers’ comments will be available to the authors if they are not desk-rejected. In case of acceptance, authors must meet the recommendations for improvement and prepare and submit the definitive version of the work up to the camera-ready paper submission deadline. In case of failure to consider the recommendations made by the reviewers, the organizing committee, the chairs and the editors reserve the right not to include these works in any of the planned conference proceedings.
 
The article’s final version must follow the appropriate style guide and contain the authors’ data (names, institutions and emails) and the ORCID details. Submitted articles will be evaluated according to their originality, technical soundness, significance of findings, contribution to knowledge, clarity of exposition and organisation and replicability.
 
Code of Ethics
Inspired by the code of ethics put forward by the Association of Computing Machinery, the programme committee, supervised by the general conference chairs and organisers, has the right to desk-reject manuscripts that perpetuate harmful stereotypes, employ unethical research practices, or uncritically present outcomes or implications that disadvantage minoritized communities. Further, reviewers of the scientific committee will be explicitly asked to consider whether the research was conducted in compliance with professional, ethical standards and applicable regulatory guidelines. Failure to do so could lead to a desk rejection.
 
Publication & indexing
 
Each accepted and presented full, short paper (for the main and special tracks), presented either as an oral presentation or as a poster, will be included in the conference proceedings by Springer in Communications in Computer and Information Science, edited by the general/PC chairs.  At least one author of each accepted paper must pay the related fees and register for the conference by the deadline. The official publication date is when the publisher makes the proceedings available online. This date will be after the conference and can take some weeks.
 
If authors want to publish their article OPEN ACCESS (upon a fee with Springer), please refer to this page.
Reply all
Reply to author
Forward
0 new messages