posted by user: grupocole || 2629 views || tracked by 6 users: [display]

DeepXplain 2025 : IJCNN Special Session on Explainable Deep Neural Networks for Responsible AI: Post-Hoc and Self-Explaining Approaches

FacebookTwitterLinkedInGoogle

Link: https://deepxplain.github.io/
 
When Jun 30, 2025 - Jul 5, 2025
Where Roma, Italy
Submission Deadline Jan 15, 2025
Notification Due Mar 15, 2025
Final Version Due May 1, 2025
Categories    NLP   computational linguistics   artificial intelligence
 

Call For Papers

We are delighted to invite you to participate in "Explainable Deep Neural Networks for Responsible AI: Post-Hoc and Self-Explaining Approaches (DeepXplain 2025)," a special session at IJCNN 2025 dedicated to innovative methodologies for improving the interpretability of Deep Neural Networks (DNNs) while maintaining high predictive accuracy.

Website: https://deepxplain.github.io/
Contributions

This special session aims to foster interdisciplinary collaboration, promote the ethical design of AI systems, and encourage the development of benchmarks and datasets for explainability research. Our goal is to advance both post-hoc and intrinsic interpretability approaches, bridging the gap between the high performance of deep neural networks and their transparency. By doing so, we seek to enhance human trust in these models and mitigate the risks of negative social impacts.

Topics of interest include, but are not limited to:

Theoretical advancements in post-hoc explanation methods (e.g., LIME, SHAP, Grad-CAM) for DNNs.

Development of inherently interpretable architectures using self-explaining mechanisms, such as attention-based or saliency-based models, prototype networks, and SENNs (Self-Explaining Neural Networks).

Post-hoc and self-explaining methods for Large Language Models (LLMs).

Application-driven explainability insights, particularly in Natural Language Processing and Computer Vision.

Ethical evaluations of DNN-based AI models with a focus on reducing bias and social impact.

Methods, metrics, and methodologies for improving interpretability and fairness in DNNs.

Ethical discussions about the social impact of non-transparent AI models.

Datasets and benchmarking tools for explainability.

Explainable AI in critical applications: healthcare, governance, misinformation, hate speech, etc.

Submission Information

We welcome submissions of academic papers (both long and short) across the spectrum of theoretical and practical work, including research ideas, methods, tools, simulations, applications or demonstrations, practical evaluations, position papers, and surveys. Submissions must be written in English, adhere to the IJCNN-2025 formatting guidelines, and be submitted as a single PDF file.

Important Dates:

Submission link: https://cmt3.research.microsoft.com/IJCNN2025/

Submission deadline: January 15, 2025

Notification date: March 15, 2025

Camera-ready submission: May 1, 2025

Organizers

Francielle Vargas, University of São Paulo, Brazil

Roseli Romero, University of São Paulo, Brazil

Jackson Trager, University of Southern California, USA

Edson Prestes, Federal University of Rio Grande do Sul, Brazil

Related Resources

CVNN 2026   2026 International Conference on Computer Vision and Neural Networks-EI/Scopus
IEEE-ICECCS 2026   2025 IEEE International Conference on Electronics, Communications and Computer Science (ICECCS 2026)
ACII 2026   Advanced Computational Intelligence: An International Journal
DEPLING 2023   International Conference on Dependency Linguistics
IJWMN 2026   International Journal of Wireless & Mobile Networks -- ERA Indexed, H index 38
Ei/Scopus-CMLDS 2026   2026 3rd International Conference on Computing, Machine Learning and Data Science (CMLDS 2026)
NeurIPS 2026   The Fortieth Annual Conference on Neural Information Processing Systems
IJESA 2026   International Journal of Embedded Systems and Applications
AMLDS 2026   IEEE--2026 2nd International Conference on Advanced Machine Learning and Data Science
VLSICS 2026   International Journal of VLSI design & Communication Systems