posted by organizer: zhangyudong || 2245 views || tracked by 3 users: [display]

XDLMIPA 2022 : Explainable Deep Learning for Medical Image Processing and Analysis


When N/A
Where N/A
Abstract Registration Due Nov 1, 2021
Submission Deadline Mar 1, 2022
Notification Due Apr 1, 2022
Final Version Due May 1, 2022
Categories    deep learning   image processing

Call For Papers

I. Aim and Scope

Deep learning (DL) is an emerging computational intelligence method that plays a major role in data science and its applications. DL provides computational models of multiple processing layers to learn and represent data with multiple levels of abstraction. It can implicitly capture intricate structures of large-scale data and is ideally suited to some of the hardware architectures currently available. DL has recently achieved outstanding performance in academic and industrial fields and become a vital utensil in a wide range of medical image processing and analysis tasks, including cancer detection, tumor classification, vessel segmentation, etc.
While DL models give impressively high predictive accuracy, they are recognized as black-boxes with deep and complicated layers. In the meantime, DLs have been recently reported as defenseless to spoofing with elegant hand-designed input samples. This principally occurs in the medical image processing field, where a single incorrect prediction might be detrimental, and the trust in the trained DL model and its capacity to deliver efficient and robust data processing must be pledged. Therefore, understanding how the DL models work by means of validation and visualization procedures and thus creating explainable DL models has become an elemental problem.
It is still unclear what information must be delivered to DL models, the best validation procedure to assess their generalization ability, and how DL models work to warrant a rapid, safe and robust prediction. Hence, experts/users request the latest research advances in explainable deep learning (XDL). This critical research topic will bring new challenges and opportunities to the AI community. This SI aims to provide a diverse but complementary set of contributions to demonstrate new developments and applications of explainable deep learning to solve problems in medical image processing. The ultimate goal is to promote research and development of explainable deep learning for multimodal biomedical images by publishing high-quality research articles and reviews in this rapidly growing interdisciplinary field.
The medical data can be acquired from multiple imaging modalities, such as Computed Tomography (CT), Magnetic Resonance (MR) Imaging, Ultrasound, Single Photon Emission Computed Tomography (SPECT), Positron Emission Tomography (PET), Magnetic Particle Imaging (MPI), EE/MEG, Optical Microscopy and Tomography, Photoacoustic Tomography, Electron Tomography, and Atomic Force Microscopy, etc.

II. Topics

Scopes (but are not limited to) the following:

• Qualification/Visualization/Interpretation of learned representations from deep neural network models
• Novel theoretical understanding of deep learning in biomedical engineering
• Explainable transfer learning and multi-task learning
• Analyzing bottleneck of efficient learning of deep neural networks
• Explainable joint semantic segmentation, object detection, and scene recognition on biomedical images
• Improvising on the computation of a deep network; exploiting parallel computation techniques and GPU programming
• Inferring and regularizing network structure for robust prediction
• Translational multimodality imaging and biomedical applications (e.g., detection, diagnostic analysis, quantitative measurements, image guidance)
• Optimization by deep neural networks, multi-dimensional deep learning
• New models of new structure of convolutional neural network
• Adversarial attacks and defending in medical-image-processing applications
• Explainable fusion of deep learning with fuzzy systems and/or evolutionary computations
• Generative Deep Learning for data augmentation
• Predictive modelling of progression based on LSTM and RNN
• How to make deep learning algorithms more explainable

III. Submissions

SUBMISSIONSManuscripts should be prepared according to the “Information for Authors” section of the journal, and submissions should be made through the journal submission website:, by selecting the Manuscript Type of “explainable deep learning for medical image processing and analysis (XDLMIPA)” and clearly marking “explainable deep learning for medical image processing and analysis (XDLMIPA)” in the comments to the Editor-in-Chief. Submitted papers will be reviewed by domain experts. Submission of a manuscript implies that it is the authors’ original, unpublished work and is not being submitted for possible publication elsewhere.

V. Guest Editors

 Prof. Dr. Yu-Dong Zhang, University of Leicester, UK,
 Prof. Dr. Juan Manuel Gorriz, Cambridge University, UK/ University of Granada, Spain
 Prof. Yi Pan, Georgia State University, USA
 Prof. Oscar Cordon, E.T.S. Ingenierías Informática y Telecomunicación/University of Granada

Related Resources

EAICI 2024   Explainable AI for Cancer Imaging
IEEE Xplore-Ei/Scopus-CCCAI 2023   2023 International Conference on Communications, Computing and Artificial Intelligence (CCCAI 2023) -EI Compendex
EAIH 2024   Explainable AI for Health
DPPR 2023   13th International Conference on Digital Image Processing and Pattern Recognition
OMLMIA 2023   Optimization and Machine Learning in Medical Image Analysis
CNWC 2023   2023 International Conference on Computer Networks and Wireless Communication (CNWC 2023)
CCVPR 2023   2023 6th International Joint Conference on Computer Vision and Pattern Recognition (CCVPR 2023)
AAIMLB 2023   Application of Artificial Intelligence and Machine Learning in Biology
CFIMA 2023   2023 3rd International Conference on Frontiers of Intelligent Manufacturing and Automation (CFIMA 2023)
CSEC 2023   2023 International Conference on Computer Science, Electronics and Communications (CSEC 2023)