posted by user: g1910 || 5090 views || tracked by 3 users: [display]

MMDLCA 2020 : International Workshop on Multi-Modal Deep Learning: Challenges and Applications 2020

FacebookTwitterLinkedInGoogle

Link: https://medical-and-multimedia-lab.github.io/MMDLCA2020/
 
When Jan 11, 2021 - Jan 11, 2021
Where Milan, Italy
Submission Deadline Oct 15, 2020
Notification Due Nov 10, 2020
Final Version Due Nov 15, 2020
Categories    deep learning   computer vision   machine learning   artificial intelligence
 

Call For Papers

INFORMATION ON MMDLCA
Deep learning is now recognized as one of the key software engines that drives the new industrial revolution. The majority of current deep learning research efforts have been dedicated to single-modal data processing. Pronounced manifestations are deep learning based visual recognition and speech recognition. Although significant progress made, single-modal data is often insufficient to derive accurate and robust deep models in many applications. Our digital world is by nature multi-modal, that combines different modalities of data such as text, audio, images, animations, videos and interactive content. Multi-modal is the most popular form for information representation and delivery. For example, posts for hot social events are typically composed of textual descriptions, images and videos. For medical diagnosis, the joint use of medical imaging and textual reports is also essential. Multi-modal data is common for human to make accurate perceptions and decisions. Multi-modal deep learning that is capable of learning from information presented in multiple modalities and consequently making predictions based on multi-modal input is much in demand.

This workshop calls for scientific works that illustrate the most recent progress on multi-modal deep learning. In particular, multi-modal data capture, integration, modelling, understanding and analysis, and how to leverage them to derive accurate and robust AI models in many applications. It is a timely topic following the rapid development of deep learning technologies and their remarkable applications to many fields. It will serve as a forum to bring together active researchers and practitioners to share their recent advances in this exciting area. In particular, we solicit original and high-quality contributions in: (1) presenting state-of-the-art theories and novel application scenarios related to multi-modal deep learning; (2) surveying the recent progress in this area; and (3) developing benchmark datasets and evaluations. We welcome contributions coming from various communities (i.e., visual computing, machine learning, multimedia analysis, distributed and cloud computing, etc.) to submit their novel results.

TOPICS
The list of topics includes, but not limited to:
• Multi-modal intelligent data acquisition and management
• Multi-modal benchmark datasets and evaluations
• Multi-modal representation learning and applications
• Multi-modal data driven visual analysis and understanding
• Multi-modal object detection, classification, recognition and segmentation
• Multi-modal information tracking, retrieval and identification
• Multi-modal social event analysis
• Multi-modal medical diagnosis
• Multi-modal machine learning from incomplete data
• Deep neural network architectures for multi-modal data processing
• Multi-modal big data analytics
• Emerging multi-modal deep learning applications

SUBMISSION GUIDELINES
Submission site: https://easychair.org/my/conference?conf=mmdlca2020#
Submissions must be formatted in accordance with the Springer's Computer Science Proceedings guidelines (https://www.springer.com/gp/computer-science/lncs/conference-proceedings-guidelines). Two types of contribution will be considered:
• Full papers (10-12 pages, including references)
• Short papers (6-8 pages, including references)
Accepted manuscripts will be included in the ICPR 2020 Workshop Proceedings Springer volume. Once accepted, at least one author is expected to attend the event and orally present the paper.

IMPORTANT DATES
• Workshop submission deadline: Oct. 15th, 2020
• Workshop author notification: Nov. 10th, 2020
• Camera-ready submission: Nov. 15th, 2020
• Finalized workshop program: Dec. 1st, 2020
• Workshop day: Jan. 11, 2021

CONTACTS
For any inquiry you may have, please send an email to:
Zhineng Chen at zhineng.chen@ia.ac.cn

Related Resources

Human Activity Recognition 2023   Multi-Resident Activity Recognition using Deep Learning Models in the Smart Home Environment
IEEE COINS 2023   IEEE COINS 2023 - Berlin, Germany - July 23-25 - Hybrid (In-Person & Virtual)
ICMLA 2023   International Conference on Machine Learning and Applications
JCRAI 2023-Ei Compendex & Scopus 2023   2023 International Joint Conference on Robotics and Artificial Intelligence (JCRAI 2023)
EAICI 2024   Explainable AI for Cancer Imaging
JCICE 2023   2023 International Joint Conference on Information and Communication Engineering(JCICE 2023)
ESANN 2023   European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning
IEEE Xplore-Ei/Scopus-CCCAI 2023   2023 International Conference on Communications, Computing and Artificial Intelligence (CCCAI 2023) -EI Compendex
EAIH 2024   Explainable AI for Health
ACM-Ei/Scopus-CWCBD 2023   2023 4th International Conference on Wireless Communications and Big Data (CWCBD 2023) -EI Compendex