posted by user: grupocole || 778 views || tracked by 2 users: [display]

BigScience-5 2022 : Workshop on Challenges & Perspectives in Creating Large Language Models


When May 27, 2022 - May 27, 2022
Where Dublin, Ireland
Submission Deadline Feb 28, 2022
Notification Due Mar 26, 2022
Final Version Due Apr 10, 2022
Categories    NLP   computational linguistics   artificial intelligene

Call For Papers

****Submission Deadline: Feb 28th****

Call for Papers: Workshop on Challenges & Perspectives in Creating Large Language Models
May 27th 2022 (w/ ACL)

Two years after the appearance of GPT-3, large language models seem to have taken over NLP. Their capabilities, limitations, societal impact and the potential new applications they unlocked have been discussed and debated at length. A handful of replication studies have been published since then, confirming some of the initial findings and discovering new limitations. This workshop aims to gather researchers and practitioners involved in the creation of these models in order to:

1. Share ideas on the next directions of research in this field, including – but not limited to – grounding, multi-modal models, continuous updates and reasoning capabilities.
2. Share best-practices, brainstorm solutions to identified limitations and discuss challenges, such as:

‍Infrastructure. What are the infrastructure and software challenges involved in scaling models to billions or trillions of parameters, and deploying training and inference on distributed servers when each model replicas is itself larger than a single node capacity?
‍Data. While the self-supervised setting dispenses with human annotation, the importance of cleaning, filtering and the bias and limitation in existing or reported corpora has become more and more apparent over the last years.
‍Ethical & Legal frameworks. What type of data can/should be used, what type of access should be provided, what filters are or should be necessary?
‍Evaluation. Investigating the diversity of intrinsic and extrinsic evaluation measures, how do they correlate and how the performances of a very large pretrained language model should be evaluated.
‍Training efficiency. Discussing the practical scaling approaches, practical questions around large scale training hyper-parameters and early-stopping conditions. Discussing measures to reduce the associated energy consumption.

This workshop is organized by the BigScience initiative and will also serve as the closing session of this one year-long initiative aimed at developing a multilingual large language model, which is currently gathering 900 researchers from more than 60 countries and 250 institutions. Its goal is to investigate the creation of a large scale dataset and model from a very wide diversity of angles.

We call for relevant contributions, either in long (8 pages) or short (4 pages) format. Accepted papers will be presented during a poster session. Submissions can be archival or non-archival.
Submissions should be made via OpenReview (

Feb. 28, 2022: Submission Deadline
‍March 26, 2022: Notification of Acceptance
‍April 10, 2022: Camera-ready papers due

Related Resources

IPM-LLMDQKG 2025   Special issue of Information Processing & Management on Large Language Models and Data Quality for Knowledge Graphs
Ei/Scopus-AACIP 2024   2024 2nd Asia Conference on Algorithms, Computing and Image Processing (AACIP 2024)-EI Compendex
KGLLM 2024   Special session on Knowledge Graphs and Large Language Models
IEEE Big Data - MMAI 2024   IEEE Big Data 2024 Workshop on Multimodal AI (Hybrid)
LLMCS 2024   The International Workshop on Large Language Models for Cybersecurity
ISEEIE 2024   2024 4th International Symposium on Electrical, Electronics and Information Engineering (ISEEIE 2024)
PromptEng 2024   The International Workshop on Prompt Engineering Large Language Models
NLPCL 2024   5th International Conference on Natural Language Processing and Computational Linguistics
GenAISE 2024   The International Workshop on Advances of GenAI in Software Engineering
NLCA 2024   5th International Conference on Natural Language Computing Advances