{ "paper_id": "2022", "header": { "generated_with": "S2ORC 1.0.0", "date_generated": "2023-01-19T01:11:32.444622Z" }, "title": "", "authors": [], "year": "", "venue": null, "identifiers": {}, "abstract": "", "pdf_parse": { "paper_id": "2022", "_pdf_hash": "", "abstract": [], "body_text": [ { "text": "Two years after the appearance of GPT-3, large language models seem to have taken over NLP. Their capabilities, limitations, societal impact and the potential new applications they unlocked have been discussed and debated at length. A handful of replication studies have been published since then, confirming some of the initial findings and discovering new limitations. This workshop aims to gather researchers and practitioners involved in the creation of these models in order to:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": null }, { "text": "1. Share ideas on the next directions of research in this field, including-but not limited to-grounding, multi-modal models, continuous updates and reasoning capabilities.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": null }, { "text": "2. Share best-practices, brainstorm solutions to identified limitations and discuss challenges, such as infrastructure, data, ethical & legal frameworks, evaluation, training efficiency, etc.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": null }, { "text": "This workshop is organized by the BigScience 1 initiative and will also serve as the closing session of this one year-long initiative aimed at developing a multilingual large language model, which is gathering 1.000+ researchers from more than 60 countries and 250 institutions and research labs. Its goal is to investigate the creation of a large scale dataset and model from a very wide diversity of angles.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": null }, { "text": "Organizing Committee ", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": null }, { "text": "https://bigscience.huggingface.co/", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null } ], "back_matter": [], "bib_entries": { "BIBREF0": { "ref_id": "b0", "title": "Lifelong Pretraining: Continually Adapting Language Models to Emerging Corpora Xisen Jin", "authors": [ { "first": "Dejiao", "middle": [], "last": "Zhang", "suffix": "" }, { "first": "Henghui", "middle": [], "last": "Zhu", "suffix": "" }, { "first": "Wei", "middle": [], "last": "Xiao", "suffix": "" }, { "first": "", "middle": [], "last": "Shang-Wen", "suffix": "" }, { "first": "Xiaokai", "middle": [], "last": "Li", "suffix": "" }, { "first": "Andrew", "middle": [], "last": "Wei", "suffix": "" }, { "first": "", "middle": [], "last": "Arnold", "suffix": "" }, { "first": ".", "middle": [ "." ], "last": "Xiang Ren", "suffix": "" } ], "year": null, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Lifelong Pretraining: Continually Adapting Language Models to Emerging Corpora Xisen Jin, Dejiao Zhang, Henghui Zhu, Wei Xiao, Shang-Wen Li, Xiaokai Wei, Andrew Arnold and Xiang Ren . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1", "links": null }, "BIBREF1": { "ref_id": "b1", "title": "You reap what you sow: On the Challenges of Bias Evaluation Under Multilingual Settings Zeerak Talat", "authors": [ { "first": "Valentin", "middle": [], "last": "Pelloin", "suffix": "" }, { "first": "Franck", "middle": [], "last": "Benoit Favre", "suffix": "" }, { "first": "Antoine", "middle": [], "last": "Dary", "suffix": "" }, { "first": "Sylvain", "middle": [], "last": "Laurent", "suffix": "" }, { "first": "Laurent", "middle": [], "last": "Meignier", "suffix": "" }, { "first": ".", "middle": [ "." ], "last": "Besacier", "suffix": "" } ], "year": null, "venue": "Using ASR-Generated Text for Spoken Language Modeling Nicolas Herv\u00e9", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Using ASR-Generated Text for Spoken Language Modeling Nicolas Herv\u00e9, Valentin Pelloin, Benoit Favre, Franck Dary, Antoine Laurent, Sylvain Meignier and Laurent Besacier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 You reap what you sow: On the Challenges of Bias Evaluation Under Multilingual Settings Zeerak Talat, Aur\u00e9lie N\u00e9v\u00e9ol, Stella Biderman, Miruna Clinciu, Manan Dey, Shayne Long- pre, Sasha Luccioni, Maraim Masoud, Margaret Mitchell, Dragomir Radev, Shanya Sharma, Arjun Subramonian, Jaesung Tae, Samson Tan, Deepak Tunuguntla and Oskar Van Der Wal . . . . . . . . . . . . . 26", "links": null }, "BIBREF3": { "ref_id": "b3", "title": "UNIREX: A Unified Learning Framework for Language Model Rationale Extraction Aaron Chan", "authors": [], "year": null, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "UNIREX: A Unified Learning Framework for Language Model Rationale Extraction Aaron Chan, Maziar Sanjabi, Lambert Mathias, Liang Tan, Shaoliang Nie, Xiaochang Peng,", "links": null }, "BIBREF4": { "ref_id": "b4", "title": "51 Pipelines for Social Bias Testing of Large Language Models Debora Nozza, Federico Bianchi and Dirk Hovy", "authors": [ { "first": "Xiang", "middle": [], "last": "Ren", "suffix": "" }, { "first": "Hamed", "middle": [ ". . ." ], "last": "Firooz", "suffix": "" } ], "year": null, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Xiang Ren and Hamed Firooz . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51 Pipelines for Social Bias Testing of Large Language Models Debora Nozza, Federico Bianchi and Dirk Hovy. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .68", "links": null }, "BIBREF5": { "ref_id": "b5", "title": "Zero-Shot on Historical Texts with T0", "authors": [ { "first": "Dates", "middle": [], "last": "Entities", "suffix": "" }, { "first": "Languages", "middle": [], "last": "", "suffix": "" } ], "year": null, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Entities, Dates, and Languages: Zero-Shot on Historical Texts with T0", "links": null }, "BIBREF7": { "ref_id": "b7", "title": "A Holistic Assessment of the Carbon Footprint of Noor, a Very Large Arabic Language Model Imad Lakim, Ebtesam Almazrouei", "authors": [], "year": null, "venue": "Ibrahim Abualhaol, Merouane Debbah and Julien Launay", "volume": "84", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "A Holistic Assessment of the Carbon Footprint of Noor, a Very Large Arabic Language Model Imad Lakim, Ebtesam Almazrouei, Ibrahim Abualhaol, Merouane Debbah and Julien Launay 84", "links": null }, "BIBREF8": { "ref_id": "b8", "title": "An Open-Source Autoregressive Language Model Sidney Black", "authors": [ { "first": "Eric", "middle": [], "last": "Gpt-Neox-20b ; Stella Biderman", "suffix": "" }, { "first": "Quentin", "middle": [ "Gregory" ], "last": "Hallahan", "suffix": "" }, { "first": "Leo", "middle": [], "last": "Anthony", "suffix": "" }, { "first": "Laurence", "middle": [], "last": "Gao", "suffix": "" }, { "first": "Horace", "middle": [], "last": "Golding", "suffix": "" }, { "first": "Connor", "middle": [], "last": "He", "suffix": "" }, { "first": "Kyle", "middle": [], "last": "Leahy", "suffix": "" }, { "first": "", "middle": [], "last": "Mcdonell", "suffix": "" } ], "year": null, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "GPT-NeoX-20B: An Open-Source Autoregressive Language Model Sidney Black, Stella Biderman, Eric Hallahan, Quentin Gregory Anthony, Leo Gao, Laurence Golding, Horace He, Connor Leahy, Kyle McDonell, Jason Phang, Michael Martin Pieler, Usvsn Sai Prashanth, Shivanshu Purohit, Laria Reynolds, Jonathan Tow, Ben Wang and Samuel Weinbach . . . . 95", "links": null }, "BIBREF9": { "ref_id": "b9", "title": "Dataset Debt in Biomedical Language Modeling Jason Alan Fries", "authors": [ { "first": "Natasha", "middle": [], "last": "Seelam", "suffix": "" }, { "first": "Gabriel", "middle": [], "last": "Altay", "suffix": "" }, { "first": "Leon", "middle": [], "last": "Weber", "suffix": "" }, { "first": "Myungsun", "middle": [], "last": "Kang", "suffix": "" }, { "first": "Debajyoti", "middle": [], "last": "Datta", "suffix": "" }, { "first": "Ruisi", "middle": [], "last": "Su", "suffix": "" }, { "first": "Samuele", "middle": [], "last": "Garda", "suffix": "" }, { "first": "Bo", "middle": [], "last": "Wang", "suffix": "" }, { "first": "Simon", "middle": [], "last": "Ott", "suffix": "" }, { "first": "Matthias", "middle": [], "last": "Samwald", "suffix": "" }, { "first": "Wojciech", "middle": [ ". . . ." ], "last": "Kusa", "suffix": "" } ], "year": null, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Dataset Debt in Biomedical Language Modeling Jason Alan Fries, Natasha Seelam, Gabriel Altay, Leon Weber, Myungsun Kang, Debajyoti Datta, Ruisi Su, Samuele Garda, Bo Wang, Simon Ott, Matthias Samwald and Wojciech Kusa . . . . . . . . . . 137", "links": null }, "BIBREF10": { "ref_id": "b10", "title": "Emergent Structures and Training Dynamics in Large Language Models", "authors": [ { "first": "Ryan", "middle": [], "last": "Teehan", "suffix": "" }, { "first": "Miruna", "middle": [], "last": "Clinciu", "suffix": "" }, { "first": "Oleg", "middle": [], "last": "Serikov", "suffix": "" }, { "first": "Eliza", "middle": [], "last": "Szczechla", "suffix": "" }, { "first": "Natasha", "middle": [], "last": "Seelam", "suffix": "" }, { "first": "Shachar", "middle": [], "last": "Mirkin", "suffix": "" }, { "first": "Aaron", "middle": [], "last": "Gokaslan", "suffix": "" }, { "first": ".", "middle": [ "." ], "last": "", "suffix": "" } ], "year": null, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Emergent Structures and Training Dynamics in Large Language Models Ryan Teehan, Miruna Clinciu, Oleg Serikov, Eliza Szczechla, Natasha Seelam, Shachar Mirkin and Aaron Gokaslan . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146", "links": null }, "BIBREF11": { "ref_id": "b11", "title": "160 Program Friday", "authors": [ { "first": "Ellyn", "middle": [], "last": "Ayton", "suffix": "" }, { "first": "Shivam", "middle": [], "last": "Sharma", "suffix": "" }, { "first": "Scott", "middle": [], "last": "Howland", "suffix": "" }, { "first": "Megha", "middle": [], "last": "Subramanian", "suffix": "" }, { "first": "Scott", "middle": [], "last": "Vasquez", "suffix": "" }, { "first": "Robin", "middle": [], "last": "Cosbey", "suffix": "" }, { "first": "Maria", "middle": [], "last": "Glenski", "suffix": "" }, { "first": "Svitlana", "middle": [ ". . . . . . . . . . . . . . . . . . . . . . . . ." ], "last": "Volkova", "suffix": "" } ], "year": 2022, "venue": "Foundation Models of Scientific Knowledge for Chemistry: Opportunities, Challenges and Lessons Learned Sameera Horawalavithana", "volume": "12", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Foundation Models of Scientific Knowledge for Chemistry: Opportunities, Challenges and Lessons Learned Sameera Horawalavithana, Ellyn Ayton, Shivam Sharma, Scott Howland, Megha Subramanian, Scott Vasquez, Robin Cosbey, Maria Glenski and Svitlana Volkova . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160 Program Friday, May 27, 2022 12:30 -11:00 Poster Session 14:00 -15:00 BigScience 15:00 -15:20 Data Governance 15:20 -15:40 Data 15:40 -16:00 Modeling 16:00 -16:20 Prompt Engineering 16:20 -16:40 Evaluation viii", "links": null } }, "ref_entries": {} } }