CNRS News
Published on CNRS News (https://news.cnrs.fr)

Home > Bigscience: a new language model for all

Bigscience: a new language model for all

You are here
Home [1]
Lire en français [2]
Digital [3]
Artificial Intelligence [4]
-A [5] +A [5]
article

Bigscience: a new language model for all

07.01.2022, by
Martin Koppe [6]
Reading time: 5 minutes
vegefox.com /Stock.Adobe.com
Open source and covering forty-six different languages, BigScience is a new artificial intelligence heavyweight offering natural language processing. Claire Gardent, a CNRS research professor at the LORIA and winner of the CNRS Silver Medal, who also sits on the project’s steering committee, explains the principle as well as its potential applications.

What is BigScience?
Claire Gardent:1 BigScience represents a major progress for language models, which calculate the probability of a sequence of words, and then generate text by linking together the terms that have the highest probability of following one another. Older models were restricted to narrow contexts, typically five consecutive words, but this limitation has been removed by neural networks that can take context – theoretically unlimited – into account. In practice, they have truly revolutionised text generation by producing content that is practically perfect in terms of grammar. For a short piece of writing, it is very difficult to distinguish between their prose and what a human would have written.

Discussion of different examples of automatic language processing, with automatic text and sign language generation with a virtual signer and latent sentence representation, at the Laboratoire interdisciplinaire des sciences du numérique (LISN).
CNRS News
Discussion of different examples of automatic language processing, with automatic text and sign language generation with a virtual signer and latent sentence representation, at the Laboratoire interdisciplinaire des sciences du numérique (LISN).
Christian MOREL / LISN / CNRS Photothèque
Christian MOREL / LISN / CNRS Photothèque
Share
Share
[7] [8] [9]

Yet until now these models were developed mostly in English and by multinationals, which made them available, but provided no information regarding their learning details nor the data used. The idea behind BigScience is to create new language models, using a multilingual and unprecedented open-source approach. 
BigScience also enjoys a very interesting collaborative aspect. Although coordinated by the private company Hugging Face, which completed a significant part of the engineering, the project has brought together a high-calibre international academic community. 

How are language models developed?
C. G.: 
They are made of neural networks that self learn from large quantities of texts. They take the first word in a sentence and try to predict the second, then the third, etc. Each prediction is compared with the term that features in the text, and in the event of an error the model's weighting is adapted using what is known as a backpropagation mechanism. Learning continues until the model converges, in other words until its predictions are reliable.

What obstacles and difficulties had to be overcome by BigScience?
C. G.:  To learn these probabilities, the model needs data, and computing power, which is provided by the Jean Zay supercomputer.2 BigScience also had to manage the scaling up resulting from its multilingual nature. Considerable reflection and engineering was required to create high-quality training data, as well as to train these large models.
The information used is automatically collected on the Internet. This poses little difficulty in the case of English, for which large bodies of text are already available. But BigScience covers forty-six languages, which are more or less present online. Significant amounts of data must therefore be assembled in tongues for which little material is available, and is not always of good quality. For example, working on Breton, I observed that a body of texts that had automatically been created for machine translation included content that was actually in English and Chinese. This must be avoided across vast volumes of data, and the presence of biases and offensive wording must also me minimised. 

The Jean Zay converged supercomputer extends the traditional uses of high-performance computing (HPC) to new uses for artificial intelligence (AI). Through two successive extensions, its capacity has been increased to 28.3 petaflops by 2022.
CNRS News
The Jean Zay converged supercomputer extends the traditional uses of high-performance computing (HPC) to new uses for artificial intelligence (AI). Through two successive extensions, its capacity has been increased to 28.3 petaflops by 2022.
Cyril FRESILLON / IDRIS / CNRS Photothèque
Cyril FRESILLON / IDRIS / CNRS Photothèque
Share
Share
[7] [10] [9]

What are the potential applications?
C. G.: All tasks relating to text generation. Machine translation remains the leading application, although simplifying, paraphrasing, or summarising are important. Generation models are also used to convert knowledge bases or numeric databases into text, for example to produce a weather report from raw meteorological data. And if it is trained on dialogue, a language model can be integrated within a conversational agent. Finally, question and answer models often use generation to provide answers based on the question and relevant information from the Internet.

A Google AI engineer recently affirmed that the language model he is working on, LaMDA, had gained consciousness. What is your reaction to such declarations?
C. G.: The intelligence of language models is an issue that arises regularly. They have truly amazing capacities, and some examples are actually quite remarkable. However, this does not mean that the programme is aware of what it is saying. Natural-language understanding involves reasoning that is complex and varied (spatial, temporal, ontological, arithmetical), based on knowledge, and can connect form and meaning (objects and actions in the real world). While a few selected examples may suggest that language models are capable of such reasoning, many incorrect cases crop up when working with them everyday.

With Angela Fan of Facebook AI Research Paris, I focused on an algorithm that generates Wikipedia pages. The wording is perfect, and seems to have been written by a human. Yet while the form is correct, the content is full of factual errors. A basketball player’s webpage sometimes indicates that he is a tennis player, certain dates are wrong, and the university where he studied is incorrect. After a few lines, an automatically-generated text often ends up contradicting itself. If language models were capable of understanding, they would not generate such incoherence.

We must be wary of hype created by a few well-chosen examples. Research should continue to develop tasks and benchmarks that truly test the capacity of natural language processing models. They should be assessed based on reasoning exercises that are multidomain, multilingual, and that interact with the world, for example via systems in which the computer has to answer questions regarding the content of an image, video, or conversation.  

There are other essential objectives, such as the models’ capacity for generalisation, in other words ensuring that they function equally well on content that is journalistic, technical, literary, etc. Models must also be designed for languages that are supposedly less well-endowed. By gathering data covering forty-six different tongues, BigScience has taken a major step in this direction.

Footnotes
  • 1. LORIA (CNRS / University of Lorraine / INRIA).
  • 2. Managed by the company GENCI (Grand équipement national de calcul intensif).

Explore more

Digital
 metamorworks / iStock.com
[11]
Article
12/02/2025
Science one step ahead [11]
[12]
Article
10/28/2025
Does AI care about us? [12]
Stephane Mallat at work
[13]
Article
09/10/2025
Stéphane Mallat, a pioneer bridging mathematics and computer... [13]
Les Linceuls de David Cronenberg 2024 © Prospero Pictures / SBS Productions / Saint Laurent / Collection ChristopheL
[14]
Article
07/22/2025
Dying with the times [14]
[15]
Article
07/02/2025
Solace of quantum [15]
Artificial Intelligence
[12]
Article
10/28/2025
Does AI care about us? [12]
[16]
Article
02/07/2025
AI needs to align with human values [16]
[17]
Article
11/29/2023
The challenges of frugal AI [17]
[18]
Article
11/29/2023
AI detects the early signs of multiple sclerosis [18]
[19]
Article
06/12/2022
IA leads the charge against multiple sclerosis [19]

Keywords

Foundation model [20] Artificial Intelligence [21] Multilingual [22]

Share this article

[23]
[24]
[7]
[9]

Source URL:https://news.cnrs.fr/articles/bigscience-a-new-language-model-for-all

Links
[1] https://news.cnrs.fr/ [2] https://lejournal.cnrs.fr/articles/bigscience-voit-grand-pour-les-modeles-de-langue [3] https://news.cnrs.fr/digital [4] https://news.cnrs.fr/artificial-intelligence [5] https://news.cnrs.fr/javascript%3A%3B [6] https://news.cnrs.fr/authors/martin-koppe [7] https://twitter.com/intent/tweet?url=https%3A//news.cnrs.fr/print/1867%2F&text=Bigscience: a new language model for all [8] http://www.facebook.com/sharer/sharer.php?s=100&p%5Burl%5D=https%3A//news.cnrs.fr/print/1867&p%5Btitle%5D=Bigscience%3A%20%20a%20new%20language%20model%20for%20all&p%5Bimages%5D%5B0%5D=https%3A//news.cnrs.fr/sites/default/files/styles/lightbox-hd/public/assets/images/272_63b1045_72dpi.jpg%3Fitok%3DzxD9RzDd&p%5Bsummary%5D= [9] https://bsky.app/intent/compose?text=Bigscience: a new language model for all%0Ahttps%3A//news.cnrs.fr/print/1867 [10] http://www.facebook.com/sharer/sharer.php?s=100&p%5Burl%5D=https%3A//news.cnrs.fr/print/1867&p%5Btitle%5D=Bigscience%3A%20%20a%20new%20language%20model%20for%20all&p%5Bimages%5D%5B0%5D=https%3A//news.cnrs.fr/sites/default/files/styles/lightbox-hd/public/assets/images/cnrs_20190051_0007_72dpi.jpg%3Fitok%3DlW-f5_AK&p%5Bsummary%5D= [11] https://news.cnrs.fr/articles/science-one-step-ahead [12] https://news.cnrs.fr/articles/does-ai-care-about-us [13] https://news.cnrs.fr/articles/stephane-mallat-a-pioneer-bridging-mathematics-and-computer-science [14] https://news.cnrs.fr/articles/dying-with-the-times [15] https://news.cnrs.fr/articles/solace-of-quantum [16] https://news.cnrs.fr/articles/ai-needs-to-align-with-human-values [17] https://news.cnrs.fr/articles/the-challenges-of-frugal-ai [18] https://news.cnrs.fr/articles/ai-detects-the-early-signs-of-multiple-sclerosis [19] https://news.cnrs.fr/articles/ia-leads-the-charge-against-multiple-sclerosis [20] https://news.cnrs.fr/foundation-model [21] https://news.cnrs.fr/artificial-intelligence-0 [22] https://news.cnrs.fr/multilingual [23] http://www.facebook.com/sharer/sharer.php?s=100&p%5Burl%5D=https%3A//news.cnrs.fr/print/1867&p%5Btitle%5D=Bigscience%3A%20%20a%20new%20language%20model%20for%20all&p%5Bimages%5D%5B0%5D=&p%5Bsummary%5D= [24] https://news.cnrs.fr/printmail/1867