A Gigamodel in a porcelain shop… Part 1

Summary

Summary

Summary

Draw me a Gigamodel

First playground: language

Data without limits with self-supervised learning

The cement of  "foundation models"

LLMs, gigantic and generative language models

A Gigamodel is computing... with or without a license

Gigamodels are displayed

Speech is not left behind

Conference: ViaDialog participates in the Artificial Intelligence Festival on November 15, 2022!


gigamodels AI
gigamodels AI
gigamodels AI

For the past two or three years, Gigamodels such as BERT, GPT3, wave2vec, or DALL-E have been a major trend in Artificial Intelligence (AI). These models, with impressive performance, exceed a billion parameters, require specific infrastructure, and are produced only by a small handful of players. How is this new landscape disrupting both AI research and industry?

Draw me a Gigamodel

What is a Gigamodel? A model that counts hundreds of millions of parameters, or even hundreds of billions of parameters for the largest ones currently. We think of BERT, GPT-3, Wav2vec (pronounced in English as "wave to vec"), DALL-E (pronounced as "Dali")... Some call them "Foundation models" or  "Pre-trained models", others speak of "Large Language Models" (LLM).  They are generally models that include self-supervised learning (self-supervised learning). But what do these terms encompass?

First playground: language

The first Gigamodels to gain attention are natural language processing models (NLP = Natural Language Processing): the BERT model (Bidirectional Encoder Representations from Transformers), released by Google in 2018, which includes several hundred million parameters, followed by the highly publicized, but much more closed, GPT3 model trained by OpenAI, released in 2020, the first to reach one hundred billion parameters.

BERT, like GPT3, takes text – or rather "embeddings" – named word2vec, which encode words as vectors based on the distribution of neighboring words and perform language processing tasks.  They share a so-called "transformer" architecture, a deep neural architecture typically including an encoder, a decoder, and an attention layer, which makes them capable of considering an extended context. They especially have the ability to learn from very large volumes of unlabeled raw data, which is called "self-supervised" learning or "self-supervised learning".

Unlimited data with self-supervised learning

Traditional models obtained through machine learning use supervised learning: input data, such as sentences, are presented to the model, which seeks to predict an output, such as a classification, which it then compares with the reference label. The model then adapts as needed to get as close as possible to the desired output. The availability of sufficiently written and labeled data for each targeted task, each language, each domain, is a limiting factor of this type of approach.

Self-supervised learning, on the other hand, involves learning a generic model from unlabeled data, optimizing it on self-supervised tasks, which depend only on the data itself: for example, as illustrated in the diagram below, guessing masked words, arbitrarily masked in a text, based on all the other surrounding words, or guessing whether two sentences follow each other or do not follow at all in the corpus.

This type of self-supervised approach frees itself from the need for labeled data and opens up the possibility of learning generic models on very large volumes of raw data collected broadly. The self-supervised model implicitly captures information about the very structure of the data, here the distribution of words in the sentences of the language, their synonyms, antonyms, analogies, agreements, etc.

The generic model, also called a pre-trained model, can then be complemented with a specific "fine-tuning" for the targeted task (classification, sentiment analysis, question-answering, language comprehension tasks, automatic summarization, translation...), and a much smaller volume of labeled data may be sufficient, as the generic model is automatically endowed with strong generalization capacity.

The cement of  "foundation models"

Upon its release in 2018, the BERT model revolutionized the world of natural language processing by allowing "fine-tuned" models to dethrone previous approaches in most NLP benchmarks, particularly in semantic analysis tasks. BERT is notably integrated into Google's search engine to provide structured answers when possible to semantically analyzed queries.

The fact that a pre-trained model can be adapted into numerous tailored models led to the term "Foundation models" or "foundation models", serving as a basis for models built atop them.

LLMs, gigantic and generative language models

After BERT,  GPT3 amazed with its native text generation capabilities and its abilities as a  "few-shot learner" or even a "zero-shot learner", meaning the ability to learn a new task with just a few examples, or even without any prior example given.

These models are called "language models" and for some time now "large language models" (LLM), due to their gigantic size.

All these LLMs: GPT3, T5, LaMDA, Bloom, OPT-175B, Megatron-Turing NLG… have generative capabilities.

Initially, a language model is simply one that assigns a probability to sequences of words or characters. Do recent LLMs exhibit unique language capabilities, going beyond this initial scope? Despite their prowess, the debate remains open...

The generative capability of LLMs is used not only to produce natural language (human language) but also other types of production, the most emblematic currently being computer code and images.

A Gigamodel does computing... with or without a license

In 2021, Microsoft, in partnership with OpenAI – in which the Mountain View firm invested 1 billion in 2019 –  and its subsidiary GitHub, a very popular code management platform, released "GitHub Copilot", a paid application that suggests lines of code to developers. A survey among users indicates that about a third of their code would already be produced by Copilot.

GitHub Copilot was trained on enormous amounts of opensource code collected from the web. A collective has recently filed a lawsuit against Microsoft, GitHub, and OpenAI, claiming that entire portions of opensource code can be "spit out" as is by Copilot without attribution to their author, in violation of opensource licenses. The proceedings are just beginning. Whatever the outcome, it will have an impact on the rights of all Gigamodels.

The Gigamodels are visible

The most visible Gigamodels currently are certainly image generation models from text such as Dall-E, released in 2021 by OpenAI, followed by Stable Diffusion, MidJourney, or Parti/Imagen (Google).

These models, trained on enormous quantities of images and texts collected from the Web,  produce based on textual statements, called "prompts", images that are generally aesthetic and highly realistic, like the famous "A teddy bear on a skateboard in Times Square", produced by DALL-E 2.

We can also mention this image produced by MidJourney to illustrate an article in The Economist titled "Huge Foundation models are turbo-charging AI progress".

A work produced with MidJourney recently won an award in Colorado, in the "digital art" category, which certainly sparked controversy.

A new feature called "outpainting" also allows the imagining of continuations to existing works beyond their frame.

Speech is not left behind

The revolution of deep learning began in 2012 with image recognition and speech recognition, as evidenced by the Turing Award awarded in 2019 jointly to Yann Le Cun (image recognition), Yoshua Bengio (speech recognition), and Geoff Hinton (their mentor to both).

Natural language processing quickly took hold of these new approaches and was the first to give rise to self-supervised Gigamodels. It was natural that speech processing and image processing in turn fed off the advances made in language processing.

It was in 2019 that wav2vec was released, followed by wav2vec2 in 2020 from Meta's laboratories — then Facebook. Wav2vec is the audio equivalent of BERT, a self-supervised model learned from large amounts of unlabeled raw audio.

The speech signal being continuous by nature, wav2vec2 must overcome this difficulty by first learning discrete audio tokens, marked by the letter "q" in the diagram below, to characterize sounds, before playing, like BERT, the game of guessing masked data portions.

Since 2020, the top places in classic speech recognition benchmarks have been swept by systems based on wav2vec2, as shown on the leaderboard below, for the "Librispeech" benchmark, a corpus of speech read by numerous speakers...

But then, what radical changes are Gigamodels poised to bring to the world of research and industry?

To be continued…

Conference: ViaDialog participates in the Festival of Artificial Intelligence on November 15, 2022!

The Computer Laboratory of Avignon (LIA) and the chair in artificial intelligence LIAvignon are hosting this year the 3rd edition of the AI festival from November 14 to 16.

ViaDialog will participate alongside Google and HugginFace by running a conference on November 15. Ariane Nabeth-Halber, Director of the AI Pole, will join the conference and debates surrounding Gigamodels in AI.

Location: Campus Hannah Arendt – Avignon University AT06 starting from 4:30 PM.

hight performer
high performer emea
best support
best support
best support

4.7/5 based on +200 reviews

+150 companies use ViaDialog to simplify their customer relations.

  • logo francetv
  • logo sg
  • logo maif
  • logo sodexo
  • logo parc asterix
  • logo eurostar
hight performer
high performer emea
best support
best support
best support

4.7/5 based on +200 reviews

+150 companies use ViaDialog to simplify their customer relations.

  • logo francetv
  • logo sg
  • logo maif
  • logo sodexo
  • logo parc asterix
  • logo eurostar
hight performer
high performer emea
best support
best support
best support

4.7/5 based on +200 reviews

+150 companies use ViaDialog to simplify their customer relations.

  • logo francetv
  • logo sg
  • logo maif
  • logo sodexo
  • logo parc asterix
  • logo eurostar

Improve customer experience and optimize the performance of your agents.

check

20 years of expertise at your service

check

Billions of interactions managed each year

check

ISO 22301 certified for your peace of mind

check

Continuous innovation with our AI lab in Brittany

check

Declared electronic communications operator to ARCEP

hight performer
high performer emea
best support
best support
best support

4.7/5 on +200 reviews

logo sodexo
logo maif
logo sg
logo francetv

Contact our sales team.

Give us some personal information and we will contact you as soon as possible. See you very soon!

Improve customer experience and optimize the performance of your agents.

check

20 years of expertise at your service

check

Billions of interactions managed each year

check

ISO 22301 certified for your peace of mind

check

Continuous innovation with our AI lab in Brittany

check

Declared electronic communications operator to ARCEP

hight performer
high performer emea
best support
best support
best support

4.7/5 on +200 reviews

logo sodexo
logo maif
logo sg
logo francetv

Contact our sales team.

Give us some personal information and we will contact you as soon as possible. See you very soon!

Improve customer experience and optimize the performance of your agents.

check

20 years of expertise at your service

check

Billions of interactions managed each year

check

ISO 22301 certified for your peace of mind

check

Continuous innovation with our AI lab in Brittany

check

Declared electronic communications operator to ARCEP

hight performer
high performer emea
best support
best support
best support

4.7/5 on +200 reviews

logo sodexo
logo maif
logo sg
logo francetv

Contact our sales team.

Give us some personal information and we will contact you as soon as possible. See you very soon!

ViaDialog Logo

Products

Resources

Company

FR: +33 (0)1 77 45 30 00

152 Boulevard Pereire, 75017 Paris

DESIGNED
AND HOSTED
IN FRANCE

logo entreprise française
logo afaq iso22301
logo rgpd
high performer
high performer emea
meilleur soutien
Users love us
Users love us

Designed and hosted in Paris 🥐

Legal Notice
Privacy Policy
Terms and Conditions
Cookies
CSR Policy
English