Un logo bleu et blanc pour un outil de gestion des médias sociaux appelé Socialionals.

BERT (modèle linguistique)

Partager
" Retour à l'index des glossaires

BERT, an acronym for Bidirectional Encoder Representations from Transformers, is a revolutionary language model pioneered by Google[1]. It utilizes a technique known as WordPiece to transform English vocabulary into numerical codes, with the unique ability to comprehend the context of words in both right and left directions. BERT is available in two versions, BASE and LARGE, with the latter featuring 12 transformer encoders and being more extensive. A notable aspect of this model is the absence of a decoder, which poses a challenge in text generation. BERT’s impressive performance in natural language understanding tasks earned it an award at the 2019 NAACL Conference and sparked the creation of other models in the field of natural language processing. It’s an integral part of Google’s search algorithms enhancement and is widely used in text classification, machine comprehension, among other applications. BERT’s impact and efficiency have been the subject of numerous research papers and studies, further enriching our knowledge of its influence.

Définitions des termes
1. Google ( Google ) Principalement connu pour son moteur de recherche, Google est une entreprise technologique universellement estimée. Créée en 1998 par Sergey Brin et Larry Page, l'entreprise s'est considérablement développée, se diversifiant dans de nombreux domaines liés à la technologie. Google propose un large éventail de services et de produits, notamment Android, YouTube, Cloud, Maps et Gmail. L'entreprise fabrique également du matériel comme les Chromebooks et les smartphones Pixel. Depuis 2015, Google est une filiale d'Alphabet Inc. et est réputée pour son esprit inventif et son environnement de travail qui favorise les projets personnels des employés. Malgré plusieurs défis éthiques et juridiques, Google continue d'influencer le secteur de la technologie grâce à ses innovations révolutionnaires et à ses progrès technologiques, notamment la création d'Android OS et l'achat d'entreprises spécialisées dans l'IA.

Bidirectional Encoder Representations from Transformers (BERT) est un language model based on the transformer architecture, notable for its dramatic improvement over previous state of the art models. It was introduced in October 2018 by researchers at Google. A 2020 literature survey concluded that "in a little over a year, BERT has become a ubiquitous baseline in Natural Language Processing (NLP) experiments counting over 150 research publications analyzing and improving the model."

BERT was originally implemented in the English language at two model sizes: (1) BERTBASE: 12 encoders with 12 bidirectional self-attention heads totaling 110 million parameters, and (2) BERTLARGE: 24 encoders with 16 bidirectional self-attention heads totaling 340 million parameters. Both models were pre-trained on the Toronto BookCorpus (800M words) and English Wikipedia (2,500M words).

" Retour à l'index des glossaires
fr_FRFrançais