Large language model
Item
- Titre (article, livre complet, chapitre de livre, publication réseaux sociaux, pageWeb, SiteWeb complet, etc.)
- Large language model
- Auteur de la référence bibliograpique
- Wikipédia
- Date de publication (AAAA ou AAAA-MM-JJ)
- s.d.
- Titre du livre (pour chapitre de livre)
- Wikipedia
- Is Referenced By
- 99DHLCSI
- Abstract
-
A large language model (LLM) is a type of computational model designed for natural language processing tasks such as language generation. As language models, LLMs acquire these abilities by learning statistical relationships from vast amounts of text during a self-supervised and semi-supervised training process.
The largest and most capable LLMs are artificial neural networks built with a decoder-only transformer-based architecture, enabling efficient processing and generation of large-scale text data. Modern models can be fine-tuned for specific tasks, or be guided by prompt engineering. These models acquire predictive power regarding syntax, semantics, and ontologies inherent in human language corpora, but they also inherit inaccuracies and biases present in the data on which they are trained. - Language
- en
- Source
- Wikipedia
- Rights
- Creative Commons Attribution-ShareAlike License
- Date Submitted
- 2024-11-05T14:24:46Z
Linked resources
Title | Class |
---|---|
La narration générative entre IA symbolique et connexionniste : une expérience transnationale | Conference |
Annotations
There are no annotations for this resource.