Encoder-only Neural Network: Difference between revisions
Jump to navigation
Jump to search
No edit summary |
m (Text replacement - ".↵----" to ". ----") |
||
Line 14: | Line 14: | ||
** [[Encoder-Decoder Neural Network Model]]. | ** [[Encoder-Decoder Neural Network Model]]. | ||
* <B>See:</B> [[Natural Language Processing]], [[Bidirectional Training]], [[Language Understanding]], [[Pre-trained Models]]. | * <B>See:</B> [[Natural Language Processing]], [[Bidirectional Training]], [[Language Understanding]], [[Pre-trained Models]]. | ||
---- | ---- | ||
---- | ---- | ||
[[Category:Concept]] | [[Category:Concept]] |
Latest revision as of 02:47, 28 November 2024
A Encoder-only Neural Network is a neural network model that exclusively uses an encoder architecture.
- Context:
- It can (typically) employ a bidirectional approach to process and understand natural language.
- It can be used in various natural language processing tasks, like language understanding and language representation.
- It can (often) be pre-trained on large corpora to learn word and sentence representations and their semantic relations.
- It can lack a decoder component, which restricts its ability in tasks like text generation and language generation.
- It can be fine-tuned on smaller datasets for specific tasks, such as sentiment analysis or question answering.
- Example(s):
- Counter-Example(s):
- See: Natural Language Processing, Bidirectional Training, Language Understanding, Pre-trained Models.