Encoder-only Transformer-based Large Language Model (LLM)

From GM-RKB
Revision as of 07:44, 3 February 2024 by Gmelli (talk | contribs) (Created page with "An Encoder-only Transformer-based Large Language Model (LLM) is a transformer-based LLM that solely employs an encoder architecture. * <B>Context:</B> ** It can focus on encoding input text into Meaningful Representation, Contextualized Representations. ** It can support NLU Tasks. ** It can use pretraining methods like masked language modeling and next-sentence prediction. ** It can be fine-tuned on smaller datasets for specific tasks suc...")
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

An Encoder-only Transformer-based Large Language Model (LLM) is a transformer-based LLM that solely employs an encoder architecture.



References

2023

  • chat
    • Encoder-only models focus on encoding input text into meaningful contextualized representations. These models are particularly well-suited for tasks where the main goal is to understand and extract information from the input text. BERT (Bidirectional Encoder Representations from Transformers) is a prominent example of an encoder-only LLM. BERT uses the Transformer's encoder layers to pretrain deep bidirectional representations, capturing the context from both left and right directions. This allows BERT to excel in tasks like sentence classification, named entity recognition, and question-answering.