Encoder-only Transformer-based Large Language Model (LLM)

From GM-RKB
Jump to navigation Jump to search

An Encoder-only Transformer-based Large Language Model (LLM) is a transformer-based LLM that solely employs an encoder architecture.



References

2023

  • chat
    • Encoder-only models focus on encoding input text into meaningful contextualized representations. These models are particularly well-suited for tasks where the main goal is to understand and extract information from the input text. BERT (Bidirectional Encoder Representations from Transformers) is a prominent example of an encoder-only LLM. BERT uses the Transformer's encoder layers to pretrain deep bidirectional representations, capturing the context from both left and right directions. This allows BERT to excel in tasks like sentence classification, named entity recognition, and question-answering.