GPT-J Model
(Redirected from GPT-J)
Jump to navigation
Jump to search
A GPT-J Model is a LNLM.
- Example(s):
- GPT-J-6B.
- …
- Counter-Example(s):
- See: ChatGPT, Language Model, EleutherAI, Open-Source Software.
References
2023
- (Wikipedia, 2023) ⇒ https://en.wikipedia.org/wiki/GPT-J Retrieved:2023-3-13.
- GPT-J is an open source artificial intelligence language model developed by EleutherAI.[1] GPT-J performs very similarly to OpenAI's GPT-3 on various zero-shot down-streaming tasks and can even outperform it on code generation tasks.[2] The newest version, GPT-J-6B is a language model based on a data set called The Pile.[3] The Pile is an open-source 825 gigibyte language modelling data set that is split into 22 smaller datasets.[4] GPT-J is similar to ChatGPT in ability, although it does not function as a chat bot, only as a text predictor.[5]
- ↑ Demo, GPT-3. "GPT-J | Discover AI use cases" (in en). https://gpt3demo.com/apps/gpt-j-6b. Retrieved 2023-02-28.
- ↑ "GPT-J-6B: An Introduction to the Largest Open Source GPT Model | Forefront" (in en). https://www.forefront.ai/blog-posts/gpt-j-6b-an-introduction-to-the-largest-open-sourced-gpt-model. Retrieved 2023-02-28.
- ↑ Template:Citation
- ↑ "The Pile". https://pile.eleuther.ai/. Retrieved 2023-02-28.
- ↑ Mueller, Vincent (2022-01-25). "How you can use GPT-J" (in en). https://towardsdatascience.com/how-you-can-use-gpt-j-9c4299dd8526. Retrieved 2023-02-28.