• Thumbnail for GPT-2
    Pre-trained Transformer 2 (GPT-2) is a large language model by OpenAI and the second in their foundational series of GPT models. GPT-2 was pre-trained on a...
    44 KB (3,260 words) - 23:33, 12 August 2024
  • Pre-trained Transformer 4 (GPT-4) is a multimodal large language model created by OpenAI, and the fourth in its series of GPT foundation models. It was...
    61 KB (5,899 words) - 17:49, 28 August 2024
  • GPT-4o (GPT-4 Omni) is a multilingual, multimodal generative pre-trained transformer designed by OpenAI. It was announced by OpenAI's CTO Mira Murati...
    17 KB (1,804 words) - 22:23, 10 September 2024
  • Generative Pre-trained Transformer 3 (GPT-3) is a large language model released by OpenAI in 2020. Like its predecessor, GPT-2, it is a decoder-only transformer...
    54 KB (4,913 words) - 21:06, 9 September 2024
  • OpenAI (section GPT-2)
    the GPT family of large language models, the DALL-E series of text-to-image models, and a text-to-video model named Sora. Its release of ChatGPT in November...
    187 KB (16,222 words) - 00:09, 14 September 2024
  • Thumbnail for Generative pre-trained transformer
    Generative pre-trained transformers (GPTs) are a type of large language model (LLM) and a prominent framework for generative artificial intelligence. They...
    47 KB (4,121 words) - 04:02, 6 September 2024
  • Thumbnail for ChatGPT
    ChatGPT is a chatbot and virtual assistant developed by OpenAI and launched on November 30, 2022. Based on large language models (LLMs), it enables users...
    197 KB (16,990 words) - 01:56, 14 September 2024
  • AI Dungeon (redirect from AI Dungeon 2)
    public in May 2019. It is not to be confused with another GPT-2-based adventure game, GPT Adventure, created by Northwestern University neuroscience...
    42 KB (3,654 words) - 02:02, 8 September 2024
  • Thumbnail for GPT-1
    Generative Pre-trained Transformer 1 (GPT-1) was the first of OpenAI's large language models following Google's invention of the transformer architecture...
    32 KB (1,064 words) - 15:45, 8 May 2024
  • trained on. Some notable LLMs are OpenAI's GPT series of models (e.g., GPT-3.5, GPT-4 and GPT-4o; used in ChatGPT and Microsoft Copilot), Google's Gemini...
    156 KB (13,419 words) - 20:33, 13 September 2024
  • Thumbnail for GPT-J
    GPT-J or GPT-J-6B is an open-source large language model (LLM) developed by EleutherAI in 2021. As the name suggests, it is a generative pre-trained transformer...
    11 KB (982 words) - 13:02, 5 September 2024
  • Thumbnail for ChatGPT in education
    Since the public release of ChatGPT by OpenAI in November 2022, the integration of chatbots in education has sparked considerable debate and exploration...
    29 KB (3,003 words) - 23:33, 10 September 2024
  • Thumbnail for DALL-E
    DALL-E (redirect from DALL-E 2)
    following year, its successor DALL-E 2 was released. DALL·E 3 was released natively into ChatGPT for ChatGPT Plus and ChatGPT Enterprise customers in October...
    52 KB (3,959 words) - 08:44, 7 September 2024
  • Thumbnail for GUID Partition Table
    The GUID Partition Table (GPT) is a standard for the layout of partition tables of a physical computer storage device, such as a hard disk drive or solid-state...
    79 KB (2,960 words) - 15:39, 19 August 2024
  • Thumbnail for Greg Brockman
    OpenAI Five, a Dota 2 bot. On February 14, 2019, OpenAI announced that they had developed a new large language model called GPT-2, but kept it private...
    14 KB (1,152 words) - 09:54, 10 September 2024
  • several subsequent state-of-the-art models in NLP, including BERT, GPT-2, and GPT-3. "Ashish Vaswani". scholar.google.com. Retrieved 2023-07-11. Vaswani...
    7 KB (551 words) - 19:57, 2 July 2024
  • latent representations of tokens in their context, similar to ELMo and GPT-2. It found applications for many many natural language processing tasks,...
    29 KB (3,310 words) - 06:04, 14 September 2024
  • Gandhe, Ankur; Gadde, Ravi Teja; Kirchhoff, Katrin (2021). "Prompt Tuning GPT-2 language model for parameter-efficient domain adaptation of ASR systems"...
    13 KB (1,370 words) - 20:36, 22 July 2024
  • Thumbnail for Generative artificial intelligence
    first generative pre-trained transformer (GPT), known as GPT-1, in 2018. This was followed in 2019 by GPT-2 which demonstrated the ability to generalize...
    135 KB (11,873 words) - 18:43, 13 September 2024
  • noteworthy foundation models include: GPT, BERT, GPT-2, T5, GPT-3, CLIP, DALL-E, Stable Diffusion, GPT-4, LLaMA, LLaMA 2, and Mistral. Each of these models...
    46 KB (5,072 words) - 13:26, 5 September 2024
  • learning models. In 2020, he revealed that large language models, like GPT-2, could memorize and output personally identifiable information. His research...
    14 KB (1,304 words) - 18:08, 8 September 2024
  • Thumbnail for Ġ
    can be expected to display correctly on most computer systems. OpenAI's GPT-2 uses U+0120 (Ġ) as a substitute for the space character in its tokens. Koryakov...
    5 KB (428 words) - 14:19, 8 August 2024
  • Thumbnail for Seq2seq
    Seq2seq (redirect from Seq-2-seq)
    times greater model capacity than OpenAI's GPT-2, whose May 2020 successor, the 175 billion parameter GPT-3, trained on a "45TB dataset of plaintext words...
    13 KB (1,429 words) - 17:44, 15 August 2024
  • investigate the scaling laws of large language models. It claimed to outperform GPT-3. It considerably simplifies downstream utilization because it requires...
    7 KB (610 words) - 05:48, 14 September 2024
  • Thumbnail for Transformer (deep learning architecture)
    (NLP). Many large language models such as GPT-2, GPT-3, GPT-4, AlbertAGPT, Claude, BERT, XLNet, RoBERTa and ChatGPT demonstrate the ability of transformers...
    99 KB (12,309 words) - 06:04, 14 September 2024
  • Thumbnail for Llama.cpp
    bfloat16; and 1.56 bit quantization. LLaMA Llama 2 Llama 3 Mistral 7B Mixtral 8x7B Mixtral 8x22B DBRX GPT-2 BLOOM Gemma Grok-1 Mamba "Initial release · ggerganov/llama...
    14 KB (1,104 words) - 23:52, 8 September 2024
  • libraries and includes implementations of notable models like BERT and GPT-2. The library was originally called "pytorch-pretrained-bert" which was then...
    14 KB (1,062 words) - 01:57, 6 August 2024
  • YandexGPT is a neural network of the GPT family developed by the Russian company Yandex LLC. YandexGPT can create and revise texts, generate new ideas...
    6 KB (618 words) - 03:38, 28 July 2024
  • Thumbnail for GPT Group
    The GPT Group is a Real Estate Investment Trust (also known as an Australian Stock Exchange listed stapled entity). The Group has been publicly listed...
    4 KB (309 words) - 00:53, 2 September 2024
  • Thumbnail for Gemini (language model)
    was announced on December 6, 2023, positioned as a competitor to OpenAI's GPT-4. It powers the chatbot of the same name. Google announced Gemini, a large...
    43 KB (3,464 words) - 23:58, 6 September 2024