• Thumbnail for GPT-2
    Pre-trained Transformer 2 (GPT-2) is a large language model by OpenAI and the second in their foundational series of GPT models. GPT-2 was pre-trained on a...
    44 KB (3,260 words) - 10:30, 25 July 2024
  • Pre-trained Transformer 4 (GPT-4) is a multimodal large language model created by OpenAI, and the fourth in its series of GPT foundation models. It was...
    61 KB (5,893 words) - 14:18, 26 July 2024
  • GPT-4o (GPT-4 Omni) is a multilingual, multimodal generative pre-trained transformer designed by OpenAI. It was announced by OpenAI's CTO Mira Murati...
    14 KB (1,536 words) - 19:37, 26 July 2024
  • Generative Pre-trained Transformer 3 (GPT-3) is a large language model released by OpenAI in 2020. Like its predecessor, GPT-2, it is a decoder-only transformer...
    54 KB (4,914 words) - 19:17, 6 July 2024
  • Thumbnail for ChatGPT
    ChatGPT is a chatbot and virtual assistant developed by OpenAI and launched on November 30, 2022. Based on large language models (LLMs), it enables users...
    191 KB (16,489 words) - 17:38, 8 August 2024
  • OpenAI (section GPT-2)
    the GPT family of large language models, the DALL-E series of text-to-image models, and a text-to-video model named Sora. Its release of ChatGPT in November...
    186 KB (16,170 words) - 23:59, 9 August 2024
  • Thumbnail for Generative pre-trained transformer
    Generative pre-trained transformers (GPT) are a type of large language model (LLM) and a prominent framework for generative artificial intelligence. They...
    46 KB (4,103 words) - 04:27, 8 August 2024
  • Thumbnail for GPT-1
    Generative Pre-trained Transformer 1 (GPT-1) was the first of OpenAI's large language models following Google's invention of the transformer architecture...
    32 KB (1,064 words) - 15:45, 8 May 2024
  • AI Dungeon (redirect from AI Dungeon 2)
    public in May 2019. It is not to be confused with another GPT-2-based adventure game, GPT Adventure, created by Northwestern University neuroscience...
    42 KB (3,654 words) - 21:36, 1 August 2024
  • trained on. Some notable LLMs are OpenAI's GPT series of models (e.g., GPT-3.5, GPT-4 and GPT-4o; used in ChatGPT and Microsoft Copilot), Google's Gemini...
    137 KB (12,402 words) - 23:00, 9 August 2024
  • Thumbnail for GPT-J
    GPT-J or GPT-J-6B is an open-source large language model (LLM) developed by EleutherAI in 2021. As the name suggests, it is a generative pre-trained transformer...
    11 KB (981 words) - 02:56, 11 June 2024
  • Thumbnail for DALL-E
    DALL-E (redirect from DALL-E 2)
    following year, its successor DALL-E 2 was released. DALL·E 3 was released natively into ChatGPT for ChatGPT Plus and ChatGPT Enterprise customers in October...
    66 KB (4,242 words) - 18:13, 25 July 2024
  • Thumbnail for GUID Partition Table
    The GUID Partition Table (GPT) is a standard for the layout of partition tables of a physical computer storage device, such as a hard disk drive or solid-state...
    79 KB (2,953 words) - 09:09, 3 August 2024
  • Thumbnail for ChatGPT in education
    Since the public release of ChatGPT by OpenAI in November 2022, the integration of chatbots in education has sparked considerable debate and exploration...
    29 KB (3,038 words) - 17:51, 8 August 2024
  • Thumbnail for Greg Brockman
    OpenAI Five, a Dota 2 bot. On February 14, 2019, OpenAI announced that they had developed a new large language model called GPT-2, but kept it private...
    13 KB (1,114 words) - 13:41, 3 August 2024
  • latent representations of tokens in their context, similar to ELMo and GPT-2. It found applications for many many natural language processing tasks,...
    25 KB (2,821 words) - 08:57, 8 August 2024
  • Thumbnail for Generative artificial intelligence
    first generative pre-trained transformer (GPT), known as GPT-1, in 2018. This was followed in 2019 by GPT-2 which demonstrated the ability to generalize...
    119 KB (10,461 words) - 22:40, 4 August 2024
  • several subsequent state-of-the-art models in NLP, including BERT, GPT-2, and GPT-3. "Ashish Vaswani". scholar.google.com. Retrieved 2023-07-11. Vaswani...
    7 KB (551 words) - 19:57, 2 July 2024
  • noteworthy foundation models include: GPT, BERT, GPT-2, T5, GPT-3, CLIP, DALL-E, Stable Diffusion, GPT-4, LLaMA, LLaMA 2, and Mistral. Each of these models...
    46 KB (5,056 words) - 19:41, 24 July 2024
  • Thumbnail for Seq2seq
    Seq2seq (redirect from Seq-2-seq)
    times greater model capacity than OpenAI's GPT-2, whose May 2020 successor, the 175 billion parameter GPT-3, trained on a "45TB dataset of plaintext words...
    13 KB (1,429 words) - 07:14, 9 August 2024
  • Thumbnail for GPT Group
    The GPT Group is a Real Estate Investment Trust (also known as an Australian Stock Exchange listed stapled entity). The Group has been publicly listed...
    4 KB (309 words) - 13:32, 4 July 2024
  • Gandhe, Ankur; Gadde, Ravi Teja; Kirchhoff, Katrin (2021). "Prompt Tuning GPT-2 language model for parameter-efficient domain adaptation of ASR systems"...
    13 KB (1,370 words) - 20:36, 22 July 2024
  • investigate the scaling laws of large language models. It claimed to outperform GPT-3. It considerably simplifies downstream utilization because it requires...
    7 KB (548 words) - 21:42, 7 August 2024
  • libraries and includes implementations of notable models like BERT and GPT-2. The library was originally called "pytorch-pretrained-bert" which was then...
    14 KB (1,062 words) - 01:57, 6 August 2024
  • Thumbnail for Ġ
    can be expected to display correctly on most computer systems. OpenAI's GPT-2 uses U+0120 (Ġ) as a substitute for the space character in its tokens. Koryakov...
    5 KB (428 words) - 14:19, 8 August 2024
  • Thumbnail for Llama.cpp
    well as 1.5-bit and 2-bit to 8-bit quantized integer types. LLaMA Llama 2 Llama 3 Mistral 7B Mixtral 8x7B Mixtral 8x22B DBRX GPT-2 BLOOM Gemma Grok-1 Mamba...
    12 KB (954 words) - 04:44, 9 August 2024
  • YandexGPT is a neural network of the GPT family developed by the Russian company Yandex LLC. YandexGPT can create and revise texts, generate new ideas...
    6 KB (618 words) - 03:38, 28 July 2024
  • Thumbnail for Gemini (language model)
    was announced on December 6, 2023, positioned as a competitor to OpenAI's GPT-4. It powers the chatbot of the same name. Google announced Gemini, a large...
    42 KB (3,324 words) - 09:10, 2 August 2024
  • Thumbnail for Transformer (deep learning architecture)
    prediction. Many large language models such as GPT-2, GPT-3, GPT-4, Claude, BERT, XLNet, RoBERTa and ChatGPT demonstrate the ability of transformers to perform...
    91 KB (11,401 words) - 18:53, 9 August 2024
  • as GPT-3, a focus of research was up-scaling models which in some instances showed major increases in emergent capabilities. The release of ChatGPT and...
    33 KB (3,395 words) - 05:41, 9 August 2024