CINXE.COM

Generative pre-trained transformer - Wikipedia

<!DOCTYPE html> <html class="client-nojs vector-feature-language-in-header-enabled vector-feature-language-in-main-page-header-disabled vector-feature-page-tools-pinned-disabled vector-feature-toc-pinned-clientpref-1 vector-feature-main-menu-pinned-disabled vector-feature-limited-width-clientpref-1 vector-feature-limited-width-content-enabled vector-feature-custom-font-size-clientpref-1 vector-feature-appearance-pinned-clientpref-1 vector-feature-night-mode-enabled skin-theme-clientpref-day vector-sticky-header-enabled vector-toc-available" lang="en" dir="ltr"> <head> <meta charset="UTF-8"> <title>Generative pre-trained transformer - Wikipedia</title> <script>(function(){var className="client-js vector-feature-language-in-header-enabled vector-feature-language-in-main-page-header-disabled vector-feature-page-tools-pinned-disabled vector-feature-toc-pinned-clientpref-1 vector-feature-main-menu-pinned-disabled vector-feature-limited-width-clientpref-1 vector-feature-limited-width-content-enabled vector-feature-custom-font-size-clientpref-1 vector-feature-appearance-pinned-clientpref-1 vector-feature-night-mode-enabled skin-theme-clientpref-day vector-sticky-header-enabled vector-toc-available";var cookie=document.cookie.match(/(?:^|; )enwikimwclientpreferences=([^;]+)/);if(cookie){cookie[1].split('%2C').forEach(function(pref){className=className.replace(new RegExp('(^| )'+pref.replace(/-clientpref-\w+$|[^\w-]+/g,'')+'-clientpref-\\w+( |$)'),'$1'+pref+'$2');});}document.documentElement.className=className;}());RLCONF={"wgBreakFrames":false,"wgSeparatorTransformTable":["",""],"wgDigitTransformTable":["",""],"wgDefaultDateFormat":"dmy", "wgMonthNames":["","January","February","March","April","May","June","July","August","September","October","November","December"],"wgRequestId":"2292b6ef-5e27-4098-b727-eba99393a477","wgCanonicalNamespace":"","wgCanonicalSpecialPageName":false,"wgNamespaceNumber":0,"wgPageName":"Generative_pre-trained_transformer","wgTitle":"Generative pre-trained transformer","wgCurRevisionId":1275868082,"wgRevisionId":1275868082,"wgArticleId":72970020,"wgIsArticle":true,"wgIsRedirect":false,"wgAction":"view","wgUserName":null,"wgUserGroups":["*"],"wgCategories":["CS1: long volume value","Articles with short description","Short description matches Wikidata","All articles with unsourced statements","Articles with unsourced statements from December 2024","Articles with unsourced statements from November 2023","Large language models","Generative artificial intelligence","Artificial neural networks","Generative pre-trained transformers","OpenAI"],"wgPageViewLanguage":"en","wgPageContentLanguage":"en", "wgPageContentModel":"wikitext","wgRelevantPageName":"Generative_pre-trained_transformer","wgRelevantArticleId":72970020,"wgIsProbablyEditable":true,"wgRelevantPageIsProbablyEditable":true,"wgRestrictionEdit":[],"wgRestrictionMove":[],"wgNoticeProject":"wikipedia","wgCiteReferencePreviewsActive":false,"wgFlaggedRevsParams":{"tags":{"status":{"levels":1}}},"wgMediaViewerOnClick":true,"wgMediaViewerEnabledByDefault":true,"wgPopupsFlags":0,"wgVisualEditor":{"pageLanguageCode":"en","pageLanguageDir":"ltr","pageVariantFallbacks":"en"},"wgMFDisplayWikibaseDescriptions":{"search":true,"watchlist":true,"tagline":false,"nearby":true},"wgWMESchemaEditAttemptStepOversample":false,"wgWMEPageLength":50000,"wgEditSubmitButtonLabelPublish":true,"wgULSPosition":"interlanguage","wgULSisCompactLinksEnabled":false,"wgVector2022LanguageInHeader":true,"wgULSisLanguageSelectorEmpty":false,"wgWikibaseItemId":"Q116777014","wgCheckUserClientHintsHeadersJsApi":["brands","architecture","bitness", "fullVersionList","mobile","model","platform","platformVersion"],"GEHomepageSuggestedEditsEnableTopics":true,"wgGETopicsMatchModeEnabled":false,"wgGEStructuredTaskRejectionReasonTextInputEnabled":false,"wgGELevelingUpEnabledForUser":false};RLSTATE={"ext.globalCssJs.user.styles":"ready","site.styles":"ready","user.styles":"ready","ext.globalCssJs.user":"ready","user":"ready","user.options":"loading","ext.cite.styles":"ready","skins.vector.search.codex.styles":"ready","skins.vector.styles":"ready","skins.vector.icons":"ready","jquery.makeCollapsible.styles":"ready","ext.wikimediamessages.styles":"ready","ext.visualEditor.desktopArticleTarget.noscript":"ready","ext.uls.interlanguage":"ready","wikibase.client.init":"ready","ext.wikimediaBadges":"ready"};RLPAGEMODULES=["ext.cite.ux-enhancements","mediawiki.page.media","site","mediawiki.page.ready","jquery.makeCollapsible","mediawiki.toc","skins.vector.js","ext.centralNotice.geoIP","ext.centralNotice.startUp","ext.gadget.ReferenceTooltips", "ext.gadget.switcher","ext.urlShortener.toolbar","ext.centralauth.centralautologin","mmv.bootstrap","ext.popups","ext.visualEditor.desktopArticleTarget.init","ext.visualEditor.targetLoader","ext.echo.centralauth","ext.eventLogging","ext.wikimediaEvents","ext.navigationTiming","ext.uls.interface","ext.cx.eventlogging.campaigns","ext.cx.uls.quick.actions","wikibase.client.vector-2022","ext.checkUser.clientHints","ext.growthExperiments.SuggestedEditSession"];</script> <script>(RLQ=window.RLQ||[]).push(function(){mw.loader.impl(function(){return["user.options@12s5i",function($,jQuery,require,module){mw.user.tokens.set({"patrolToken":"+\\","watchToken":"+\\","csrfToken":"+\\"}); }];});});</script> <link rel="stylesheet" href="/w/load.php?lang=en&amp;modules=ext.cite.styles%7Cext.uls.interlanguage%7Cext.visualEditor.desktopArticleTarget.noscript%7Cext.wikimediaBadges%7Cext.wikimediamessages.styles%7Cjquery.makeCollapsible.styles%7Cskins.vector.icons%2Cstyles%7Cskins.vector.search.codex.styles%7Cwikibase.client.init&amp;only=styles&amp;skin=vector-2022"> <script async="" src="/w/load.php?lang=en&amp;modules=startup&amp;only=scripts&amp;raw=1&amp;skin=vector-2022"></script> <meta name="ResourceLoaderDynamicStyles" content=""> <link rel="stylesheet" href="/w/load.php?lang=en&amp;modules=site.styles&amp;only=styles&amp;skin=vector-2022"> <meta name="generator" content="MediaWiki 1.44.0-wmf.17"> <meta name="referrer" content="origin"> <meta name="referrer" content="origin-when-cross-origin"> <meta name="robots" content="max-image-preview:standard"> <meta name="format-detection" content="telephone=no"> <meta property="og:image" content="https://upload.wikimedia.org/wikipedia/commons/thumb/5/51/Full_GPT_architecture.svg/1200px-Full_GPT_architecture.svg.png"> <meta property="og:image:width" content="1200"> <meta property="og:image:height" content="1440"> <meta property="og:image" content="https://upload.wikimedia.org/wikipedia/commons/thumb/5/51/Full_GPT_architecture.svg/800px-Full_GPT_architecture.svg.png"> <meta property="og:image:width" content="800"> <meta property="og:image:height" content="960"> <meta property="og:image" content="https://upload.wikimedia.org/wikipedia/commons/thumb/5/51/Full_GPT_architecture.svg/640px-Full_GPT_architecture.svg.png"> <meta property="og:image:width" content="640"> <meta property="og:image:height" content="768"> <meta name="viewport" content="width=1120"> <meta property="og:title" content="Generative pre-trained transformer - Wikipedia"> <meta property="og:type" content="website"> <link rel="preconnect" href="//upload.wikimedia.org"> <link rel="alternate" media="only screen and (max-width: 640px)" href="//en.m.wikipedia.org/wiki/Generative_pre-trained_transformer"> <link rel="alternate" type="application/x-wiki" title="Edit this page" href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit"> <link rel="apple-touch-icon" href="/static/apple-touch/wikipedia.png"> <link rel="icon" href="/static/favicon/wikipedia.ico"> <link rel="search" type="application/opensearchdescription+xml" href="/w/rest.php/v1/search" title="Wikipedia (en)"> <link rel="EditURI" type="application/rsd+xml" href="//en.wikipedia.org/w/api.php?action=rsd"> <link rel="canonical" href="https://en.wikipedia.org/wiki/Generative_pre-trained_transformer"> <link rel="license" href="https://creativecommons.org/licenses/by-sa/4.0/deed.en"> <link rel="alternate" type="application/atom+xml" title="Wikipedia Atom feed" href="/w/index.php?title=Special:RecentChanges&amp;feed=atom"> <link rel="dns-prefetch" href="//meta.wikimedia.org" /> <link rel="dns-prefetch" href="login.wikimedia.org"> </head> <body class="skin--responsive skin-vector skin-vector-search-vue mediawiki ltr sitedir-ltr mw-hide-empty-elt ns-0 ns-subject mw-editable page-Generative_pre-trained_transformer rootpage-Generative_pre-trained_transformer skin-vector-2022 action-view"><a class="mw-jump-link" href="#bodyContent">Jump to content</a> <div class="vector-header-container"> <header class="vector-header mw-header"> <div class="vector-header-start"> <nav class="vector-main-menu-landmark" aria-label="Site"> <div id="vector-main-menu-dropdown" class="vector-dropdown vector-main-menu-dropdown vector-button-flush-left vector-button-flush-right" title="Main menu" > <input type="checkbox" id="vector-main-menu-dropdown-checkbox" role="button" aria-haspopup="true" data-event-name="ui.dropdown-vector-main-menu-dropdown" class="vector-dropdown-checkbox " aria-label="Main menu" > <label id="vector-main-menu-dropdown-label" for="vector-main-menu-dropdown-checkbox" class="vector-dropdown-label cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only " aria-hidden="true" ><span class="vector-icon mw-ui-icon-menu mw-ui-icon-wikimedia-menu"></span> <span class="vector-dropdown-label-text">Main menu</span> </label> <div class="vector-dropdown-content"> <div id="vector-main-menu-unpinned-container" class="vector-unpinned-container"> <div id="vector-main-menu" class="vector-main-menu vector-pinnable-element"> <div class="vector-pinnable-header vector-main-menu-pinnable-header vector-pinnable-header-unpinned" data-feature-name="main-menu-pinned" data-pinnable-element-id="vector-main-menu" data-pinned-container-id="vector-main-menu-pinned-container" data-unpinned-container-id="vector-main-menu-unpinned-container" > <div class="vector-pinnable-header-label">Main menu</div> <button class="vector-pinnable-header-toggle-button vector-pinnable-header-pin-button" data-event-name="pinnable-header.vector-main-menu.pin">move to sidebar</button> <button class="vector-pinnable-header-toggle-button vector-pinnable-header-unpin-button" data-event-name="pinnable-header.vector-main-menu.unpin">hide</button> </div> <div id="p-navigation" class="vector-menu mw-portlet mw-portlet-navigation" > <div class="vector-menu-heading"> Navigation </div> <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li id="n-mainpage-description" class="mw-list-item"><a href="/wiki/Main_Page" title="Visit the main page [z]" accesskey="z"><span>Main page</span></a></li><li id="n-contents" class="mw-list-item"><a href="/wiki/Wikipedia:Contents" title="Guides to browsing Wikipedia"><span>Contents</span></a></li><li id="n-currentevents" class="mw-list-item"><a href="/wiki/Portal:Current_events" title="Articles related to current events"><span>Current events</span></a></li><li id="n-randompage" class="mw-list-item"><a href="/wiki/Special:Random" title="Visit a randomly selected article [x]" accesskey="x"><span>Random article</span></a></li><li id="n-aboutsite" class="mw-list-item"><a href="/wiki/Wikipedia:About" title="Learn about Wikipedia and how it works"><span>About Wikipedia</span></a></li><li id="n-contactpage" class="mw-list-item"><a href="//en.wikipedia.org/wiki/Wikipedia:Contact_us" title="How to contact Wikipedia"><span>Contact us</span></a></li> </ul> </div> </div> <div id="p-interaction" class="vector-menu mw-portlet mw-portlet-interaction" > <div class="vector-menu-heading"> Contribute </div> <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li id="n-help" class="mw-list-item"><a href="/wiki/Help:Contents" title="Guidance on how to use and edit Wikipedia"><span>Help</span></a></li><li id="n-introduction" class="mw-list-item"><a href="/wiki/Help:Introduction" title="Learn how to edit Wikipedia"><span>Learn to edit</span></a></li><li id="n-portal" class="mw-list-item"><a href="/wiki/Wikipedia:Community_portal" title="The hub for editors"><span>Community portal</span></a></li><li id="n-recentchanges" class="mw-list-item"><a href="/wiki/Special:RecentChanges" title="A list of recent changes to Wikipedia [r]" accesskey="r"><span>Recent changes</span></a></li><li id="n-upload" class="mw-list-item"><a href="/wiki/Wikipedia:File_upload_wizard" title="Add images or other media for use on Wikipedia"><span>Upload file</span></a></li><li id="n-specialpages" class="mw-list-item"><a href="/wiki/Special:SpecialPages"><span>Special pages</span></a></li> </ul> </div> </div> </div> </div> </div> </div> </nav> <a href="/wiki/Main_Page" class="mw-logo"> <img class="mw-logo-icon" src="/static/images/icons/wikipedia.png" alt="" aria-hidden="true" height="50" width="50"> <span class="mw-logo-container skin-invert"> <img class="mw-logo-wordmark" alt="Wikipedia" src="/static/images/mobile/copyright/wikipedia-wordmark-en.svg" style="width: 7.5em; height: 1.125em;"> <img class="mw-logo-tagline" alt="The Free Encyclopedia" src="/static/images/mobile/copyright/wikipedia-tagline-en.svg" width="117" height="13" style="width: 7.3125em; height: 0.8125em;"> </span> </a> </div> <div class="vector-header-end"> <div id="p-search" role="search" class="vector-search-box-vue vector-search-box-collapses vector-search-box-show-thumbnail vector-search-box-auto-expand-width vector-search-box"> <a href="/wiki/Special:Search" class="cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only search-toggle" title="Search Wikipedia [f]" accesskey="f"><span class="vector-icon mw-ui-icon-search mw-ui-icon-wikimedia-search"></span> <span>Search</span> </a> <div class="vector-typeahead-search-container"> <div class="cdx-typeahead-search cdx-typeahead-search--show-thumbnail cdx-typeahead-search--auto-expand-width"> <form action="/w/index.php" id="searchform" class="cdx-search-input cdx-search-input--has-end-button"> <div id="simpleSearch" class="cdx-search-input__input-wrapper" data-search-loc="header-moved"> <div class="cdx-text-input cdx-text-input--has-start-icon"> <input class="cdx-text-input__input" type="search" name="search" placeholder="Search Wikipedia" aria-label="Search Wikipedia" autocapitalize="sentences" title="Search Wikipedia [f]" accesskey="f" id="searchInput" > <span class="cdx-text-input__icon cdx-text-input__start-icon"></span> </div> <input type="hidden" name="title" value="Special:Search"> </div> <button class="cdx-button cdx-search-input__end-button">Search</button> </form> </div> </div> </div> <nav class="vector-user-links vector-user-links-wide" aria-label="Personal tools"> <div class="vector-user-links-main"> <div id="p-vector-user-menu-preferences" class="vector-menu mw-portlet emptyPortlet" > <div class="vector-menu-content"> <ul class="vector-menu-content-list"> </ul> </div> </div> <div id="p-vector-user-menu-userpage" class="vector-menu mw-portlet emptyPortlet" > <div class="vector-menu-content"> <ul class="vector-menu-content-list"> </ul> </div> </div> <nav class="vector-appearance-landmark" aria-label="Appearance"> <div id="vector-appearance-dropdown" class="vector-dropdown " title="Change the appearance of the page&#039;s font size, width, and color" > <input type="checkbox" id="vector-appearance-dropdown-checkbox" role="button" aria-haspopup="true" data-event-name="ui.dropdown-vector-appearance-dropdown" class="vector-dropdown-checkbox " aria-label="Appearance" > <label id="vector-appearance-dropdown-label" for="vector-appearance-dropdown-checkbox" class="vector-dropdown-label cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only " aria-hidden="true" ><span class="vector-icon mw-ui-icon-appearance mw-ui-icon-wikimedia-appearance"></span> <span class="vector-dropdown-label-text">Appearance</span> </label> <div class="vector-dropdown-content"> <div id="vector-appearance-unpinned-container" class="vector-unpinned-container"> </div> </div> </div> </nav> <div id="p-vector-user-menu-notifications" class="vector-menu mw-portlet emptyPortlet" > <div class="vector-menu-content"> <ul class="vector-menu-content-list"> </ul> </div> </div> <div id="p-vector-user-menu-overflow" class="vector-menu mw-portlet" > <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li id="pt-sitesupport-2" class="user-links-collapsible-item mw-list-item user-links-collapsible-item"><a data-mw="interface" href="https://donate.wikimedia.org/?wmf_source=donate&amp;wmf_medium=sidebar&amp;wmf_campaign=en.wikipedia.org&amp;uselang=en" class=""><span>Donate</span></a> </li> <li id="pt-createaccount-2" class="user-links-collapsible-item mw-list-item user-links-collapsible-item"><a data-mw="interface" href="/w/index.php?title=Special:CreateAccount&amp;returnto=Generative+pre-trained+transformer" title="You are encouraged to create an account and log in; however, it is not mandatory" class=""><span>Create account</span></a> </li> <li id="pt-login-2" class="user-links-collapsible-item mw-list-item user-links-collapsible-item"><a data-mw="interface" href="/w/index.php?title=Special:UserLogin&amp;returnto=Generative+pre-trained+transformer" title="You&#039;re encouraged to log in; however, it&#039;s not mandatory. [o]" accesskey="o" class=""><span>Log in</span></a> </li> </ul> </div> </div> </div> <div id="vector-user-links-dropdown" class="vector-dropdown vector-user-menu vector-button-flush-right vector-user-menu-logged-out" title="Log in and more options" > <input type="checkbox" id="vector-user-links-dropdown-checkbox" role="button" aria-haspopup="true" data-event-name="ui.dropdown-vector-user-links-dropdown" class="vector-dropdown-checkbox " aria-label="Personal tools" > <label id="vector-user-links-dropdown-label" for="vector-user-links-dropdown-checkbox" class="vector-dropdown-label cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only " aria-hidden="true" ><span class="vector-icon mw-ui-icon-ellipsis mw-ui-icon-wikimedia-ellipsis"></span> <span class="vector-dropdown-label-text">Personal tools</span> </label> <div class="vector-dropdown-content"> <div id="p-personal" class="vector-menu mw-portlet mw-portlet-personal user-links-collapsible-item" title="User menu" > <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li id="pt-sitesupport" class="user-links-collapsible-item mw-list-item"><a href="https://donate.wikimedia.org/?wmf_source=donate&amp;wmf_medium=sidebar&amp;wmf_campaign=en.wikipedia.org&amp;uselang=en"><span>Donate</span></a></li><li id="pt-createaccount" class="user-links-collapsible-item mw-list-item"><a href="/w/index.php?title=Special:CreateAccount&amp;returnto=Generative+pre-trained+transformer" title="You are encouraged to create an account and log in; however, it is not mandatory"><span class="vector-icon mw-ui-icon-userAdd mw-ui-icon-wikimedia-userAdd"></span> <span>Create account</span></a></li><li id="pt-login" class="user-links-collapsible-item mw-list-item"><a href="/w/index.php?title=Special:UserLogin&amp;returnto=Generative+pre-trained+transformer" title="You&#039;re encouraged to log in; however, it&#039;s not mandatory. [o]" accesskey="o"><span class="vector-icon mw-ui-icon-logIn mw-ui-icon-wikimedia-logIn"></span> <span>Log in</span></a></li> </ul> </div> </div> <div id="p-user-menu-anon-editor" class="vector-menu mw-portlet mw-portlet-user-menu-anon-editor" > <div class="vector-menu-heading"> Pages for logged out editors <a href="/wiki/Help:Introduction" aria-label="Learn more about editing"><span>learn more</span></a> </div> <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li id="pt-anoncontribs" class="mw-list-item"><a href="/wiki/Special:MyContributions" title="A list of edits made from this IP address [y]" accesskey="y"><span>Contributions</span></a></li><li id="pt-anontalk" class="mw-list-item"><a href="/wiki/Special:MyTalk" title="Discussion about edits from this IP address [n]" accesskey="n"><span>Talk</span></a></li> </ul> </div> </div> </div> </div> </nav> </div> </header> </div> <div class="mw-page-container"> <div class="mw-page-container-inner"> <div class="vector-sitenotice-container"> <div id="siteNotice"><!-- CentralNotice --></div> </div> <div class="vector-column-start"> <div class="vector-main-menu-container"> <div id="mw-navigation"> <nav id="mw-panel" class="vector-main-menu-landmark" aria-label="Site"> <div id="vector-main-menu-pinned-container" class="vector-pinned-container"> </div> </nav> </div> </div> <div class="vector-sticky-pinned-container"> <nav id="mw-panel-toc" aria-label="Contents" data-event-name="ui.sidebar-toc" class="mw-table-of-contents-container vector-toc-landmark"> <div id="vector-toc-pinned-container" class="vector-pinned-container"> <div id="vector-toc" class="vector-toc vector-pinnable-element"> <div class="vector-pinnable-header vector-toc-pinnable-header vector-pinnable-header-pinned" data-feature-name="toc-pinned" data-pinnable-element-id="vector-toc" > <h2 class="vector-pinnable-header-label">Contents</h2> <button class="vector-pinnable-header-toggle-button vector-pinnable-header-pin-button" data-event-name="pinnable-header.vector-toc.pin">move to sidebar</button> <button class="vector-pinnable-header-toggle-button vector-pinnable-header-unpin-button" data-event-name="pinnable-header.vector-toc.unpin">hide</button> </div> <ul class="vector-toc-contents" id="mw-panel-toc-list"> <li id="toc-mw-content-text" class="vector-toc-list-item vector-toc-level-1"> <a href="#" class="vector-toc-link"> <div class="vector-toc-text">(Top)</div> </a> </li> <li id="toc-History" class="vector-toc-list-item vector-toc-level-1 vector-toc-list-item-expanded"> <a class="vector-toc-link" href="#History"> <div class="vector-toc-text"> <span class="vector-toc-numb">1</span> <span>History</span> </div> </a> <button aria-controls="toc-History-sublist" class="cdx-button cdx-button--weight-quiet cdx-button--icon-only vector-toc-toggle"> <span class="vector-icon mw-ui-icon-wikimedia-expand"></span> <span>Toggle History subsection</span> </button> <ul id="toc-History-sublist" class="vector-toc-list"> <li id="toc-Initial_developments" class="vector-toc-list-item vector-toc-level-2"> <a class="vector-toc-link" href="#Initial_developments"> <div class="vector-toc-text"> <span class="vector-toc-numb">1.1</span> <span>Initial developments</span> </div> </a> <ul id="toc-Initial_developments-sublist" class="vector-toc-list"> </ul> </li> <li id="toc-Later_developments" class="vector-toc-list-item vector-toc-level-2"> <a class="vector-toc-link" href="#Later_developments"> <div class="vector-toc-text"> <span class="vector-toc-numb">1.2</span> <span>Later developments</span> </div> </a> <ul id="toc-Later_developments-sublist" class="vector-toc-list"> </ul> </li> </ul> </li> <li id="toc-Foundation_models" class="vector-toc-list-item vector-toc-level-1 vector-toc-list-item-expanded"> <a class="vector-toc-link" href="#Foundation_models"> <div class="vector-toc-text"> <span class="vector-toc-numb">2</span> <span>Foundation models</span> </div> </a> <ul id="toc-Foundation_models-sublist" class="vector-toc-list"> </ul> </li> <li id="toc-Task-specific_models" class="vector-toc-list-item vector-toc-level-1 vector-toc-list-item-expanded"> <a class="vector-toc-link" href="#Task-specific_models"> <div class="vector-toc-text"> <span class="vector-toc-numb">3</span> <span>Task-specific models</span> </div> </a> <button aria-controls="toc-Task-specific_models-sublist" class="cdx-button cdx-button--weight-quiet cdx-button--icon-only vector-toc-toggle"> <span class="vector-icon mw-ui-icon-wikimedia-expand"></span> <span>Toggle Task-specific models subsection</span> </button> <ul id="toc-Task-specific_models-sublist" class="vector-toc-list"> <li id="toc-Multimodality" class="vector-toc-list-item vector-toc-level-2"> <a class="vector-toc-link" href="#Multimodality"> <div class="vector-toc-text"> <span class="vector-toc-numb">3.1</span> <span>Multimodality</span> </div> </a> <ul id="toc-Multimodality-sublist" class="vector-toc-list"> </ul> </li> <li id="toc-Domain-specificity" class="vector-toc-list-item vector-toc-level-2"> <a class="vector-toc-link" href="#Domain-specificity"> <div class="vector-toc-text"> <span class="vector-toc-numb">3.2</span> <span>Domain-specificity</span> </div> </a> <ul id="toc-Domain-specificity-sublist" class="vector-toc-list"> </ul> </li> </ul> </li> <li id="toc-Brand_issues" class="vector-toc-list-item vector-toc-level-1 vector-toc-list-item-expanded"> <a class="vector-toc-link" href="#Brand_issues"> <div class="vector-toc-text"> <span class="vector-toc-numb">4</span> <span>Brand issues</span> </div> </a> <ul id="toc-Brand_issues-sublist" class="vector-toc-list"> </ul> </li> <li id="toc-Selected_bibliography" class="vector-toc-list-item vector-toc-level-1 vector-toc-list-item-expanded"> <a class="vector-toc-link" href="#Selected_bibliography"> <div class="vector-toc-text"> <span class="vector-toc-numb">5</span> <span>Selected bibliography</span> </div> </a> <ul id="toc-Selected_bibliography-sublist" class="vector-toc-list"> </ul> </li> <li id="toc-See_also" class="vector-toc-list-item vector-toc-level-1 vector-toc-list-item-expanded"> <a class="vector-toc-link" href="#See_also"> <div class="vector-toc-text"> <span class="vector-toc-numb">6</span> <span>See also</span> </div> </a> <ul id="toc-See_also-sublist" class="vector-toc-list"> </ul> </li> <li id="toc-References" class="vector-toc-list-item vector-toc-level-1 vector-toc-list-item-expanded"> <a class="vector-toc-link" href="#References"> <div class="vector-toc-text"> <span class="vector-toc-numb">7</span> <span>References</span> </div> </a> <ul id="toc-References-sublist" class="vector-toc-list"> </ul> </li> </ul> </div> </div> </nav> </div> </div> <div class="mw-content-container"> <main id="content" class="mw-body"> <header class="mw-body-header vector-page-titlebar"> <nav aria-label="Contents" class="vector-toc-landmark"> <div id="vector-page-titlebar-toc" class="vector-dropdown vector-page-titlebar-toc vector-button-flush-left" title="Table of Contents" > <input type="checkbox" id="vector-page-titlebar-toc-checkbox" role="button" aria-haspopup="true" data-event-name="ui.dropdown-vector-page-titlebar-toc" class="vector-dropdown-checkbox " aria-label="Toggle the table of contents" > <label id="vector-page-titlebar-toc-label" for="vector-page-titlebar-toc-checkbox" class="vector-dropdown-label cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only " aria-hidden="true" ><span class="vector-icon mw-ui-icon-listBullet mw-ui-icon-wikimedia-listBullet"></span> <span class="vector-dropdown-label-text">Toggle the table of contents</span> </label> <div class="vector-dropdown-content"> <div id="vector-page-titlebar-toc-unpinned-container" class="vector-unpinned-container"> </div> </div> </div> </nav> <h1 id="firstHeading" class="firstHeading mw-first-heading"><span class="mw-page-title-main">Generative pre-trained transformer</span></h1> <div id="p-lang-btn" class="vector-dropdown mw-portlet mw-portlet-lang" > <input type="checkbox" id="p-lang-btn-checkbox" role="button" aria-haspopup="true" data-event-name="ui.dropdown-p-lang-btn" class="vector-dropdown-checkbox mw-interlanguage-selector" aria-label="Go to an article in another language. Available in 26 languages" > <label id="p-lang-btn-label" for="p-lang-btn-checkbox" class="vector-dropdown-label cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--action-progressive mw-portlet-lang-heading-26" aria-hidden="true" ><span class="vector-icon mw-ui-icon-language-progressive mw-ui-icon-wikimedia-language-progressive"></span> <span class="vector-dropdown-label-text">26 languages</span> </label> <div class="vector-dropdown-content"> <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li class="interlanguage-link interwiki-ar mw-list-item"><a href="https://ar.wikipedia.org/wiki/%D9%85%D8%AD%D9%88%D9%84_%D9%85%D9%88%D9%84%D8%AF_%D9%85%D8%B3%D8%A8%D9%82_%D8%A7%D9%84%D8%AA%D8%AF%D8%B1%D9%8A%D8%A8" title="محول مولد مسبق التدريب – Arabic" lang="ar" hreflang="ar" data-title="محول مولد مسبق التدريب" data-language-autonym="العربية" data-language-local-name="Arabic" class="interlanguage-link-target"><span>العربية</span></a></li><li class="interlanguage-link interwiki-bar mw-list-item"><a href="https://bar.wikipedia.org/wiki/Generative_pre-trained_transformer" title="Generative pre-trained transformer – Bavarian" lang="bar" hreflang="bar" data-title="Generative pre-trained transformer" data-language-autonym="Boarisch" data-language-local-name="Bavarian" class="interlanguage-link-target"><span>Boarisch</span></a></li><li class="interlanguage-link interwiki-ca mw-list-item"><a href="https://ca.wikipedia.org/wiki/Transformadors_generatius_pre-entrenats" title="Transformadors generatius pre-entrenats – Catalan" lang="ca" hreflang="ca" data-title="Transformadors generatius pre-entrenats" data-language-autonym="Català" data-language-local-name="Catalan" class="interlanguage-link-target"><span>Català</span></a></li><li class="interlanguage-link interwiki-cs mw-list-item"><a href="https://cs.wikipedia.org/wiki/Generative_pre-trained_transformer" title="Generative pre-trained transformer – Czech" lang="cs" hreflang="cs" data-title="Generative pre-trained transformer" data-language-autonym="Čeština" data-language-local-name="Czech" class="interlanguage-link-target"><span>Čeština</span></a></li><li class="interlanguage-link interwiki-de mw-list-item"><a href="https://de.wikipedia.org/wiki/Generativer_vortrainierter_Transformer" title="Generativer vortrainierter Transformer – German" lang="de" hreflang="de" data-title="Generativer vortrainierter Transformer" data-language-autonym="Deutsch" data-language-local-name="German" class="interlanguage-link-target"><span>Deutsch</span></a></li><li class="interlanguage-link interwiki-es mw-list-item"><a href="https://es.wikipedia.org/wiki/Transformador_generativo_preentrenado" title="Transformador generativo preentrenado – Spanish" lang="es" hreflang="es" data-title="Transformador generativo preentrenado" data-language-autonym="Español" data-language-local-name="Spanish" class="interlanguage-link-target"><span>Español</span></a></li><li class="interlanguage-link interwiki-fa mw-list-item"><a href="https://fa.wikipedia.org/wiki/%D8%AA%D8%B1%D9%86%D8%B3%D9%81%D9%88%D8%B1%D9%85%D8%B1_%D8%AA%D9%88%D9%84%DB%8C%D8%AF%DA%AF%D8%B1_%D8%A7%D8%B2_%D9%BE%DB%8C%D8%B4_%D8%A2%D9%85%D9%88%D8%B2%D8%B4%E2%80%8C%D8%AF%DB%8C%D8%AF%D9%87" title="ترنسفورمر تولیدگر از پیش آموزش‌دیده – Persian" lang="fa" hreflang="fa" data-title="ترنسفورمر تولیدگر از پیش آموزش‌دیده" data-language-autonym="فارسی" data-language-local-name="Persian" class="interlanguage-link-target"><span>فارسی</span></a></li><li class="interlanguage-link interwiki-fr mw-list-item"><a href="https://fr.wikipedia.org/wiki/Transformeur_g%C3%A9n%C3%A9ratif_pr%C3%A9entra%C3%AEn%C3%A9" title="Transformeur génératif préentraîné – French" lang="fr" hreflang="fr" data-title="Transformeur génératif préentraîné" data-language-autonym="Français" data-language-local-name="French" class="interlanguage-link-target"><span>Français</span></a></li><li class="interlanguage-link interwiki-ko mw-list-item"><a href="https://ko.wikipedia.org/wiki/GPT_(%EC%96%B8%EC%96%B4_%EB%AA%A8%EB%8D%B8)" title="GPT (언어 모델) – Korean" lang="ko" hreflang="ko" data-title="GPT (언어 모델)" data-language-autonym="한국어" data-language-local-name="Korean" class="interlanguage-link-target"><span>한국어</span></a></li><li class="interlanguage-link interwiki-hi mw-list-item"><a href="https://hi.wikipedia.org/wiki/%E0%A4%89%E0%A4%A4%E0%A5%8D%E0%A4%AA%E0%A4%BE%E0%A4%A6%E0%A4%95_%E0%A4%AA%E0%A5%82%E0%A4%B0%E0%A5%8D%E0%A4%B5-%E0%A4%AA%E0%A5%8D%E0%A4%B0%E0%A4%B6%E0%A4%BF%E0%A4%95%E0%A5%8D%E0%A4%B7%E0%A4%BF%E0%A4%A4_%E0%A4%9F%E0%A5%8D%E0%A4%B0%E0%A4%BE%E0%A4%82%E0%A4%B8%E0%A4%AB%E0%A5%89%E0%A4%B0%E0%A5%8D%E0%A4%AE%E0%A4%B0" title="उत्पादक पूर्व-प्रशिक्षित ट्रांसफॉर्मर – Hindi" lang="hi" hreflang="hi" data-title="उत्पादक पूर्व-प्रशिक्षित ट्रांसफॉर्मर" data-language-autonym="हिन्दी" data-language-local-name="Hindi" class="interlanguage-link-target"><span>हिन्दी</span></a></li><li class="interlanguage-link interwiki-id mw-list-item"><a href="https://id.wikipedia.org/wiki/Penyulihbentuk_praterlatih_generatif" title="Penyulihbentuk praterlatih generatif – Indonesian" lang="id" hreflang="id" data-title="Penyulihbentuk praterlatih generatif" data-language-autonym="Bahasa Indonesia" data-language-local-name="Indonesian" class="interlanguage-link-target"><span>Bahasa Indonesia</span></a></li><li class="interlanguage-link interwiki-zu mw-list-item"><a href="https://zu.wikipedia.org/wiki/IsiGwabuzi_esiPhehlayo_esiQeqeshiwe" title="IsiGwabuzi esiPhehlayo esiQeqeshiwe – Zulu" lang="zu" hreflang="zu" data-title="IsiGwabuzi esiPhehlayo esiQeqeshiwe" data-language-autonym="IsiZulu" data-language-local-name="Zulu" class="interlanguage-link-target"><span>IsiZulu</span></a></li><li class="interlanguage-link interwiki-he mw-list-item"><a href="https://he.wikipedia.org/wiki/GPT" title="GPT – Hebrew" lang="he" hreflang="he" data-title="GPT" data-language-autonym="עברית" data-language-local-name="Hebrew" class="interlanguage-link-target"><span>עברית</span></a></li><li class="interlanguage-link interwiki-nl mw-list-item"><a href="https://nl.wikipedia.org/wiki/Generative_pre-trained_transformer" title="Generative pre-trained transformer – Dutch" lang="nl" hreflang="nl" data-title="Generative pre-trained transformer" data-language-autonym="Nederlands" data-language-local-name="Dutch" class="interlanguage-link-target"><span>Nederlands</span></a></li><li class="interlanguage-link interwiki-ja mw-list-item"><a href="https://ja.wikipedia.org/wiki/GPT_(%E8%A8%80%E8%AA%9E%E3%83%A2%E3%83%87%E3%83%AB)" title="GPT (言語モデル) – Japanese" lang="ja" hreflang="ja" data-title="GPT (言語モデル)" data-language-autonym="日本語" data-language-local-name="Japanese" class="interlanguage-link-target"><span>日本語</span></a></li><li class="interlanguage-link interwiki-pl mw-list-item"><a href="https://pl.wikipedia.org/wiki/Wst%C4%99pnie_przeszkolony_transformator_generatywny" title="Wstępnie przeszkolony transformator generatywny – Polish" lang="pl" hreflang="pl" data-title="Wstępnie przeszkolony transformator generatywny" data-language-autonym="Polski" data-language-local-name="Polish" class="interlanguage-link-target"><span>Polski</span></a></li><li class="interlanguage-link interwiki-kaa mw-list-item"><a href="https://kaa.wikipedia.org/wiki/Generativ_ald%C4%B1nnan_tayarlan%C7%B5an_transformator" title="Generativ aldınnan tayarlanǵan transformator – Kara-Kalpak" lang="kaa" hreflang="kaa" data-title="Generativ aldınnan tayarlanǵan transformator" data-language-autonym="Qaraqalpaqsha" data-language-local-name="Kara-Kalpak" class="interlanguage-link-target"><span>Qaraqalpaqsha</span></a></li><li class="interlanguage-link interwiki-ro mw-list-item"><a href="https://ro.wikipedia.org/wiki/Generative_pre-trained_transformer" title="Generative pre-trained transformer – Romanian" lang="ro" hreflang="ro" data-title="Generative pre-trained transformer" data-language-autonym="Română" data-language-local-name="Romanian" class="interlanguage-link-target"><span>Română</span></a></li><li class="interlanguage-link interwiki-ru mw-list-item"><a href="https://ru.wikipedia.org/wiki/Generative_pre-trained_transformer" title="Generative pre-trained transformer – Russian" lang="ru" hreflang="ru" data-title="Generative pre-trained transformer" data-language-autonym="Русский" data-language-local-name="Russian" class="interlanguage-link-target"><span>Русский</span></a></li><li class="interlanguage-link interwiki-sr mw-list-item"><a href="https://sr.wikipedia.org/wiki/Generativni_prethodno_obu%C4%8Deni_transformator" title="Generativni prethodno obučeni transformator – Serbian" lang="sr" hreflang="sr" data-title="Generativni prethodno obučeni transformator" data-language-autonym="Српски / srpski" data-language-local-name="Serbian" class="interlanguage-link-target"><span>Српски / srpski</span></a></li><li class="interlanguage-link interwiki-fi mw-list-item"><a href="https://fi.wikipedia.org/wiki/GPT_(kielimalliperhe)" title="GPT (kielimalliperhe) – Finnish" lang="fi" hreflang="fi" data-title="GPT (kielimalliperhe)" data-language-autonym="Suomi" data-language-local-name="Finnish" class="interlanguage-link-target"><span>Suomi</span></a></li><li class="interlanguage-link interwiki-th mw-list-item"><a href="https://th.wikipedia.org/wiki/%E0%B8%97%E0%B8%A3%E0%B8%B2%E0%B8%99%E0%B8%AA%E0%B9%8C%E0%B8%9F%E0%B8%AD%E0%B8%A3%E0%B9%8C%E0%B9%80%E0%B8%A1%E0%B8%AD%E0%B8%A3%E0%B9%8C%E0%B8%9D%E0%B8%B6%E0%B8%81%E0%B8%A5%E0%B9%88%E0%B8%A7%E0%B8%87%E0%B8%AB%E0%B8%99%E0%B9%89%E0%B8%B2%E0%B8%81%E0%B9%88%E0%B8%AD%E0%B8%81%E0%B8%B3%E0%B9%80%E0%B8%99%E0%B8%B4%E0%B8%94" title="ทรานส์ฟอร์เมอร์ฝึกล่วงหน้าก่อกำเนิด – Thai" lang="th" hreflang="th" data-title="ทรานส์ฟอร์เมอร์ฝึกล่วงหน้าก่อกำเนิด" data-language-autonym="ไทย" data-language-local-name="Thai" class="interlanguage-link-target"><span>ไทย</span></a></li><li class="interlanguage-link interwiki-tr mw-list-item"><a href="https://tr.wikipedia.org/wiki/%C3%96nceden_e%C4%9Fitilmi%C5%9F_%C3%BCretken_d%C3%B6n%C3%BC%C5%9Ft%C3%BCr%C3%BCc%C3%BC" title="Önceden eğitilmiş üretken dönüştürücü – Turkish" lang="tr" hreflang="tr" data-title="Önceden eğitilmiş üretken dönüştürücü" data-language-autonym="Türkçe" data-language-local-name="Turkish" class="interlanguage-link-target"><span>Türkçe</span></a></li><li class="interlanguage-link interwiki-uk mw-list-item"><a href="https://uk.wikipedia.org/wiki/Generative_pre-trained_transformer" title="Generative pre-trained transformer – Ukrainian" lang="uk" hreflang="uk" data-title="Generative pre-trained transformer" data-language-autonym="Українська" data-language-local-name="Ukrainian" class="interlanguage-link-target"><span>Українська</span></a></li><li class="interlanguage-link interwiki-ug mw-list-item"><a href="https://ug.wikipedia.org/wiki/%D8%A6%D8%A7%D9%84%D8%AF%D9%89%D9%86_%D8%A6%DB%86%DA%AF%D9%89%D8%AA%D9%89%D9%84%DA%AF%DB%95%D9%86_%DA%BE%D8%A7%D8%B3%D9%89%D9%84%D8%A7%D8%AA%D9%84%D9%89%D9%82_%D8%AA%D9%89%D8%B1%D8%A7%D9%86%D8%B3%D9%81%D9%88%D8%B1%D9%85%D8%A7%D8%AA%D9%88%D8%B1" title="ئالدىن ئۆگىتىلگەن ھاسىلاتلىق تىرانسفورماتور – Uyghur" lang="ug" hreflang="ug" data-title="ئالدىن ئۆگىتىلگەن ھاسىلاتلىق تىرانسفورماتور" data-language-autonym="ئۇيغۇرچە / Uyghurche" data-language-local-name="Uyghur" class="interlanguage-link-target"><span>ئۇيغۇرچە / Uyghurche</span></a></li><li class="interlanguage-link interwiki-zh mw-list-item"><a href="https://zh.wikipedia.org/wiki/GPT_(%E8%AF%AD%E8%A8%80%E6%A8%A1%E5%9E%8B)" title="GPT (语言模型) – Chinese" lang="zh" hreflang="zh" data-title="GPT (语言模型)" data-language-autonym="中文" data-language-local-name="Chinese" class="interlanguage-link-target"><span>中文</span></a></li> </ul> <div class="after-portlet after-portlet-lang"><span class="wb-langlinks-edit wb-langlinks-link"><a href="https://www.wikidata.org/wiki/Special:EntityPage/Q116777014#sitelinks-wikipedia" title="Edit interlanguage links" class="wbc-editpage">Edit links</a></span></div> </div> </div> </div> </header> <div class="vector-page-toolbar"> <div class="vector-page-toolbar-container"> <div id="left-navigation"> <nav aria-label="Namespaces"> <div id="p-associated-pages" class="vector-menu vector-menu-tabs mw-portlet mw-portlet-associated-pages" > <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li id="ca-nstab-main" class="selected vector-tab-noicon mw-list-item"><a href="/wiki/Generative_pre-trained_transformer" title="View the content page [c]" accesskey="c"><span>Article</span></a></li><li id="ca-talk" class="vector-tab-noicon mw-list-item"><a href="/wiki/Talk:Generative_pre-trained_transformer" rel="discussion" title="Discuss improvements to the content page [t]" accesskey="t"><span>Talk</span></a></li> </ul> </div> </div> <div id="vector-variants-dropdown" class="vector-dropdown emptyPortlet" > <input type="checkbox" id="vector-variants-dropdown-checkbox" role="button" aria-haspopup="true" data-event-name="ui.dropdown-vector-variants-dropdown" class="vector-dropdown-checkbox " aria-label="Change language variant" > <label id="vector-variants-dropdown-label" for="vector-variants-dropdown-checkbox" class="vector-dropdown-label cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet" aria-hidden="true" ><span class="vector-dropdown-label-text">English</span> </label> <div class="vector-dropdown-content"> <div id="p-variants" class="vector-menu mw-portlet mw-portlet-variants emptyPortlet" > <div class="vector-menu-content"> <ul class="vector-menu-content-list"> </ul> </div> </div> </div> </div> </nav> </div> <div id="right-navigation" class="vector-collapsible"> <nav aria-label="Views"> <div id="p-views" class="vector-menu vector-menu-tabs mw-portlet mw-portlet-views" > <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li id="ca-view" class="selected vector-tab-noicon mw-list-item"><a href="/wiki/Generative_pre-trained_transformer"><span>Read</span></a></li><li id="ca-edit" class="vector-tab-noicon mw-list-item"><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit" title="Edit this page [e]" accesskey="e"><span>Edit</span></a></li><li id="ca-history" class="vector-tab-noicon mw-list-item"><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=history" title="Past revisions of this page [h]" accesskey="h"><span>View history</span></a></li> </ul> </div> </div> </nav> <nav class="vector-page-tools-landmark" aria-label="Page tools"> <div id="vector-page-tools-dropdown" class="vector-dropdown vector-page-tools-dropdown" > <input type="checkbox" id="vector-page-tools-dropdown-checkbox" role="button" aria-haspopup="true" data-event-name="ui.dropdown-vector-page-tools-dropdown" class="vector-dropdown-checkbox " aria-label="Tools" > <label id="vector-page-tools-dropdown-label" for="vector-page-tools-dropdown-checkbox" class="vector-dropdown-label cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet" aria-hidden="true" ><span class="vector-dropdown-label-text">Tools</span> </label> <div class="vector-dropdown-content"> <div id="vector-page-tools-unpinned-container" class="vector-unpinned-container"> <div id="vector-page-tools" class="vector-page-tools vector-pinnable-element"> <div class="vector-pinnable-header vector-page-tools-pinnable-header vector-pinnable-header-unpinned" data-feature-name="page-tools-pinned" data-pinnable-element-id="vector-page-tools" data-pinned-container-id="vector-page-tools-pinned-container" data-unpinned-container-id="vector-page-tools-unpinned-container" > <div class="vector-pinnable-header-label">Tools</div> <button class="vector-pinnable-header-toggle-button vector-pinnable-header-pin-button" data-event-name="pinnable-header.vector-page-tools.pin">move to sidebar</button> <button class="vector-pinnable-header-toggle-button vector-pinnable-header-unpin-button" data-event-name="pinnable-header.vector-page-tools.unpin">hide</button> </div> <div id="p-cactions" class="vector-menu mw-portlet mw-portlet-cactions emptyPortlet vector-has-collapsible-items" title="More options" > <div class="vector-menu-heading"> Actions </div> <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li id="ca-more-view" class="selected vector-more-collapsible-item mw-list-item"><a href="/wiki/Generative_pre-trained_transformer"><span>Read</span></a></li><li id="ca-more-edit" class="vector-more-collapsible-item mw-list-item"><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit" title="Edit this page [e]" accesskey="e"><span>Edit</span></a></li><li id="ca-more-history" class="vector-more-collapsible-item mw-list-item"><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=history"><span>View history</span></a></li> </ul> </div> </div> <div id="p-tb" class="vector-menu mw-portlet mw-portlet-tb" > <div class="vector-menu-heading"> General </div> <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li id="t-whatlinkshere" class="mw-list-item"><a href="/wiki/Special:WhatLinksHere/Generative_pre-trained_transformer" title="List of all English Wikipedia pages containing links to this page [j]" accesskey="j"><span>What links here</span></a></li><li id="t-recentchangeslinked" class="mw-list-item"><a href="/wiki/Special:RecentChangesLinked/Generative_pre-trained_transformer" rel="nofollow" title="Recent changes in pages linked from this page [k]" accesskey="k"><span>Related changes</span></a></li><li id="t-upload" class="mw-list-item"><a href="//en.wikipedia.org/wiki/Wikipedia:File_Upload_Wizard" title="Upload files [u]" accesskey="u"><span>Upload file</span></a></li><li id="t-permalink" class="mw-list-item"><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;oldid=1275868082" title="Permanent link to this revision of this page"><span>Permanent link</span></a></li><li id="t-info" class="mw-list-item"><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=info" title="More information about this page"><span>Page information</span></a></li><li id="t-cite" class="mw-list-item"><a href="/w/index.php?title=Special:CiteThisPage&amp;page=Generative_pre-trained_transformer&amp;id=1275868082&amp;wpFormIdentifier=titleform" title="Information on how to cite this page"><span>Cite this page</span></a></li><li id="t-urlshortener" class="mw-list-item"><a href="/w/index.php?title=Special:UrlShortener&amp;url=https%3A%2F%2Fen.wikipedia.org%2Fwiki%2FGenerative_pre-trained_transformer"><span>Get shortened URL</span></a></li><li id="t-urlshortener-qrcode" class="mw-list-item"><a href="/w/index.php?title=Special:QrCode&amp;url=https%3A%2F%2Fen.wikipedia.org%2Fwiki%2FGenerative_pre-trained_transformer"><span>Download QR code</span></a></li> </ul> </div> </div> <div id="p-coll-print_export" class="vector-menu mw-portlet mw-portlet-coll-print_export" > <div class="vector-menu-heading"> Print/export </div> <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li id="coll-download-as-rl" class="mw-list-item"><a href="/w/index.php?title=Special:DownloadAsPdf&amp;page=Generative_pre-trained_transformer&amp;action=show-download-screen" title="Download this page as a PDF file"><span>Download as PDF</span></a></li><li id="t-print" class="mw-list-item"><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;printable=yes" title="Printable version of this page [p]" accesskey="p"><span>Printable version</span></a></li> </ul> </div> </div> <div id="p-wikibase-otherprojects" class="vector-menu mw-portlet mw-portlet-wikibase-otherprojects" > <div class="vector-menu-heading"> In other projects </div> <div class="vector-menu-content"> <ul class="vector-menu-content-list"> <li id="t-wikibase" class="wb-otherproject-link wb-otherproject-wikibase-dataitem mw-list-item"><a href="https://www.wikidata.org/wiki/Special:EntityPage/Q116777014" title="Structured data on this page hosted by Wikidata [g]" accesskey="g"><span>Wikidata item</span></a></li> </ul> </div> </div> </div> </div> </div> </div> </nav> </div> </div> </div> <div class="vector-column-end"> <div class="vector-sticky-pinned-container"> <nav class="vector-page-tools-landmark" aria-label="Page tools"> <div id="vector-page-tools-pinned-container" class="vector-pinned-container"> </div> </nav> <nav class="vector-appearance-landmark" aria-label="Appearance"> <div id="vector-appearance-pinned-container" class="vector-pinned-container"> <div id="vector-appearance" class="vector-appearance vector-pinnable-element"> <div class="vector-pinnable-header vector-appearance-pinnable-header vector-pinnable-header-pinned" data-feature-name="appearance-pinned" data-pinnable-element-id="vector-appearance" data-pinned-container-id="vector-appearance-pinned-container" data-unpinned-container-id="vector-appearance-unpinned-container" > <div class="vector-pinnable-header-label">Appearance</div> <button class="vector-pinnable-header-toggle-button vector-pinnable-header-pin-button" data-event-name="pinnable-header.vector-appearance.pin">move to sidebar</button> <button class="vector-pinnable-header-toggle-button vector-pinnable-header-unpin-button" data-event-name="pinnable-header.vector-appearance.unpin">hide</button> </div> </div> </div> </nav> </div> </div> <div id="bodyContent" class="vector-body" aria-labelledby="firstHeading" data-mw-ve-target-container> <div class="vector-body-before-content"> <div class="mw-indicators"> </div> <div id="siteSub" class="noprint">From Wikipedia, the free encyclopedia</div> </div> <div id="contentSub"><div id="mw-content-subtitle"></div></div> <div id="mw-content-text" class="mw-body-content"><div class="mw-content-ltr mw-parser-output" lang="en" dir="ltr"><div class="shortdescription nomobile noexcerpt noprint searchaux" style="display:none">Type of large language model</div><style data-mw-deduplicate="TemplateStyles:r1236090951">.mw-parser-output .hatnote{font-style:italic}.mw-parser-output div.hatnote{padding-left:1.6em;margin-bottom:0.5em}.mw-parser-output .hatnote i{font-style:normal}.mw-parser-output .hatnote+link+.hatnote{margin-top:-0.5em}@media print{body.ns-0 .mw-parser-output .hatnote{display:none!important}}</style><div role="note" class="hatnote navigation-not-searchable">Not to be confused with <a href="/wiki/ChatGPT" title="ChatGPT">ChatGPT</a>.</div> <style data-mw-deduplicate="TemplateStyles:r1244144826">.mw-parser-output .machine-learning-list-title{background-color:#ddddff}html.skin-theme-clientpref-night .mw-parser-output .machine-learning-list-title{background-color:#222}@media(prefers-color-scheme:dark){html.skin-theme-clientpref-os .mw-parser-output .machine-learning-list-title{background-color:#222}}</style> <style data-mw-deduplicate="TemplateStyles:r1129693374">.mw-parser-output .hlist dl,.mw-parser-output .hlist ol,.mw-parser-output .hlist ul{margin:0;padding:0}.mw-parser-output .hlist dd,.mw-parser-output .hlist dt,.mw-parser-output .hlist li{margin:0;display:inline}.mw-parser-output .hlist.inline,.mw-parser-output .hlist.inline dl,.mw-parser-output .hlist.inline ol,.mw-parser-output .hlist.inline ul,.mw-parser-output .hlist dl dl,.mw-parser-output .hlist dl ol,.mw-parser-output .hlist dl ul,.mw-parser-output .hlist ol dl,.mw-parser-output .hlist ol ol,.mw-parser-output .hlist ol ul,.mw-parser-output .hlist ul dl,.mw-parser-output .hlist ul ol,.mw-parser-output .hlist ul ul{display:inline}.mw-parser-output .hlist .mw-empty-li{display:none}.mw-parser-output .hlist dt::after{content:": "}.mw-parser-output .hlist dd::after,.mw-parser-output .hlist li::after{content:" · ";font-weight:bold}.mw-parser-output .hlist dd:last-child::after,.mw-parser-output .hlist dt:last-child::after,.mw-parser-output .hlist li:last-child::after{content:none}.mw-parser-output .hlist dd dd:first-child::before,.mw-parser-output .hlist dd dt:first-child::before,.mw-parser-output .hlist dd li:first-child::before,.mw-parser-output .hlist dt dd:first-child::before,.mw-parser-output .hlist dt dt:first-child::before,.mw-parser-output .hlist dt li:first-child::before,.mw-parser-output .hlist li dd:first-child::before,.mw-parser-output .hlist li dt:first-child::before,.mw-parser-output .hlist li li:first-child::before{content:" (";font-weight:normal}.mw-parser-output .hlist dd dd:last-child::after,.mw-parser-output .hlist dd dt:last-child::after,.mw-parser-output .hlist dd li:last-child::after,.mw-parser-output .hlist dt dd:last-child::after,.mw-parser-output .hlist dt dt:last-child::after,.mw-parser-output .hlist dt li:last-child::after,.mw-parser-output .hlist li dd:last-child::after,.mw-parser-output .hlist li dt:last-child::after,.mw-parser-output .hlist li li:last-child::after{content:")";font-weight:normal}.mw-parser-output .hlist ol{counter-reset:listitem}.mw-parser-output .hlist ol>li{counter-increment:listitem}.mw-parser-output .hlist ol>li::before{content:" "counter(listitem)"\a0 "}.mw-parser-output .hlist dd ol>li:first-child::before,.mw-parser-output .hlist dt ol>li:first-child::before,.mw-parser-output .hlist li ol>li:first-child::before{content:" ("counter(listitem)"\a0 "}</style><style data-mw-deduplicate="TemplateStyles:r1246091330">.mw-parser-output .sidebar{width:22em;float:right;clear:right;margin:0.5em 0 1em 1em;background:var(--background-color-neutral-subtle,#f8f9fa);border:1px solid var(--border-color-base,#a2a9b1);padding:0.2em;text-align:center;line-height:1.4em;font-size:88%;border-collapse:collapse;display:table}body.skin-minerva .mw-parser-output .sidebar{display:table!important;float:right!important;margin:0.5em 0 1em 1em!important}.mw-parser-output .sidebar-subgroup{width:100%;margin:0;border-spacing:0}.mw-parser-output .sidebar-left{float:left;clear:left;margin:0.5em 1em 1em 0}.mw-parser-output .sidebar-none{float:none;clear:both;margin:0.5em 1em 1em 0}.mw-parser-output .sidebar-outer-title{padding:0 0.4em 0.2em;font-size:125%;line-height:1.2em;font-weight:bold}.mw-parser-output .sidebar-top-image{padding:0.4em}.mw-parser-output .sidebar-top-caption,.mw-parser-output .sidebar-pretitle-with-top-image,.mw-parser-output .sidebar-caption{padding:0.2em 0.4em 0;line-height:1.2em}.mw-parser-output .sidebar-pretitle{padding:0.4em 0.4em 0;line-height:1.2em}.mw-parser-output .sidebar-title,.mw-parser-output .sidebar-title-with-pretitle{padding:0.2em 0.8em;font-size:145%;line-height:1.2em}.mw-parser-output .sidebar-title-with-pretitle{padding:0.1em 0.4em}.mw-parser-output .sidebar-image{padding:0.2em 0.4em 0.4em}.mw-parser-output .sidebar-heading{padding:0.1em 0.4em}.mw-parser-output .sidebar-content{padding:0 0.5em 0.4em}.mw-parser-output .sidebar-content-with-subgroup{padding:0.1em 0.4em 0.2em}.mw-parser-output .sidebar-above,.mw-parser-output .sidebar-below{padding:0.3em 0.8em;font-weight:bold}.mw-parser-output .sidebar-collapse .sidebar-above,.mw-parser-output .sidebar-collapse .sidebar-below{border-top:1px solid #aaa;border-bottom:1px solid #aaa}.mw-parser-output .sidebar-navbar{text-align:right;font-size:115%;padding:0 0.4em 0.4em}.mw-parser-output .sidebar-list-title{padding:0 0.4em;text-align:left;font-weight:bold;line-height:1.6em;font-size:105%}.mw-parser-output .sidebar-list-title-c{padding:0 0.4em;text-align:center;margin:0 3.3em}@media(max-width:640px){body.mediawiki .mw-parser-output .sidebar{width:100%!important;clear:both;float:none!important;margin-left:0!important;margin-right:0!important}}body.skin--responsive .mw-parser-output .sidebar a>img{max-width:none!important}@media screen{html.skin-theme-clientpref-night .mw-parser-output .sidebar:not(.notheme) .sidebar-list-title,html.skin-theme-clientpref-night .mw-parser-output .sidebar:not(.notheme) .sidebar-title-with-pretitle{background:transparent!important}html.skin-theme-clientpref-night .mw-parser-output .sidebar:not(.notheme) .sidebar-title-with-pretitle a{color:var(--color-progressive)!important}}@media screen and (prefers-color-scheme:dark){html.skin-theme-clientpref-os .mw-parser-output .sidebar:not(.notheme) .sidebar-list-title,html.skin-theme-clientpref-os .mw-parser-output .sidebar:not(.notheme) .sidebar-title-with-pretitle{background:transparent!important}html.skin-theme-clientpref-os .mw-parser-output .sidebar:not(.notheme) .sidebar-title-with-pretitle a{color:var(--color-progressive)!important}}@media print{body.ns-0 .mw-parser-output .sidebar{display:none!important}}</style><style data-mw-deduplicate="TemplateStyles:r886047488">.mw-parser-output .nobold{font-weight:normal}</style><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r886047488"><table class="sidebar sidebar-collapse nomobile nowraplinks"><tbody><tr><td class="sidebar-pretitle">Part of a series on</td></tr><tr><th class="sidebar-title-with-pretitle"><a href="/wiki/Machine_learning" title="Machine learning">Machine learning</a><br />and <a href="/wiki/Data_mining" title="Data mining">data mining</a></th></tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)">Paradigms</div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Supervised_learning" title="Supervised learning">Supervised learning</a></li> <li><a href="/wiki/Unsupervised_learning" title="Unsupervised learning">Unsupervised learning</a></li> <li><a href="/wiki/Semi-supervised_learning" class="mw-redirect" title="Semi-supervised learning">Semi-supervised learning</a></li> <li><a href="/wiki/Self-supervised_learning" title="Self-supervised learning">Self-supervised learning</a></li> <li><a href="/wiki/Reinforcement_learning" title="Reinforcement learning">Reinforcement learning</a></li> <li><a href="/wiki/Meta-learning_(computer_science)" title="Meta-learning (computer science)">Meta-learning</a></li> <li><a href="/wiki/Online_machine_learning" title="Online machine learning">Online learning</a></li> <li><a href="/wiki/Batch_learning" class="mw-redirect" title="Batch learning">Batch learning</a></li> <li><a href="/wiki/Curriculum_learning" title="Curriculum learning">Curriculum learning</a></li> <li><a href="/wiki/Rule-based_machine_learning" title="Rule-based machine learning">Rule-based learning</a></li> <li><a href="/wiki/Neuro-symbolic_AI" title="Neuro-symbolic AI">Neuro-symbolic AI</a></li> <li><a href="/wiki/Neuromorphic_engineering" class="mw-redirect" title="Neuromorphic engineering">Neuromorphic engineering</a></li> <li><a href="/wiki/Quantum_machine_learning" title="Quantum machine learning">Quantum machine learning</a></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)">Problems</div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Statistical_classification" title="Statistical classification">Classification</a></li> <li><a href="/wiki/Generative_model" title="Generative model">Generative modeling</a></li> <li><a href="/wiki/Regression_analysis" title="Regression analysis">Regression</a></li> <li><a href="/wiki/Cluster_analysis" title="Cluster analysis">Clustering</a></li> <li><a href="/wiki/Dimensionality_reduction" title="Dimensionality reduction">Dimensionality reduction</a></li> <li><a href="/wiki/Density_estimation" title="Density estimation">Density estimation</a></li> <li><a href="/wiki/Anomaly_detection" title="Anomaly detection">Anomaly detection</a></li> <li><a href="/wiki/Data_cleaning" class="mw-redirect" title="Data cleaning">Data cleaning</a></li> <li><a href="/wiki/Automated_machine_learning" title="Automated machine learning">AutoML</a></li> <li><a href="/wiki/Association_rule_learning" title="Association rule learning">Association rules</a></li> <li><a href="/wiki/Semantic_analysis_(machine_learning)" title="Semantic analysis (machine learning)">Semantic analysis</a></li> <li><a href="/wiki/Structured_prediction" title="Structured prediction">Structured prediction</a></li> <li><a href="/wiki/Feature_engineering" title="Feature engineering">Feature engineering</a></li> <li><a href="/wiki/Feature_learning" title="Feature learning">Feature learning</a></li> <li><a href="/wiki/Learning_to_rank" title="Learning to rank">Learning to rank</a></li> <li><a href="/wiki/Grammar_induction" title="Grammar induction">Grammar induction</a></li> <li><a href="/wiki/Ontology_learning" title="Ontology learning">Ontology learning</a></li> <li><a href="/wiki/Multimodal_learning" title="Multimodal learning">Multimodal learning</a></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)"><div style="display: inline-block; line-height: 1.2em; padding: .1em 0;"><a href="/wiki/Supervised_learning" title="Supervised learning">Supervised learning</a><br /><span class="nobold"><span style="font-size:85%;">(<b><a href="/wiki/Statistical_classification" title="Statistical classification">classification</a></b>&#160;&#8226;&#32;<b><a href="/wiki/Regression_analysis" title="Regression analysis">regression</a></b>)</span></span> </div></div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Apprenticeship_learning" title="Apprenticeship learning">Apprenticeship learning</a></li> <li><a href="/wiki/Decision_tree_learning" title="Decision tree learning">Decision trees</a></li> <li><a href="/wiki/Ensemble_learning" title="Ensemble learning">Ensembles</a> <ul><li><a href="/wiki/Bootstrap_aggregating" title="Bootstrap aggregating">Bagging</a></li> <li><a href="/wiki/Boosting_(machine_learning)" title="Boosting (machine learning)">Boosting</a></li> <li><a href="/wiki/Random_forest" title="Random forest">Random forest</a></li></ul></li> <li><a href="/wiki/K-nearest_neighbors_algorithm" title="K-nearest neighbors algorithm"><i>k</i>-NN</a></li> <li><a href="/wiki/Linear_regression" title="Linear regression">Linear regression</a></li> <li><a href="/wiki/Naive_Bayes_classifier" title="Naive Bayes classifier">Naive Bayes</a></li> <li><a href="/wiki/Artificial_neural_network" class="mw-redirect" title="Artificial neural network">Artificial neural networks</a></li> <li><a href="/wiki/Logistic_regression" title="Logistic regression">Logistic regression</a></li> <li><a href="/wiki/Perceptron" title="Perceptron">Perceptron</a></li> <li><a href="/wiki/Relevance_vector_machine" title="Relevance vector machine">Relevance vector machine (RVM)</a></li> <li><a href="/wiki/Support_vector_machine" title="Support vector machine">Support vector machine (SVM)</a></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)"><a href="/wiki/Cluster_analysis" title="Cluster analysis">Clustering</a></div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/BIRCH" title="BIRCH">BIRCH</a></li> <li><a href="/wiki/CURE_algorithm" title="CURE algorithm">CURE</a></li> <li><a href="/wiki/Hierarchical_clustering" title="Hierarchical clustering">Hierarchical</a></li> <li><a href="/wiki/K-means_clustering" title="K-means clustering"><i>k</i>-means</a></li> <li><a href="/wiki/Fuzzy_clustering" title="Fuzzy clustering">Fuzzy</a></li> <li><a href="/wiki/Expectation%E2%80%93maximization_algorithm" title="Expectation–maximization algorithm">Expectation–maximization (EM)</a></li> <li><br /><a href="/wiki/DBSCAN" title="DBSCAN">DBSCAN</a></li> <li><a href="/wiki/OPTICS_algorithm" title="OPTICS algorithm">OPTICS</a></li> <li><a href="/wiki/Mean_shift" title="Mean shift">Mean shift</a></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)"><a href="/wiki/Dimensionality_reduction" title="Dimensionality reduction">Dimensionality reduction</a></div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Factor_analysis" title="Factor analysis">Factor analysis</a></li> <li><a href="/wiki/Canonical_correlation" title="Canonical correlation">CCA</a></li> <li><a href="/wiki/Independent_component_analysis" title="Independent component analysis">ICA</a></li> <li><a href="/wiki/Linear_discriminant_analysis" title="Linear discriminant analysis">LDA</a></li> <li><a href="/wiki/Non-negative_matrix_factorization" title="Non-negative matrix factorization">NMF</a></li> <li><a href="/wiki/Principal_component_analysis" title="Principal component analysis">PCA</a></li> <li><a href="/wiki/Proper_generalized_decomposition" title="Proper generalized decomposition">PGD</a></li> <li><a href="/wiki/T-distributed_stochastic_neighbor_embedding" title="T-distributed stochastic neighbor embedding">t-SNE</a></li> <li><a href="/wiki/Sparse_dictionary_learning" title="Sparse dictionary learning">SDL</a></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)"><a href="/wiki/Structured_prediction" title="Structured prediction">Structured prediction</a></div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Graphical_model" title="Graphical model">Graphical models</a> <ul><li><a href="/wiki/Bayesian_network" title="Bayesian network">Bayes net</a></li> <li><a href="/wiki/Conditional_random_field" title="Conditional random field">Conditional random field</a></li> <li><a href="/wiki/Hidden_Markov_model" title="Hidden Markov model">Hidden Markov</a></li></ul></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)"><a href="/wiki/Anomaly_detection" title="Anomaly detection">Anomaly detection</a></div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Random_sample_consensus" title="Random sample consensus">RANSAC</a></li> <li><a href="/wiki/K-nearest_neighbors_algorithm" title="K-nearest neighbors algorithm"><i>k</i>-NN</a></li> <li><a href="/wiki/Local_outlier_factor" title="Local outlier factor">Local outlier factor</a></li> <li><a href="/wiki/Isolation_forest" title="Isolation forest">Isolation forest</a></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)"><a href="/wiki/Artificial_neural_network" class="mw-redirect" title="Artificial neural network">Artificial neural network</a></div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Autoencoder" title="Autoencoder">Autoencoder</a></li> <li><a href="/wiki/Deep_learning" title="Deep learning">Deep learning</a></li> <li><a href="/wiki/Feedforward_neural_network" title="Feedforward neural network">Feedforward neural network</a></li> <li><a href="/wiki/Recurrent_neural_network" title="Recurrent neural network">Recurrent neural network</a> <ul><li><a href="/wiki/Long_short-term_memory" title="Long short-term memory">LSTM</a></li> <li><a href="/wiki/Gated_recurrent_unit" title="Gated recurrent unit">GRU</a></li> <li><a href="/wiki/Echo_state_network" title="Echo state network">ESN</a></li> <li><a href="/wiki/Reservoir_computing" title="Reservoir computing">reservoir computing</a></li></ul></li> <li><a href="/wiki/Boltzmann_machine" title="Boltzmann machine">Boltzmann machine</a> <ul><li><a href="/wiki/Restricted_Boltzmann_machine" title="Restricted Boltzmann machine">Restricted</a></li></ul></li> <li><a href="/wiki/Generative_adversarial_network" title="Generative adversarial network">GAN</a></li> <li><a href="/wiki/Diffusion_model" title="Diffusion model">Diffusion model</a></li> <li><a href="/wiki/Self-organizing_map" title="Self-organizing map">SOM</a></li> <li><a href="/wiki/Convolutional_neural_network" title="Convolutional neural network">Convolutional neural network</a> <ul><li><a href="/wiki/U-Net" title="U-Net">U-Net</a></li> <li><a href="/wiki/LeNet" title="LeNet">LeNet</a></li> <li><a href="/wiki/AlexNet" title="AlexNet">AlexNet</a></li> <li><a href="/wiki/DeepDream" title="DeepDream">DeepDream</a></li></ul></li> <li><a href="/wiki/Neural_radiance_field" title="Neural radiance field">Neural radiance field</a></li> <li><a href="/wiki/Transformer_(machine_learning_model)" class="mw-redirect" title="Transformer (machine learning model)">Transformer</a> <ul><li><a href="/wiki/Vision_transformer" title="Vision transformer">Vision</a></li></ul></li> <li><a href="/wiki/Mamba_(deep_learning_architecture)" title="Mamba (deep learning architecture)">Mamba</a></li> <li><a href="/wiki/Spiking_neural_network" title="Spiking neural network">Spiking neural network</a></li> <li><a href="/wiki/Memtransistor" title="Memtransistor">Memtransistor</a></li> <li><a href="/wiki/Electrochemical_RAM" title="Electrochemical RAM">Electrochemical RAM</a> (ECRAM)</li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)"><a href="/wiki/Reinforcement_learning" title="Reinforcement learning">Reinforcement learning</a></div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Q-learning" title="Q-learning">Q-learning</a></li> <li><a href="/wiki/State%E2%80%93action%E2%80%93reward%E2%80%93state%E2%80%93action" title="State–action–reward–state–action">SARSA</a></li> <li><a href="/wiki/Temporal_difference_learning" title="Temporal difference learning">Temporal difference (TD)</a></li> <li><a href="/wiki/Multi-agent_reinforcement_learning" title="Multi-agent reinforcement learning">Multi-agent</a> <ul><li><a href="/wiki/Self-play_(reinforcement_learning_technique)" class="mw-redirect" title="Self-play (reinforcement learning technique)">Self-play</a></li></ul></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)">Learning with humans</div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Active_learning_(machine_learning)" title="Active learning (machine learning)">Active learning</a></li> <li><a href="/wiki/Crowdsourcing" title="Crowdsourcing">Crowdsourcing</a></li> <li><a href="/wiki/Human-in-the-loop" title="Human-in-the-loop">Human-in-the-loop</a></li> <li><a href="/wiki/Reinforcement_learning_from_human_feedback" title="Reinforcement learning from human feedback">RLHF</a></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)">Model diagnostics</div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Coefficient_of_determination" title="Coefficient of determination">Coefficient of determination</a></li> <li><a href="/wiki/Confusion_matrix" title="Confusion matrix">Confusion matrix</a></li> <li><a href="/wiki/Learning_curve_(machine_learning)" title="Learning curve (machine learning)">Learning curve</a></li> <li><a href="/wiki/Receiver_operating_characteristic" title="Receiver operating characteristic">ROC curve</a></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)">Mathematical foundations</div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Kernel_machines" class="mw-redirect" title="Kernel machines">Kernel machines</a></li> <li><a href="/wiki/Bias%E2%80%93variance_tradeoff" title="Bias–variance tradeoff">Bias–variance tradeoff</a></li> <li><a href="/wiki/Computational_learning_theory" title="Computational learning theory">Computational learning theory</a></li> <li><a href="/wiki/Empirical_risk_minimization" title="Empirical risk minimization">Empirical risk minimization</a></li> <li><a href="/wiki/Occam_learning" title="Occam learning">Occam learning</a></li> <li><a href="/wiki/Probably_approximately_correct_learning" title="Probably approximately correct learning">PAC learning</a></li> <li><a href="/wiki/Statistical_learning_theory" title="Statistical learning theory">Statistical learning</a></li> <li><a href="/wiki/Vapnik%E2%80%93Chervonenkis_theory" title="Vapnik–Chervonenkis theory">VC theory</a></li> <li><a href="/wiki/Topological_deep_learning" title="Topological deep learning">Topological deep learning</a></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)">Journals and conferences</div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/ECML_PKDD" title="ECML PKDD">ECML PKDD</a></li> <li><a href="/wiki/Conference_on_Neural_Information_Processing_Systems" title="Conference on Neural Information Processing Systems">NeurIPS</a></li> <li><a href="/wiki/International_Conference_on_Machine_Learning" title="International Conference on Machine Learning">ICML</a></li> <li><a href="/wiki/International_Conference_on_Learning_Representations" title="International Conference on Learning Representations">ICLR</a></li> <li><a href="/wiki/International_Joint_Conference_on_Artificial_Intelligence" title="International Joint Conference on Artificial Intelligence">IJCAI</a></li> <li><a href="/wiki/Machine_Learning_(journal)" title="Machine Learning (journal)">ML</a></li> <li><a href="/wiki/Journal_of_Machine_Learning_Research" title="Journal of Machine Learning Research">JMLR</a></li></ul></div></div></td> </tr><tr><td class="sidebar-content"> <div class="sidebar-list mw-collapsible mw-collapsed machine-learning-list-title"><div class="sidebar-list-title" style="border-top:1px solid #aaa; text-align:center;;color: var(--color-base)">Related articles</div><div class="sidebar-list-content mw-collapsible-content hlist"> <ul><li><a href="/wiki/Glossary_of_artificial_intelligence" title="Glossary of artificial intelligence">Glossary of artificial intelligence</a></li> <li><a href="/wiki/List_of_datasets_for_machine-learning_research" title="List of datasets for machine-learning research">List of datasets for machine-learning research</a> <ul><li><a href="/wiki/List_of_datasets_in_computer_vision_and_image_processing" title="List of datasets in computer vision and image processing">List of datasets in computer vision and image processing</a></li></ul></li> <li><a href="/wiki/Outline_of_machine_learning" title="Outline of machine learning">Outline of machine learning</a></li></ul></div></div></td> </tr><tr><td class="sidebar-navbar"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1129693374"><style data-mw-deduplicate="TemplateStyles:r1239400231">.mw-parser-output .navbar{display:inline;font-size:88%;font-weight:normal}.mw-parser-output .navbar-collapse{float:left;text-align:left}.mw-parser-output .navbar-boxtext{word-spacing:0}.mw-parser-output .navbar ul{display:inline-block;white-space:nowrap;line-height:inherit}.mw-parser-output .navbar-brackets::before{margin-right:-0.125em;content:"[ "}.mw-parser-output .navbar-brackets::after{margin-left:-0.125em;content:" ]"}.mw-parser-output .navbar li{word-spacing:-0.125em}.mw-parser-output .navbar a>span,.mw-parser-output .navbar a>abbr{text-decoration:inherit}.mw-parser-output .navbar-mini abbr{font-variant:small-caps;border-bottom:none;text-decoration:none;cursor:inherit}.mw-parser-output .navbar-ct-full{font-size:114%;margin:0 7em}.mw-parser-output .navbar-ct-mini{font-size:114%;margin:0 4em}html.skin-theme-clientpref-night .mw-parser-output .navbar li a abbr{color:var(--color-base)!important}@media(prefers-color-scheme:dark){html.skin-theme-clientpref-os .mw-parser-output .navbar li a abbr{color:var(--color-base)!important}}@media print{.mw-parser-output .navbar{display:none!important}}</style><div class="navbar plainlinks hlist navbar-mini"><ul><li class="nv-view"><a href="/wiki/Template:Machine_learning" title="Template:Machine learning"><abbr title="View this template">v</abbr></a></li><li class="nv-talk"><a href="/wiki/Template_talk:Machine_learning" title="Template talk:Machine learning"><abbr title="Discuss this template">t</abbr></a></li><li class="nv-edit"><a href="/wiki/Special:EditPage/Template:Machine_learning" title="Special:EditPage/Template:Machine learning"><abbr title="Edit this template">e</abbr></a></li></ul></div></td></tr></tbody></table> <figure class="mw-default-size mw-halign-right" typeof="mw:File/Thumb"><a href="/wiki/File:Full_GPT_architecture.svg" class="mw-file-description"><img src="//upload.wikimedia.org/wikipedia/commons/thumb/5/51/Full_GPT_architecture.svg/220px-Full_GPT_architecture.svg.png" decoding="async" width="220" height="264" class="mw-file-element" srcset="//upload.wikimedia.org/wikipedia/commons/thumb/5/51/Full_GPT_architecture.svg/330px-Full_GPT_architecture.svg.png 1.5x, //upload.wikimedia.org/wikipedia/commons/thumb/5/51/Full_GPT_architecture.svg/440px-Full_GPT_architecture.svg.png 2x" data-file-width="500" data-file-height="600" /></a><figcaption>Original GPT model</figcaption></figure><p>A <b>generative pre-trained transformer</b> (<b>GPT</b>) is a type of <a href="/wiki/Large_language_model" title="Large language model">large language model</a> (LLM)<sup id="cite_ref-:1_1-0" class="reference"><a href="#cite_note-:1-1"><span class="cite-bracket">&#91;</span>1<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-:0_2-0" class="reference"><a href="#cite_note-:0-2"><span class="cite-bracket">&#91;</span>2<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-:4_3-0" class="reference"><a href="#cite_note-:4-3"><span class="cite-bracket">&#91;</span>3<span class="cite-bracket">&#93;</span></a></sup> and a prominent framework for <a href="/wiki/Generative_artificial_intelligence" title="Generative artificial intelligence">generative artificial intelligence</a>.<sup id="cite_ref-4" class="reference"><a href="#cite_note-4"><span class="cite-bracket">&#91;</span>4<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-5" class="reference"><a href="#cite_note-5"><span class="cite-bracket">&#91;</span>5<span class="cite-bracket">&#93;</span></a></sup> It is an <a href="/wiki/Neural_network_(machine_learning)" title="Neural network (machine learning)">artificial neural network</a> that is used in <a href="/wiki/Natural_language_processing" title="Natural language processing">natural language processing</a> by machines.<sup id="cite_ref-6" class="reference"><a href="#cite_note-6"><span class="cite-bracket">&#91;</span>6<span class="cite-bracket">&#93;</span></a></sup> It is based on the <a href="/wiki/Transformer_(deep_learning_architecture)" title="Transformer (deep learning architecture)">transformer deep learning architecture</a>, pre-trained on large <a href="/wiki/Data_set" title="Data set">data sets</a> of unlabeled text, and able to generate novel human-like content.<sup id="cite_ref-:0_2-1" class="reference"><a href="#cite_note-:0-2"><span class="cite-bracket">&#91;</span>2<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-:4_3-1" class="reference"><a href="#cite_note-:4-3"><span class="cite-bracket">&#91;</span>3<span class="cite-bracket">&#93;</span></a></sup> As of 2023, most LLMs had these characteristics<sup id="cite_ref-7" class="reference"><a href="#cite_note-7"><span class="cite-bracket">&#91;</span>7<span class="cite-bracket">&#93;</span></a></sup> and are sometimes referred to broadly as GPTs.<sup id="cite_ref-8" class="reference"><a href="#cite_note-8"><span class="cite-bracket">&#91;</span>8<span class="cite-bracket">&#93;</span></a></sup> </p><p>The first GPT was introduced in 2018 by <a href="/wiki/OpenAI" title="OpenAI">OpenAI</a>.<sup id="cite_ref-gpt1_9-0" class="reference"><a href="#cite_note-gpt1-9"><span class="cite-bracket">&#91;</span>9<span class="cite-bracket">&#93;</span></a></sup> OpenAI has released significant <a href="#Foundational_models">GPT foundation models</a> that have been sequentially numbered, to comprise its "GPT-<i>n</i>" series.<sup id="cite_ref-10" class="reference"><a href="#cite_note-10"><span class="cite-bracket">&#91;</span>10<span class="cite-bracket">&#93;</span></a></sup> Each of these was significantly more capable than the previous, due to increased size (number of trainable parameters) and training. The most recent of these, <a href="/wiki/GPT-4o" title="GPT-4o">GPT-4o</a>, was released in May 2024.<sup id="cite_ref-11" class="reference"><a href="#cite_note-11"><span class="cite-bracket">&#91;</span>11<span class="cite-bracket">&#93;</span></a></sup> Such models have been the basis for their more <a href="#Task-specific_models">task-specific GPT systems</a>, including models <a href="/wiki/Instruction_tuning" class="mw-redirect" title="Instruction tuning">fine-tuned for instruction following</a>—which in turn power the <a href="/wiki/ChatGPT" title="ChatGPT">ChatGPT</a> <a href="/wiki/Chatbot" title="Chatbot">chatbot</a> service.<sup id="cite_ref-:1_1-1" class="reference"><a href="#cite_note-:1-1"><span class="cite-bracket">&#91;</span>1<span class="cite-bracket">&#93;</span></a></sup> </p><p>The term "GPT" is also used in the names and descriptions of such models developed by others. For example, other GPT foundation models include <a href="/wiki/EleutherAI#GPT_models" title="EleutherAI">a series of models</a> created by <a href="/wiki/EleutherAI" title="EleutherAI">EleutherAI</a>,<sup id="cite_ref-:9_12-0" class="reference"><a href="#cite_note-:9-12"><span class="cite-bracket">&#91;</span>12<span class="cite-bracket">&#93;</span></a></sup> and seven models created by <a href="/wiki/Cerebras" title="Cerebras">Cerebras</a> in 2023.<sup id="cite_ref-:10_13-0" class="reference"><a href="#cite_note-:10-13"><span class="cite-bracket">&#91;</span>13<span class="cite-bracket">&#93;</span></a></sup> Companies in different industries have developed task-specific GPTs in their respective fields, such as <a href="/wiki/Salesforce" title="Salesforce">Salesforce</a>'s "EinsteinGPT" (for <a href="/wiki/Customer_relationship_management" title="Customer relationship management">CRM</a>)<sup id="cite_ref-14" class="reference"><a href="#cite_note-14"><span class="cite-bracket">&#91;</span>14<span class="cite-bracket">&#93;</span></a></sup> and <a href="/wiki/Bloomberg_L.P." title="Bloomberg L.P.">Bloomberg</a>'s "BloombergGPT" (for finance).<sup id="cite_ref-15" class="reference"><a href="#cite_note-15"><span class="cite-bracket">&#91;</span>15<span class="cite-bracket">&#93;</span></a></sup> </p> <meta property="mw:PageProp/toc" /> <div class="mw-heading mw-heading2"><h2 id="History">History</h2><span class="mw-editsection"><span class="mw-editsection-bracket">[</span><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit&amp;section=1" title="Edit section: History"><span>edit</span></a><span class="mw-editsection-bracket">]</span></span></div> <div class="mw-heading mw-heading3"><h3 id="Initial_developments">Initial developments</h3><span class="mw-editsection"><span class="mw-editsection-bracket">[</span><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit&amp;section=2" title="Edit section: Initial developments"><span>edit</span></a><span class="mw-editsection-bracket">]</span></span></div> <p>Generative pretraining (GP) was a long-established concept in machine learning applications.<sup id="cite_ref-16" class="reference"><a href="#cite_note-16"><span class="cite-bracket">&#91;</span>16<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-17" class="reference"><a href="#cite_note-17"><span class="cite-bracket">&#91;</span>17<span class="cite-bracket">&#93;</span></a></sup> It was originally used as a form of <a href="/wiki/Semi-supervised_learning" class="mw-redirect" title="Semi-supervised learning">semi-supervised learning</a>, as the model is trained first on an unlabelled dataset (<i>pretraining</i> step) by learning to <i>generate</i> datapoints in the dataset, and then it is trained to classify a labelled dataset.<sup id="cite_ref-18" class="reference"><a href="#cite_note-18"><span class="cite-bracket">&#91;</span>18<span class="cite-bracket">&#93;</span></a></sup> </p><p>There were three main types of early GP. The <a href="/wiki/Hidden_Markov_model" title="Hidden Markov model">hidden Markov models</a> learn a generative model of sequences for downstream applications. For example, in <a href="/wiki/Speech_recognition" title="Speech recognition">speech recognition</a>, a trained HMM infers the most likely hidden sequence for a speech signal, and the hidden sequence is taken as the phonemes of the speech signal. These were developed in the 1970s and became widely applied in speech recognition in the 1980s.<sup id="cite_ref-19" class="reference"><a href="#cite_note-19"><span class="cite-bracket">&#91;</span>19<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-20" class="reference"><a href="#cite_note-20"><span class="cite-bracket">&#91;</span>20<span class="cite-bracket">&#93;</span></a></sup> </p><p>The compressors learn to compress data such as images and textual sequences, and the compressed data serves as a good representation for downstream applications such as <a href="/wiki/Facial_recognition_system" title="Facial recognition system">facial recognition</a>.<sup id="cite_ref-21" class="reference"><a href="#cite_note-21"><span class="cite-bracket">&#91;</span>21<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-22" class="reference"><a href="#cite_note-22"><span class="cite-bracket">&#91;</span>22<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-schmidhuber1992_23-0" class="reference"><a href="#cite_note-schmidhuber1992-23"><span class="cite-bracket">&#91;</span>23<span class="cite-bracket">&#93;</span></a></sup> The <a href="/wiki/Autoencoder" title="Autoencoder">autoencoders</a> similarly learn a latent representation of data for later downstream applications such as speech recognition.<sup id="cite_ref-24" class="reference"><a href="#cite_note-24"><span class="cite-bracket">&#91;</span>24<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-:14_25-0" class="reference"><a href="#cite_note-:14-25"><span class="cite-bracket">&#91;</span>25<span class="cite-bracket">&#93;</span></a></sup> The connection between autoencoders and algorithmic compressors was noted in 1993.<sup id="cite_ref-26" class="reference"><a href="#cite_note-26"><span class="cite-bracket">&#91;</span>26<span class="cite-bracket">&#93;</span></a></sup> </p> <link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1236090951"><div role="note" class="hatnote navigation-not-searchable">See also: <a href="/wiki/Transformer_(deep_learning_architecture)#History" title="Transformer (deep learning architecture)">Transformer (deep learning architecture) §&#160;History</a></div> <p>During the 2010s, the problem of machine translation was solved<sup class="noprint Inline-Template Template-Fact" style="white-space:nowrap;">&#91;<i><a href="/wiki/Wikipedia:Citation_needed" title="Wikipedia:Citation needed"><span title="This claim needs references to reliable sources. (December 2024)">citation needed</span></a></i>&#93;</sup> by <a href="/wiki/Recurrent_neural_network" title="Recurrent neural network">recurrent neural networks</a>, with <a href="/wiki/Attention_(machine_learning)" title="Attention (machine learning)">attention mechanism</a> added. This was optimized into the <a href="/wiki/Transformer_(deep_learning_architecture)" title="Transformer (deep learning architecture)">transformer</a> architecture, published by <a href="/wiki/Google" title="Google">Google</a> researchers in <i><a href="/wiki/Attention_Is_All_You_Need" title="Attention Is All You Need">Attention Is All You Need</a></i> (2017).<sup id="cite_ref-27" class="reference"><a href="#cite_note-27"><span class="cite-bracket">&#91;</span>27<span class="cite-bracket">&#93;</span></a></sup> That development led to the emergence of <a href="/wiki/Large_language_model" title="Large language model">large language models</a> such as <a href="/wiki/BERT_(language_model)" title="BERT (language model)">BERT</a> (2018)<sup id="cite_ref-28" class="reference"><a href="#cite_note-28"><span class="cite-bracket">&#91;</span>28<span class="cite-bracket">&#93;</span></a></sup> which was a pre-trained transformer (PT) but not designed to be <a href="/wiki/Generative_artificial_intelligence" title="Generative artificial intelligence">generative</a> (BERT was an "<a href="/wiki/Transformer_(deep_learning_architecture)#encoder-only" title="Transformer (deep learning architecture)">encoder-only</a>" model). Also in 2018, <a href="/wiki/OpenAI" title="OpenAI">OpenAI</a> published <i>Improving Language Understanding by Generative Pre-Training</i>, which introduced <a href="/wiki/GPT-1" title="GPT-1">GPT-1</a>, the first in its GPT series.<sup id="cite_ref-gpt1paper_29-0" class="reference"><a href="#cite_note-gpt1paper-29"><span class="cite-bracket">&#91;</span>29<span class="cite-bracket">&#93;</span></a></sup> </p><p>Previously in 2017, some of the authors who would later work on GPT-1 worked on generative pre-training of language with <a href="/wiki/Long_short-term_memory" title="Long short-term memory">LSTM</a>, which resulted in a model that could represent text with vectors that could easily be fine-tuned for downstream applications.<sup id="cite_ref-30" class="reference"><a href="#cite_note-30"><span class="cite-bracket">&#91;</span>30<span class="cite-bracket">&#93;</span></a></sup> </p><p>Prior to transformer-based architectures, the best-performing neural NLP (<a href="/wiki/Natural_language_processing" title="Natural language processing">natural language processing</a>) models commonly employed <a href="/wiki/Supervised_learning" title="Supervised learning">supervised learning</a> from large amounts of manually-labeled data. The reliance on supervised learning limited their use on datasets that were not well-annotated, and also made it prohibitively expensive and time-consuming to train extremely large language models.<sup id="cite_ref-gpt1paper_29-1" class="reference"><a href="#cite_note-gpt1paper-29"><span class="cite-bracket">&#91;</span>29<span class="cite-bracket">&#93;</span></a></sup> </p><p>The <a href="/wiki/Semi-supervised_learning" class="mw-redirect" title="Semi-supervised learning">semi-supervised</a> approach OpenAI employed to make a large-scale generative system—and was first to do with a transformer model—involved two stages: an <a href="/wiki/Unsupervised_learning" title="Unsupervised learning">unsupervised</a> <a href="/wiki/Generative_model" title="Generative model">generative</a> "pretraining" stage to set initial parameters using a language modeling objective, and a supervised <a href="/wiki/Discriminative_model" title="Discriminative model">discriminative</a> "<a href="/wiki/Fine-tuning_(machine_learning)" class="mw-redirect" title="Fine-tuning (machine learning)">fine-tuning</a>" stage to adapt these parameters to a target task.<sup id="cite_ref-gpt1paper_29-2" class="reference"><a href="#cite_note-gpt1paper-29"><span class="cite-bracket">&#91;</span>29<span class="cite-bracket">&#93;</span></a></sup> </p> <div class="mw-heading mw-heading3"><h3 id="Later_developments">Later developments</h3><span class="mw-editsection"><span class="mw-editsection-bracket">[</span><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit&amp;section=3" title="Edit section: Later developments"><span>edit</span></a><span class="mw-editsection-bracket">]</span></span></div> <p>Regarding more recent <a href="#Foundational_models">GPT foundation models</a>, <a href="/wiki/OpenAI" title="OpenAI">OpenAI</a> published its first versions of <a href="/wiki/GPT-3" title="GPT-3">GPT-3</a> in July 2020. There were three models, with 1B, 6.7B, 175B parameters, respectively named <i>babbage, curie, and davinci</i> (giving initials B, C, and D).<sup class="noprint Inline-Template Template-Fact" style="white-space:nowrap;">&#91;<i><a href="/wiki/Wikipedia:Citation_needed" title="Wikipedia:Citation needed"><span title="This claim needs references to reliable sources. (November 2023)">citation needed</span></a></i>&#93;</sup> </p><p>In July 2021, OpenAI published <a href="/wiki/OpenAI_Codex" title="OpenAI Codex">Codex</a>, a <a href="#Task-specific_models">task-specific GPT model</a> targeted for programming applications. This was developed by fine-tuning a 12B parameter version of GPT-3 (different from previous GPT-3 models) using code from <a href="/wiki/GitHub" title="GitHub">GitHub</a>.<sup id="cite_ref-31" class="reference"><a href="#cite_note-31"><span class="cite-bracket">&#91;</span>31<span class="cite-bracket">&#93;</span></a></sup> </p><p>In March 2022, OpenAI published two versions of GPT-3 that were fine-tuned for instruction-following (instruction-tuned), named <i>davinci-instruct-beta</i> (175B) and <i>text-davinci-001</i>,<sup id="cite_ref-32" class="reference"><a href="#cite_note-32"><span class="cite-bracket">&#91;</span>32<span class="cite-bracket">&#93;</span></a></sup> and then started beta testing <i>code-davinci-002</i>.<sup id="cite_ref-33" class="reference"><a href="#cite_note-33"><span class="cite-bracket">&#91;</span>33<span class="cite-bracket">&#93;</span></a></sup> <i>text-davinci-002</i> was instruction-tuned from <i>code-davinci-002</i>. Both <i>text-davinci-003</i> and <a href="/wiki/ChatGPT" title="ChatGPT">ChatGPT</a> were released in November 2022, with both building upon <i>text-davinci-002</i> via reinforcement learning from human feedback (RLHF). <i>text-davinci-003</i> is trained for following instructions (like its predecessors), whereas ChatGPT is further trained for conversational interaction with a human user.<sup id="cite_ref-fu2022_34-0" class="reference"><a href="#cite_note-fu2022-34"><span class="cite-bracket">&#91;</span>34<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-35" class="reference"><a href="#cite_note-35"><span class="cite-bracket">&#91;</span>35<span class="cite-bracket">&#93;</span></a></sup> </p><p>OpenAI's most recent GPT foundation model, <a href="/wiki/GPT-4" title="GPT-4">GPT-4</a>, was released on March 14, 2023. It can be accessed directly by users via a premium version of ChatGPT, and is available to developers for incorporation into other products and services via OpenAI's <a href="/wiki/API" title="API">API</a>. Other producers of GPT foundation models include <a href="/wiki/EleutherAI" title="EleutherAI">EleutherAI</a> (with <a href="/wiki/EleutherAI#GPT_models" title="EleutherAI">a series of models</a> starting in March 2021)<sup id="cite_ref-:9_12-1" class="reference"><a href="#cite_note-:9-12"><span class="cite-bracket">&#91;</span>12<span class="cite-bracket">&#93;</span></a></sup> and <a href="/wiki/Cerebras" title="Cerebras">Cerebras</a> (with seven models released in March 2023).<sup id="cite_ref-:10_13-1" class="reference"><a href="#cite_note-:10-13"><span class="cite-bracket">&#91;</span>13<span class="cite-bracket">&#93;</span></a></sup> </p> <div class="mw-heading mw-heading2"><h2 id="Foundation_models">Foundation models</h2><span class="mw-editsection"><span class="mw-editsection-bracket">[</span><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit&amp;section=4" title="Edit section: Foundation models"><span>edit</span></a><span class="mw-editsection-bracket">]</span></span></div> <p>A <a href="/wiki/Foundation_model" title="Foundation model">foundation model</a> is an AI model trained on broad data at scale such that it can be adapted to a wide range of downstream tasks.<sup id="cite_ref-36" class="reference"><a href="#cite_note-36"><span class="cite-bracket">&#91;</span>36<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-37" class="reference"><a href="#cite_note-37"><span class="cite-bracket">&#91;</span>37<span class="cite-bracket">&#93;</span></a></sup> </p><p>Thus far, the most notable GPT foundation models have been from <a href="/wiki/OpenAI" title="OpenAI">OpenAI</a>'s <i>GPT-n</i> series. The most recent from that is <a href="/wiki/GPT-4" title="GPT-4">GPT-4</a>, for which OpenAI declined to publish the size or training details (citing "the competitive landscape and the safety implications of large-scale models").<sup id="cite_ref-gpt4-report_38-0" class="reference"><a href="#cite_note-gpt4-report-38"><span class="cite-bracket">&#91;</span>38<span class="cite-bracket">&#93;</span></a></sup> </p> <table class="wikitable"> <caption>OpenAI's <i>GPT-n</i> series </caption> <tbody><tr> <th>Model </th> <th>Architecture </th> <th>Parameter count </th> <th>Training data </th> <th>Release date </th> <th>Training cost </th></tr> <tr> <td><a href="/wiki/GPT-1" title="GPT-1">GPT-1</a> </td> <td>12-level, 12-headed Transformer decoder (no encoder), followed by linear-softmax. </td> <td>117 million </td> <td><a href="/wiki/BookCorpus" title="BookCorpus">BookCorpus</a>:<sup id="cite_ref-39" class="reference"><a href="#cite_note-39"><span class="cite-bracket">&#91;</span>39<span class="cite-bracket">&#93;</span></a></sup> 4.5 GB of text, from 7,000 unpublished books of various genres. </td> <td><span data-sort-value="000000002018-06-11-0000" style="white-space:nowrap">June 11, 2018</span><sup id="cite_ref-gpt1_9-1" class="reference"><a href="#cite_note-gpt1-9"><span class="cite-bracket">&#91;</span>9<span class="cite-bracket">&#93;</span></a></sup> </td> <td>30 days on 8 <a href="/wiki/Quadro" title="Quadro">P600</a> graphics cards, or 1 peta<a href="/wiki/FLOPS" class="mw-redirect" title="FLOPS">FLOPS</a>-day.<sup id="cite_ref-gpt1_9-2" class="reference"><a href="#cite_note-gpt1-9"><span class="cite-bracket">&#91;</span>9<span class="cite-bracket">&#93;</span></a></sup> </td></tr> <tr> <td><a href="/wiki/GPT-2" title="GPT-2">GPT-2</a> </td> <td>GPT-1, but with modified normalization </td> <td>1.5 billion </td> <td>WebText: 40 GB of text, 8 million documents, from 45 million webpages <a href="/wiki/Upvote" class="mw-redirect" title="Upvote">upvoted</a> on <a href="/wiki/Reddit" title="Reddit">Reddit</a>. </td> <td><span data-sort-value="000000002019-02-14-0000" style="white-space:nowrap">February 14, 2019</span> (initial/limited version) and <span data-sort-value="000000002019-11-05-0000" style="white-space:nowrap">November 5, 2019</span> (full version)<sup id="cite_ref-40" class="reference"><a href="#cite_note-40"><span class="cite-bracket">&#91;</span>40<span class="cite-bracket">&#93;</span></a></sup> </td> <td>"tens of petaflop/s-day",<sup id="cite_ref-:2_41-0" class="reference"><a href="#cite_note-:2-41"><span class="cite-bracket">&#91;</span>41<span class="cite-bracket">&#93;</span></a></sup> or 1.5e21 FLOPS.<sup id="cite_ref-:3_42-0" class="reference"><a href="#cite_note-:3-42"><span class="cite-bracket">&#91;</span>42<span class="cite-bracket">&#93;</span></a></sup> </td></tr> <tr> <td><a href="/wiki/GPT-3" title="GPT-3">GPT-3</a> </td> <td>GPT-2, but with modification to allow larger scaling </td> <td>175 billion<sup id="cite_ref-:8_43-0" class="reference"><a href="#cite_note-:8-43"><span class="cite-bracket">&#91;</span>43<span class="cite-bracket">&#93;</span></a></sup> </td> <td>499 billion tokens consisting of <a href="/wiki/Common_Crawl" title="Common Crawl">CommonCrawl</a> (570 GB), WebText, English Wikipedia, and two books corpora (Books1 and Books2). </td> <td><span data-sort-value="000000002020-05-28-0000" style="white-space:nowrap">May 28, 2020</span><sup id="cite_ref-:2_41-1" class="reference"><a href="#cite_note-:2-41"><span class="cite-bracket">&#91;</span>41<span class="cite-bracket">&#93;</span></a></sup> </td> <td>3640 petaflop/s-day (Table D.1<sup id="cite_ref-:2_41-2" class="reference"><a href="#cite_note-:2-41"><span class="cite-bracket">&#91;</span>41<span class="cite-bracket">&#93;</span></a></sup>), or 3.1e23 FLOPS.<sup id="cite_ref-:3_42-1" class="reference"><a href="#cite_note-:3-42"><span class="cite-bracket">&#91;</span>42<span class="cite-bracket">&#93;</span></a></sup> </td></tr> <tr> <td><a href="/wiki/GPT-3.5" class="mw-redirect" title="GPT-3.5">GPT-3.5</a> </td> <td>Undisclosed </td> <td>175 billion<sup id="cite_ref-:8_43-1" class="reference"><a href="#cite_note-:8-43"><span class="cite-bracket">&#91;</span>43<span class="cite-bracket">&#93;</span></a></sup> </td> <td>Undisclosed </td> <td>March 15, 2022 </td> <td>Undisclosed </td></tr> <tr> <td><a href="/wiki/GPT-4" title="GPT-4">GPT-4</a> </td> <td>Also trained with both text prediction and <a href="/wiki/Reinforcement_learning_from_human_feedback" title="Reinforcement learning from human feedback">RLHF</a>; accepts <a href="/wiki/Multimodal_learning" title="Multimodal learning">both text and images</a> as input. Further details are not public.<sup id="cite_ref-gpt4-report_38-1" class="reference"><a href="#cite_note-gpt4-report-38"><span class="cite-bracket">&#91;</span>38<span class="cite-bracket">&#93;</span></a></sup> </td> <td>Undisclosed. Estimated 1.7 trillion.<sup id="cite_ref-44" class="reference"><a href="#cite_note-44"><span class="cite-bracket">&#91;</span>44<span class="cite-bracket">&#93;</span></a></sup> </td> <td>Undisclosed </td> <td><span data-sort-value="000000002023-03-14-0000" style="white-space:nowrap">March 14, 2023</span> </td> <td>Undisclosed. Estimated 2.1 × 10<sup>25</sup> FLOPS.<sup id="cite_ref-:3_42-2" class="reference"><a href="#cite_note-:3-42"><span class="cite-bracket">&#91;</span>42<span class="cite-bracket">&#93;</span></a></sup> </td></tr></tbody></table> <p>Other such models include <a href="/wiki/Google" title="Google">Google</a>'s <a href="/wiki/PaLM" title="PaLM">PaLM</a>, a broad foundation model that has been compared to <a href="/wiki/GPT-3" title="GPT-3">GPT-3</a> and have been made available to developers via an <a href="/wiki/API" title="API">API</a>,<sup id="cite_ref-45" class="reference"><a href="#cite_note-45"><span class="cite-bracket">&#91;</span>45<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-46" class="reference"><a href="#cite_note-46"><span class="cite-bracket">&#91;</span>46<span class="cite-bracket">&#93;</span></a></sup> and Together's <b>GPT-JT</b>, which has been reported as the closest-performing <a href="/wiki/Open-source" class="mw-redirect" title="Open-source">open-source</a> alternative to <a href="/wiki/GPT-3" title="GPT-3">GPT-3</a> (and is derived from <a href="/wiki/EleutherAI#GPT_models" title="EleutherAI">earlier open-source GPTs</a>).<sup id="cite_ref-47" class="reference"><a href="#cite_note-47"><span class="cite-bracket">&#91;</span>47<span class="cite-bracket">&#93;</span></a></sup> <a href="/wiki/Meta_AI" title="Meta AI">Meta AI</a> (formerly <a href="/wiki/Facebook" title="Facebook">Facebook</a>) also has a generative transformer-based foundational large language model, known as <a href="/wiki/LLaMA" class="mw-redirect" title="LLaMA">LLaMA</a>.<sup id="cite_ref-48" class="reference"><a href="#cite_note-48"><span class="cite-bracket">&#91;</span>48<span class="cite-bracket">&#93;</span></a></sup> </p><p>Foundational GPTs can also employ <a href="/wiki/Modality_(human%E2%80%93computer_interaction)" title="Modality (human–computer interaction)">modalities</a> other than text, for input and/or output. <a href="/wiki/GPT-4" title="GPT-4">GPT-4</a> is a multi-modal LLM that is capable of processing text and image input (though its output is limited to text).<sup id="cite_ref-49" class="reference"><a href="#cite_note-49"><span class="cite-bracket">&#91;</span>49<span class="cite-bracket">&#93;</span></a></sup> Regarding multimodal <i>output</i>, some generative transformer-based models are used for <a href="/wiki/Text-to-image_model" title="Text-to-image model">text-to-image</a> technologies such as <a href="/wiki/Diffusion_model" title="Diffusion model">diffusion</a><sup id="cite_ref-50" class="reference"><a href="#cite_note-50"><span class="cite-bracket">&#91;</span>50<span class="cite-bracket">&#93;</span></a></sup> and parallel decoding.<sup id="cite_ref-51" class="reference"><a href="#cite_note-51"><span class="cite-bracket">&#91;</span>51<span class="cite-bracket">&#93;</span></a></sup> Such kinds of models can serve as <b>visual foundation models</b> (VFMs) for developing downstream systems that can work with images.<sup id="cite_ref-52" class="reference"><a href="#cite_note-52"><span class="cite-bracket">&#91;</span>52<span class="cite-bracket">&#93;</span></a></sup> </p> <div class="mw-heading mw-heading2"><h2 id="Task-specific_models">Task-specific models</h2><span class="mw-editsection"><span class="mw-editsection-bracket">[</span><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit&amp;section=5" title="Edit section: Task-specific models"><span>edit</span></a><span class="mw-editsection-bracket">]</span></span></div> <p>A foundational GPT model can be further adapted to produce more targeted systems directed to specific tasks and/or subject-matter domains. Methods for such adaptation can include additional <a href="/wiki/Fine-tuning_(machine_learning)" class="mw-redirect" title="Fine-tuning (machine learning)">fine-tuning</a> (beyond that done for the foundation model) as well as certain forms of <a href="/wiki/Prompt_engineering" title="Prompt engineering">prompt engineering</a>.<sup id="cite_ref-53" class="reference"><a href="#cite_note-53"><span class="cite-bracket">&#91;</span>53<span class="cite-bracket">&#93;</span></a></sup> </p><p>An important example of this is <a href="/wiki/Instruction_tuning" class="mw-redirect" title="Instruction tuning">fine-tuning models to follow instructions</a>, which is of course a fairly broad task but more targeted than a foundation model. In January 2022, <a href="/wiki/OpenAI" title="OpenAI">OpenAI</a> introduced "InstructGPT"—a series of models which were fine-tuned to follow instructions using a combination of <a href="/wiki/Supervised_learning" title="Supervised learning">supervised</a> training and <a href="/wiki/Reinforcement_learning_from_human_feedback" title="Reinforcement learning from human feedback">reinforcement learning from human feedback</a> (RLHF) on base GPT-3 language models.<sup id="cite_ref-instructgpt-blog_54-0" class="reference"><a href="#cite_note-instructgpt-blog-54"><span class="cite-bracket">&#91;</span>54<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-instructgpt-paper_55-0" class="reference"><a href="#cite_note-instructgpt-paper-55"><span class="cite-bracket">&#91;</span>55<span class="cite-bracket">&#93;</span></a></sup> Advantages this had over the bare foundational models included higher accuracy, less negative/toxic sentiment, and generally better alignment with user needs. Hence, OpenAI began using this as the basis for its <a href="/wiki/API" title="API">API</a> service offerings.<sup id="cite_ref-56" class="reference"><a href="#cite_note-56"><span class="cite-bracket">&#91;</span>56<span class="cite-bracket">&#93;</span></a></sup> Other instruction-tuned models have been released by others, including a fully open version.<sup id="cite_ref-57" class="reference"><a href="#cite_note-57"><span class="cite-bracket">&#91;</span>57<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-58" class="reference"><a href="#cite_note-58"><span class="cite-bracket">&#91;</span>58<span class="cite-bracket">&#93;</span></a></sup> </p><p>Another (related) kind of task-specific models are <a href="/wiki/Chatbots" class="mw-redirect" title="Chatbots">chatbots</a>, which engage in human-like conversation. In November 2022, OpenAI launched <a href="/wiki/ChatGPT" title="ChatGPT">ChatGPT</a>—an online chat interface powered by an instruction-tuned language model trained in a similar fashion to InstructGPT.<sup id="cite_ref-chatgpt-blog_59-0" class="reference"><a href="#cite_note-chatgpt-blog-59"><span class="cite-bracket">&#91;</span>59<span class="cite-bracket">&#93;</span></a></sup> They trained this model using RLHF, with human AI trainers providing conversations in which they played both the user and the AI, and mixed this new dialogue dataset with the InstructGPT dataset for a conversational format suitable for a chatbot. Other major chatbots currently include <a href="/wiki/Microsoft" title="Microsoft">Microsoft</a>'s <a href="/wiki/Bing_Chat" class="mw-redirect" title="Bing Chat">Bing Chat</a>, which uses OpenAI's <a href="/wiki/GPT-4" title="GPT-4">GPT-4</a> (as part of a broader close collaboration between OpenAI and Microsoft),<sup id="cite_ref-60" class="reference"><a href="#cite_note-60"><span class="cite-bracket">&#91;</span>60<span class="cite-bracket">&#93;</span></a></sup> and <a href="/wiki/Google" title="Google">Google</a>'s competing chatbot <a href="/wiki/Gemini_(chatbot)" title="Gemini (chatbot)">Gemini</a> (initially based on their <a href="/wiki/LaMDA" title="LaMDA">LaMDA</a> family of conversation-trained language models, with plans to switch to <a href="/wiki/PaLM" title="PaLM">PaLM</a>).<sup id="cite_ref-61" class="reference"><a href="#cite_note-61"><span class="cite-bracket">&#91;</span>61<span class="cite-bracket">&#93;</span></a></sup> </p><p>Yet another kind of task that a GPT can be used for is the <a href="/wiki/Meta_(prefix)" title="Meta (prefix)">meta</a>-task of generating <i>its own</i> instructions, like developing a series of prompts for 'itself' to be able to effectuate a more general goal given by a human user.<sup id="cite_ref-62" class="reference"><a href="#cite_note-62"><span class="cite-bracket">&#91;</span>62<span class="cite-bracket">&#93;</span></a></sup> This is known as an AI <a href="/wiki/Software_agent" title="Software agent">agent</a>, and more specifically a recursive one because it uses results from its previous self-instructions to help it form its subsequent prompts; the first major example of this was <a href="/wiki/Auto-GPT" class="mw-redirect" title="Auto-GPT">Auto-GPT</a> (which uses OpenAI's GPT models), and others have since been developed as well.<sup id="cite_ref-63" class="reference"><a href="#cite_note-63"><span class="cite-bracket">&#91;</span>63<span class="cite-bracket">&#93;</span></a></sup> </p> <div class="mw-heading mw-heading3"><h3 id="Multimodality">Multimodality</h3><span class="mw-editsection"><span class="mw-editsection-bracket">[</span><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit&amp;section=6" title="Edit section: Multimodality"><span>edit</span></a><span class="mw-editsection-bracket">]</span></span></div> <p>Generative transformer-based systems can also be targeted for tasks involving <a href="/wiki/Modality_(human%E2%80%93computer_interaction)" title="Modality (human–computer interaction)">modalities</a> beyond text. For example, <a href="/wiki/Microsoft" title="Microsoft">Microsoft</a><span class="nowrap" style="padding-left:0.1em;">&#39;s</span> "Visual ChatGPT" combines ChatGPT with visual foundation models (VFMs) to enable input or output comprising images as well as text.<sup id="cite_ref-64" class="reference"><a href="#cite_note-64"><span class="cite-bracket">&#91;</span>64<span class="cite-bracket">&#93;</span></a></sup> Also, advances in <i>text-to-speech</i> technology offer tools for audio content creation when used in conjunction with foundational GPT language models.<sup id="cite_ref-65" class="reference"><a href="#cite_note-65"><span class="cite-bracket">&#91;</span>65<span class="cite-bracket">&#93;</span></a></sup> </p> <div class="mw-heading mw-heading3"><h3 id="Domain-specificity">Domain-specificity</h3><span class="mw-editsection"><span class="mw-editsection-bracket">[</span><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit&amp;section=7" title="Edit section: Domain-specificity"><span>edit</span></a><span class="mw-editsection-bracket">]</span></span></div> <p>GPT systems can be directed toward particular fields or domains. Some reported examples of such models and apps are as follows: </p> <ul><li>EinsteinGPT – for sales and marketing domains, to aid with customer relationship management (uses <a href="/wiki/GPT-3.5" class="mw-redirect" title="GPT-3.5">GPT-3.5</a>)<sup id="cite_ref-66" class="reference"><a href="#cite_note-66"><span class="cite-bracket">&#91;</span>66<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-67" class="reference"><a href="#cite_note-67"><span class="cite-bracket">&#91;</span>67<span class="cite-bracket">&#93;</span></a></sup></li> <li>BloombergGPT – for the financial domain, to aid with financial news and information (uses "freely available" AI methods, combined with their proprietary data)<sup id="cite_ref-68" class="reference"><a href="#cite_note-68"><span class="cite-bracket">&#91;</span>68<span class="cite-bracket">&#93;</span></a></sup></li> <li>Khanmigo – described as a GPT version for tutoring, in the education domain, it aids students using <a href="/wiki/Khan_Academy" title="Khan Academy">Khan Academy</a> by guiding them through their studies without directly providing answers (powered by <a href="/wiki/GPT-4" title="GPT-4">GPT-4</a>)<sup id="cite_ref-69" class="reference"><a href="#cite_note-69"><span class="cite-bracket">&#91;</span>69<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-70" class="reference"><a href="#cite_note-70"><span class="cite-bracket">&#91;</span>70<span class="cite-bracket">&#93;</span></a></sup></li> <li>SlackGPT – for the <a href="/wiki/Slack_(software)" title="Slack (software)">Slack</a> instant-messaging service, to aid with navigating and summarizing discussions on it (uses <a href="/wiki/OpenAI" title="OpenAI">OpenAI</a>'s <a href="/wiki/API" title="API">API</a>)<sup id="cite_ref-71" class="reference"><a href="#cite_note-71"><span class="cite-bracket">&#91;</span>71<span class="cite-bracket">&#93;</span></a></sup></li> <li>BioGPT – for the biomedical domain, to aid with biomedical literature text generation and mining (uses <a href="/wiki/GPT-2" title="GPT-2">GPT-2</a>)<sup id="cite_ref-72" class="reference"><a href="#cite_note-72"><span class="cite-bracket">&#91;</span>72<span class="cite-bracket">&#93;</span></a></sup></li></ul> <p>Sometimes domain-specificity is accomplished via software <a href="/wiki/Plug-in_(computing)" title="Plug-in (computing)">plug-ins or add-ons</a>. For example, several different companies have developed particular plugins that interact directly with OpenAI's <a href="/wiki/ChatGPT" title="ChatGPT">ChatGPT</a> interface,<sup id="cite_ref-73" class="reference"><a href="#cite_note-73"><span class="cite-bracket">&#91;</span>73<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-74" class="reference"><a href="#cite_note-74"><span class="cite-bracket">&#91;</span>74<span class="cite-bracket">&#93;</span></a></sup> and <a href="/wiki/Google_Workspace" title="Google Workspace">Google Workspace</a> has available add-ons such as "GPT for Sheets and Docs"—which is reported to aid use of <a href="/wiki/Spreadsheet" title="Spreadsheet">spreadsheet</a> functionality in <a href="/wiki/Google_Sheets" title="Google Sheets">Google Sheets</a>.<sup id="cite_ref-75" class="reference"><a href="#cite_note-75"><span class="cite-bracket">&#91;</span>75<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-76" class="reference"><a href="#cite_note-76"><span class="cite-bracket">&#91;</span>76<span class="cite-bracket">&#93;</span></a></sup> </p><p>In November 2023, OpenAI announced that ChatGPT Plus subscribers would be able to create <i>custom versions of ChatGPT</i> (being called <i>GPTs</i>).<sup id="cite_ref-77" class="reference"><a href="#cite_note-77"><span class="cite-bracket">&#91;</span>77<span class="cite-bracket">&#93;</span></a></sup> These can be tailored for specific domains via prompt engineering, curated datasets, and/or targeted interaction with external tools. Users who register as verified builders are able to publish their custom GPTs for other users, with monetization potential. (This is notably distinct from OpenAI's API service, as this is based internally within OpenAI's platform.) </p> <div class="mw-heading mw-heading2"><h2 id="Brand_issues">Brand issues</h2><span class="mw-editsection"><span class="mw-editsection-bracket">[</span><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit&amp;section=8" title="Edit section: Brand issues"><span>edit</span></a><span class="mw-editsection-bracket">]</span></span></div> <p><a href="/wiki/OpenAI" title="OpenAI">OpenAI</a>, which created the first generative pre-trained transformer (GPT) in 2018, has recently asserted that "GPT" should be regarded as a <i>brand</i> of OpenAI.<sup id="cite_ref-:5_78-0" class="reference"><a href="#cite_note-:5-78"><span class="cite-bracket">&#91;</span>78<span class="cite-bracket">&#93;</span></a></sup> In April 2023, OpenAI revised the brand guidelines in its <a href="/wiki/Terms_of_service" title="Terms of service">terms of service</a> to indicate that other businesses using its <a href="/wiki/API" title="API">API</a> to run their artificial intelligence (AI) services would no longer be able to include "GPT" in such names or branding.<sup id="cite_ref-79" class="reference"><a href="#cite_note-79"><span class="cite-bracket">&#91;</span>79<span class="cite-bracket">&#93;</span></a></sup> In May 2023, OpenAI engaged a brand management service to notify its API customers of this policy, although these notifications stopped short of making overt legal claims (such as allegations of <a href="/wiki/Trademark_infringement" title="Trademark infringement">trademark infringement</a> or demands to <a href="/wiki/Cease_and_desist" title="Cease and desist">cease and desist</a>).<sup id="cite_ref-:5_78-1" class="reference"><a href="#cite_note-:5-78"><span class="cite-bracket">&#91;</span>78<span class="cite-bracket">&#93;</span></a></sup> As of November 2023, OpenAI still prohibits its API licensees from naming their own products with "GPT",<sup id="cite_ref-:11_80-0" class="reference"><a href="#cite_note-:11-80"><span class="cite-bracket">&#91;</span>80<span class="cite-bracket">&#93;</span></a></sup> but it has begun enabling its ChatGPT Plus subscribers to make "custom versions of ChatGPT" that are being called <i>GPTs</i> on the OpenAI site.<sup id="cite_ref-81" class="reference"><a href="#cite_note-81"><span class="cite-bracket">&#91;</span>81<span class="cite-bracket">&#93;</span></a></sup> OpenAI's terms of service says that its subscribers may use "GPT" in the names of these, although it's "discouraged".<sup id="cite_ref-:11_80-1" class="reference"><a href="#cite_note-:11-80"><span class="cite-bracket">&#91;</span>80<span class="cite-bracket">&#93;</span></a></sup> </p><p>Relatedly, OpenAI has applied to the <a href="/wiki/United_States_Patent_and_Trademark_Office" title="United States Patent and Trademark Office">United States Patent and Trademark Office</a> (USPTO) to seek domestic <a href="/wiki/Trademark_registration" class="mw-redirect" title="Trademark registration">trademark registration</a> for the term "GPT" in the field of AI.<sup id="cite_ref-:5_78-2" class="reference"><a href="#cite_note-:5-78"><span class="cite-bracket">&#91;</span>78<span class="cite-bracket">&#93;</span></a></sup> OpenAI sought to expedite handling of its application, but the USPTO declined that request in April 2023.<sup id="cite_ref-:6_82-0" class="reference"><a href="#cite_note-:6-82"><span class="cite-bracket">&#91;</span>82<span class="cite-bracket">&#93;</span></a></sup> In May 2023, the USPTO responded to the application with a determination that "GPT" was both descriptive and generic.<sup id="cite_ref-83" class="reference"><a href="#cite_note-83"><span class="cite-bracket">&#91;</span>83<span class="cite-bracket">&#93;</span></a></sup> As of November 2023, OpenAI continues to pursue its argument through the available processes. Regardless, failure to obtain a <i>registered</i> U.S. trademark does not preclude some level of <i>common-law</i> trademark rights in the U.S.,<sup id="cite_ref-84" class="reference"><a href="#cite_note-84"><span class="cite-bracket">&#91;</span>84<span class="cite-bracket">&#93;</span></a></sup> and/or trademark rights in other countries.<sup id="cite_ref-85" class="reference"><a href="#cite_note-85"><span class="cite-bracket">&#91;</span>85<span class="cite-bracket">&#93;</span></a></sup> </p><p>For any given type or scope of trademark protection in the U.S., OpenAI would need to establish that the term is actually "<a href="/wiki/Trademark_distinctiveness" title="Trademark distinctiveness">distinctive</a>" to their specific offerings in addition to being a broader technical term for the kind of technology. Some media reports suggested that OpenAI may be able to obtain trademark registration based indirectly on the fame of its GPT-based <a href="/wiki/Chatbot" title="Chatbot">chatbot</a> product, <a href="/wiki/ChatGPT" title="ChatGPT">ChatGPT</a>,<sup id="cite_ref-:6_82-1" class="reference"><a href="#cite_note-:6-82"><span class="cite-bracket">&#91;</span>82<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-86" class="reference"><a href="#cite_note-86"><span class="cite-bracket">&#91;</span>86<span class="cite-bracket">&#93;</span></a></sup> for which OpenAI has <i>separately</i> sought protection (and which it has sought to enforce more strongly).<sup id="cite_ref-87" class="reference"><a href="#cite_note-87"><span class="cite-bracket">&#91;</span>87<span class="cite-bracket">&#93;</span></a></sup> Other reports have indicated that registration for the bare term "GPT" seems unlikely to be granted,<sup id="cite_ref-:5_78-3" class="reference"><a href="#cite_note-:5-78"><span class="cite-bracket">&#91;</span>78<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-:7_88-0" class="reference"><a href="#cite_note-:7-88"><span class="cite-bracket">&#91;</span>88<span class="cite-bracket">&#93;</span></a></sup> as it is used frequently as a common term to refer simply to AI systems that involve generative pre-trained transformers.<sup id="cite_ref-:4_3-2" class="reference"><a href="#cite_note-:4-3"><span class="cite-bracket">&#91;</span>3<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-89" class="reference"><a href="#cite_note-89"><span class="cite-bracket">&#91;</span>89<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-90" class="reference"><a href="#cite_note-90"><span class="cite-bracket">&#91;</span>90<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-91" class="reference"><a href="#cite_note-91"><span class="cite-bracket">&#91;</span>91<span class="cite-bracket">&#93;</span></a></sup> In any event, to whatever extent exclusive rights in the term may occur the U.S., others would need to avoid using it for similar products or services in ways likely to cause confusion.<sup id="cite_ref-:7_88-1" class="reference"><a href="#cite_note-:7-88"><span class="cite-bracket">&#91;</span>88<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-92" class="reference"><a href="#cite_note-92"><span class="cite-bracket">&#91;</span>92<span class="cite-bracket">&#93;</span></a></sup> If such rights ever became broad enough to implicate other well-established uses in the field, the trademark doctrine of <i>descriptive fair use</i> could still continue non-brand-related usage.<sup id="cite_ref-93" class="reference"><a href="#cite_note-93"><span class="cite-bracket">&#91;</span>93<span class="cite-bracket">&#93;</span></a></sup> </p> <div class="mw-heading mw-heading2"><h2 id="Selected_bibliography">Selected bibliography</h2><span class="mw-editsection"><span class="mw-editsection-bracket">[</span><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit&amp;section=9" title="Edit section: Selected bibliography"><span>edit</span></a><span class="mw-editsection-bracket">]</span></span></div> <p>This section lists the main official publications from OpenAI and Microsoft on their GPT models. </p> <ul><li>GPT-1: report,<sup id="cite_ref-gpt1_9-3" class="reference"><a href="#cite_note-gpt1-9"><span class="cite-bracket">&#91;</span>9<span class="cite-bracket">&#93;</span></a></sup> GitHub release.<sup id="cite_ref-94" class="reference"><a href="#cite_note-94"><span class="cite-bracket">&#91;</span>94<span class="cite-bracket">&#93;</span></a></sup></li> <li>GPT-2: blog announcement,<sup id="cite_ref-95" class="reference"><a href="#cite_note-95"><span class="cite-bracket">&#91;</span>95<span class="cite-bracket">&#93;</span></a></sup> report on its decision of "staged release",<sup id="cite_ref-96" class="reference"><a href="#cite_note-96"><span class="cite-bracket">&#91;</span>96<span class="cite-bracket">&#93;</span></a></sup> GitHub release.<sup id="cite_ref-97" class="reference"><a href="#cite_note-97"><span class="cite-bracket">&#91;</span>97<span class="cite-bracket">&#93;</span></a></sup></li> <li>GPT-3: report.<sup id="cite_ref-:2_41-3" class="reference"><a href="#cite_note-:2-41"><span class="cite-bracket">&#91;</span>41<span class="cite-bracket">&#93;</span></a></sup> No GitHub or any other form of code release thenceforth.</li> <li>WebGPT: blog announcement,<sup id="cite_ref-98" class="reference"><a href="#cite_note-98"><span class="cite-bracket">&#91;</span>98<span class="cite-bracket">&#93;</span></a></sup> report,<sup id="cite_ref-99" class="reference"><a href="#cite_note-99"><span class="cite-bracket">&#91;</span>99<span class="cite-bracket">&#93;</span></a></sup></li> <li>InstructGPT: blog announcement,<sup id="cite_ref-instructgpt-blog_54-1" class="reference"><a href="#cite_note-instructgpt-blog-54"><span class="cite-bracket">&#91;</span>54<span class="cite-bracket">&#93;</span></a></sup> report.<sup id="cite_ref-instructgpt-paper_55-1" class="reference"><a href="#cite_note-instructgpt-paper-55"><span class="cite-bracket">&#91;</span>55<span class="cite-bracket">&#93;</span></a></sup></li> <li>ChatGPT: blog announcement (no report).<sup id="cite_ref-chatgpt-blog_59-1" class="reference"><a href="#cite_note-chatgpt-blog-59"><span class="cite-bracket">&#91;</span>59<span class="cite-bracket">&#93;</span></a></sup></li> <li>GPT-4: blog announcement,<sup id="cite_ref-100" class="reference"><a href="#cite_note-100"><span class="cite-bracket">&#91;</span>100<span class="cite-bracket">&#93;</span></a></sup> reports,<sup id="cite_ref-101" class="reference"><a href="#cite_note-101"><span class="cite-bracket">&#91;</span>101<span class="cite-bracket">&#93;</span></a></sup><sup id="cite_ref-102" class="reference"><a href="#cite_note-102"><span class="cite-bracket">&#91;</span>102<span class="cite-bracket">&#93;</span></a></sup> model card.<sup id="cite_ref-103" class="reference"><a href="#cite_note-103"><span class="cite-bracket">&#91;</span>103<span class="cite-bracket">&#93;</span></a></sup></li> <li>GPT-4o: blog announcement.<sup id="cite_ref-104" class="reference"><a href="#cite_note-104"><span class="cite-bracket">&#91;</span>104<span class="cite-bracket">&#93;</span></a></sup></li></ul> <div class="mw-heading mw-heading2"><h2 id="See_also">See also</h2><span class="mw-editsection"><span class="mw-editsection-bracket">[</span><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit&amp;section=10" title="Edit section: See also"><span>edit</span></a><span class="mw-editsection-bracket">]</span></span></div> <ul><li><a href="/wiki/Cyc" title="Cyc">Cyc</a></li> <li><a href="/wiki/Gemini_(language_model)" title="Gemini (language model)">Gemini</a></li></ul> <div class="mw-heading mw-heading2"><h2 id="References">References</h2><span class="mw-editsection"><span class="mw-editsection-bracket">[</span><a href="/w/index.php?title=Generative_pre-trained_transformer&amp;action=edit&amp;section=11" title="Edit section: References"><span>edit</span></a><span class="mw-editsection-bracket">]</span></span></div> <style data-mw-deduplicate="TemplateStyles:r1239543626">.mw-parser-output .reflist{margin-bottom:0.5em;list-style-type:decimal}@media screen{.mw-parser-output .reflist{font-size:90%}}.mw-parser-output .reflist .references{font-size:100%;margin-bottom:0;list-style-type:inherit}.mw-parser-output .reflist-columns-2{column-width:30em}.mw-parser-output .reflist-columns-3{column-width:25em}.mw-parser-output .reflist-columns{margin-top:0.3em}.mw-parser-output .reflist-columns ol{margin-top:0}.mw-parser-output .reflist-columns li{page-break-inside:avoid;break-inside:avoid-column}.mw-parser-output .reflist-upper-alpha{list-style-type:upper-alpha}.mw-parser-output .reflist-upper-roman{list-style-type:upper-roman}.mw-parser-output .reflist-lower-alpha{list-style-type:lower-alpha}.mw-parser-output .reflist-lower-greek{list-style-type:lower-greek}.mw-parser-output .reflist-lower-roman{list-style-type:lower-roman}</style><div class="reflist"> <div class="mw-references-wrap mw-references-columns"><ol class="references"> <li id="cite_note-:1-1"><span class="mw-cite-backlink">^ <a href="#cite_ref-:1_1-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:1_1-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><style data-mw-deduplicate="TemplateStyles:r1238218222">.mw-parser-output cite.citation{font-style:inherit;word-wrap:break-word}.mw-parser-output .citation q{quotes:"\"""\"""'""'"}.mw-parser-output .citation:target{background-color:rgba(0,127,255,0.133)}.mw-parser-output .id-lock-free.id-lock-free a{background:url("//upload.wikimedia.org/wikipedia/commons/6/65/Lock-green.svg")right 0.1em center/9px no-repeat}.mw-parser-output .id-lock-limited.id-lock-limited a,.mw-parser-output .id-lock-registration.id-lock-registration a{background:url("//upload.wikimedia.org/wikipedia/commons/d/d6/Lock-gray-alt-2.svg")right 0.1em center/9px no-repeat}.mw-parser-output .id-lock-subscription.id-lock-subscription a{background:url("//upload.wikimedia.org/wikipedia/commons/a/aa/Lock-red-alt-2.svg")right 0.1em center/9px no-repeat}.mw-parser-output .cs1-ws-icon a{background:url("//upload.wikimedia.org/wikipedia/commons/4/4c/Wikisource-logo.svg")right 0.1em center/12px no-repeat}body:not(.skin-timeless):not(.skin-minerva) .mw-parser-output .id-lock-free a,body:not(.skin-timeless):not(.skin-minerva) .mw-parser-output .id-lock-limited a,body:not(.skin-timeless):not(.skin-minerva) .mw-parser-output .id-lock-registration a,body:not(.skin-timeless):not(.skin-minerva) .mw-parser-output .id-lock-subscription a,body:not(.skin-timeless):not(.skin-minerva) .mw-parser-output .cs1-ws-icon a{background-size:contain;padding:0 1em 0 0}.mw-parser-output .cs1-code{color:inherit;background:inherit;border:none;padding:inherit}.mw-parser-output .cs1-hidden-error{display:none;color:var(--color-error,#d33)}.mw-parser-output .cs1-visible-error{color:var(--color-error,#d33)}.mw-parser-output .cs1-maint{display:none;color:#085;margin-left:0.3em}.mw-parser-output .cs1-kern-left{padding-left:0.2em}.mw-parser-output .cs1-kern-right{padding-right:0.2em}.mw-parser-output .citation .mw-selflink{font-weight:inherit}@media screen{.mw-parser-output .cs1-format{font-size:95%}html.skin-theme-clientpref-night .mw-parser-output .cs1-maint{color:#18911f}}@media screen and (prefers-color-scheme:dark){html.skin-theme-clientpref-os .mw-parser-output .cs1-maint{color:#18911f}}</style><cite id="CITEREFHaddad" class="citation web cs1">Haddad, Mohammed. <a rel="nofollow" class="external text" href="https://www.aljazeera.com/news/2023/3/15/how-do-ai-models-like-gpt-4-work-and-how-can-you-start-using-it">"How does GPT-4 work and how can you start using it in ChatGPT?"</a>. <i>www.aljazeera.com</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=www.aljazeera.com&amp;rft.atitle=How+does+GPT-4+work+and+how+can+you+start+using+it+in+ChatGPT%3F&amp;rft.aulast=Haddad&amp;rft.aufirst=Mohammed&amp;rft_id=https%3A%2F%2Fwww.aljazeera.com%2Fnews%2F2023%2F3%2F15%2Fhow-do-ai-models-like-gpt-4-work-and-how-can-you-start-using-it&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:0-2"><span class="mw-cite-backlink">^ <a href="#cite_ref-:0_2-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:0_2-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.weforum.org/agenda/2023/01/davos23-generative-ai-a-game-changer-industries-and-society-code-developers/">"Generative AI: a game-changer society needs to be ready for"</a>. <i>World Economic Forum</i>. 9 January 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=World+Economic+Forum&amp;rft.atitle=Generative+AI%3A+a+game-changer+society+needs+to+be+ready+for&amp;rft.date=2023-01-09&amp;rft_id=https%3A%2F%2Fwww.weforum.org%2Fagenda%2F2023%2F01%2Fdavos23-generative-ai-a-game-changer-industries-and-society-code-developers%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:4-3"><span class="mw-cite-backlink">^ <a href="#cite_ref-:4_3-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:4_3-1"><sup><i><b>b</b></i></sup></a> <a href="#cite_ref-:4_3-2"><sup><i><b>c</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation magazine cs1"><a rel="nofollow" class="external text" href="https://time.com/6271657/a-to-z-of-artificial-intelligence/">"The A to Z of Artificial Intelligence"</a>. <i>Time</i>. April 13, 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Time&amp;rft.atitle=The+A+to+Z+of+Artificial+Intelligence&amp;rft.date=2023-04-13&amp;rft_id=https%3A%2F%2Ftime.com%2F6271657%2Fa-to-z-of-artificial-intelligence%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-4"><span class="mw-cite-backlink"><b><a href="#cite_ref-4">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFHu2022" class="citation web cs1">Hu, Luhui (November 15, 2022). <a rel="nofollow" class="external text" href="https://pub.towardsai.net/generative-ai-and-future-c3b1695876f2">"Generative AI and Future"</a>. <i>Medium</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Medium&amp;rft.atitle=Generative+AI+and+Future&amp;rft.date=2022-11-15&amp;rft.aulast=Hu&amp;rft.aufirst=Luhui&amp;rft_id=https%3A%2F%2Fpub.towardsai.net%2Fgenerative-ai-and-future-c3b1695876f2&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-5"><span class="mw-cite-backlink"><b><a href="#cite_ref-5">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.computer.org/csdl/magazine/co/2022/10/09903869/1H0G6xvtREk">"CSDL &#124; IEEE Computer Society"</a>. <i>www.computer.org</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=www.computer.org&amp;rft.atitle=CSDL+%26%23124%3B+IEEE+Computer+Society&amp;rft_id=https%3A%2F%2Fwww.computer.org%2Fcsdl%2Fmagazine%2Fco%2F2022%2F10%2F09903869%2F1H0G6xvtREk&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-6"><span class="mw-cite-backlink"><b><a href="#cite_ref-6">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://hallmark.libguides.com/c.php?g=1312147&amp;p=9644939">"LibGuides: Using AI Language Models&#160;: ChatGPT"</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=LibGuides%3A+Using+AI+Language+Models+%3A+ChatGPT&amp;rft_id=https%3A%2F%2Fhallmark.libguides.com%2Fc.php%3Fg%3D1312147%26p%3D9644939&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-7"><span class="mw-cite-backlink"><b><a href="#cite_ref-7">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFToews" class="citation web cs1">Toews, Rob. <a rel="nofollow" class="external text" href="https://www.forbes.com/sites/robtoews/2023/02/07/the-next-generation-of-large-language-models/">"The Next Generation Of Large Language Models"</a>. <i>Forbes</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Forbes&amp;rft.atitle=The+Next+Generation+Of+Large+Language+Models&amp;rft.aulast=Toews&amp;rft.aufirst=Rob&amp;rft_id=https%3A%2F%2Fwww.forbes.com%2Fsites%2Frobtoews%2F2023%2F02%2F07%2Fthe-next-generation-of-large-language-models%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-8"><span class="mw-cite-backlink"><b><a href="#cite_ref-8">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFMckendrick2023" class="citation web cs1">Mckendrick, Joe (March 13, 2023). <a rel="nofollow" class="external text" href="https://www.forbes.com/sites/joemckendrick/2023/03/26/most-jobs-soon-to-be-influenced-by-artificial-intelligence-research-out-of-openai-and-university-of-pennsylvania-suggests/?sh=420f9c8f73c7">"Most Jobs Soon To Be 'Influenced' By Artificial Intelligence, Research Out Of OpenAI And University Of Pennsylvania Suggests"</a>. <i><a href="/wiki/Forbes" title="Forbes">Forbes</a></i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Forbes&amp;rft.atitle=Most+Jobs+Soon+To+Be+%27Influenced%27+By+Artificial+Intelligence%2C+Research+Out+Of+OpenAI+And+University+Of+Pennsylvania+Suggests&amp;rft.date=2023-03-13&amp;rft.aulast=Mckendrick&amp;rft.aufirst=Joe&amp;rft_id=https%3A%2F%2Fwww.forbes.com%2Fsites%2Fjoemckendrick%2F2023%2F03%2F26%2Fmost-jobs-soon-to-be-influenced-by-artificial-intelligence-research-out-of-openai-and-university-of-pennsylvania-suggests%2F%3Fsh%3D420f9c8f73c7&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-gpt1-9"><span class="mw-cite-backlink">^ <a href="#cite_ref-gpt1_9-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-gpt1_9-1"><sup><i><b>b</b></i></sup></a> <a href="#cite_ref-gpt1_9-2"><sup><i><b>c</b></i></sup></a> <a href="#cite_ref-gpt1_9-3"><sup><i><b>d</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://openai.com/research/language-unsupervised">"Improving language understanding with unsupervised learning"</a>. <i>openai.com</i>. June 11, 2018. <a rel="nofollow" class="external text" href="https://web.archive.org/web/20230318210736/https://openai.com/research/language-unsupervised">Archived</a> from the original on 2023-03-18<span class="reference-accessdate">. Retrieved <span class="nowrap">2023-03-18</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=openai.com&amp;rft.atitle=Improving+language+understanding+with+unsupervised+learning&amp;rft.date=2018-06-11&amp;rft_id=https%3A%2F%2Fopenai.com%2Fresearch%2Flanguage-unsupervised&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-10"><span class="mw-cite-backlink"><b><a href="#cite_ref-10">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.makeuseof.com/gpt-models-explained-and-compared/">"GPT-1 to GPT-4: Each of OpenAI's GPT Models Explained and Compared"</a>. <i>MUO</i>. April 11, 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=MUO&amp;rft.atitle=GPT-1+to+GPT-4%3A+Each+of+OpenAI%27s+GPT+Models+Explained+and+Compared&amp;rft.date=2023-04-11&amp;rft_id=https%3A%2F%2Fwww.makeuseof.com%2Fgpt-models-explained-and-compared%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-11"><span class="mw-cite-backlink"><b><a href="#cite_ref-11">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://openai.com/research/gpt-4">"GPT-4"</a>. <i>openai.com</i><span class="reference-accessdate">. Retrieved <span class="nowrap">2023-12-08</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=openai.com&amp;rft.atitle=GPT-4&amp;rft_id=https%3A%2F%2Fopenai.com%2Fresearch%2Fgpt-4&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:9-12"><span class="mw-cite-backlink">^ <a href="#cite_ref-:9_12-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:9_12-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFAlford2021" class="citation web cs1">Alford, Anthony (July 13, 2021). <a rel="nofollow" class="external text" href="https://www.infoq.com/news/2021/07/eleutherai-gpt-j/">"EleutherAI Open-Sources Six Billion Parameter GPT-3 Clone GPT-J"</a>. <i>InfoQ</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=InfoQ&amp;rft.atitle=EleutherAI+Open-Sources+Six+Billion+Parameter+GPT-3+Clone+GPT-J&amp;rft.date=2021-07-13&amp;rft.aulast=Alford&amp;rft.aufirst=Anthony&amp;rft_id=https%3A%2F%2Fwww.infoq.com%2Fnews%2F2021%2F07%2Feleutherai-gpt-j%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:10-13"><span class="mw-cite-backlink">^ <a href="#cite_ref-:10_13-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:10_13-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation pressrelease cs1"><a rel="nofollow" class="external text" href="https://www.businesswire.com/news/home/20230328005366/en/Cerebras-Systems-Releases-Seven-New-GPT-Models-Trained-on-CS-2-Wafer-Scale-Systems">"News"</a> (Press release).</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=News&amp;rft_id=https%3A%2F%2Fwww.businesswire.com%2Fnews%2Fhome%2F20230328005366%2Fen%2FCerebras-Systems-Releases-Seven-New-GPT-Models-Trained-on-CS-2-Wafer-Scale-Systems&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-14"><span class="mw-cite-backlink"><b><a href="#cite_ref-14">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFMorrison2023" class="citation web cs1">Morrison, Ryan (7 March 2023). <a rel="nofollow" class="external text" href="https://techmonitor.ai/technology/ai-and-automation/salesforce-einsteingpt-openai-chatgpt">"Salesforce launches EinsteinGPT built with OpenAI technology"</a>. <i>Tech Monitor</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Tech+Monitor&amp;rft.atitle=Salesforce+launches+EinsteinGPT+built+with+OpenAI+technology&amp;rft.date=2023-03-07&amp;rft.aulast=Morrison&amp;rft.aufirst=Ryan&amp;rft_id=https%3A%2F%2Ftechmonitor.ai%2Ftechnology%2Fai-and-automation%2Fsalesforce-einsteingpt-openai-chatgpt&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-15"><span class="mw-cite-backlink"><b><a href="#cite_ref-15">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.forbes.com/sites/jamielsheikh/2023/04/05/the-chatgpt-of-finance-is-here-bloomberg-is-combining-ai-and-fintech/?sh=43b4385e3081">"The ChatGPT of Finance is Here, Bloomberg is Combining AI and Fintech"</a>. <i><a href="/wiki/Forbes" title="Forbes">Forbes</a></i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Forbes&amp;rft.atitle=The+ChatGPT+of+Finance+is+Here%2C+Bloomberg+is+Combining+AI+and+Fintech&amp;rft_id=https%3A%2F%2Fwww.forbes.com%2Fsites%2Fjamielsheikh%2F2023%2F04%2F05%2Fthe-chatgpt-of-finance-is-here-bloomberg-is-combining-ai-and-fintech%2F%3Fsh%3D43b4385e3081&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-16"><span class="mw-cite-backlink"><b><a href="#cite_ref-16">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFHinton_(et-al)2012" class="citation journal cs1 cs1-prop-long-vol">Hinton (et-al), Geoffrey (October 15, 2012). <a rel="nofollow" class="external text" href="http://cs224d.stanford.edu/papers/maas_paper.pdf">"Deep neural networks for acoustic modeling in speech recognition"</a> <span class="cs1-format">(PDF)</span>. <i>IEEE Signal Processing Magazine</i>. Digital Object Identifier 10.1109/MSP.2012.2205597. <a href="/wiki/Doi_(identifier)" class="mw-redirect" title="Doi (identifier)">doi</a>:<a rel="nofollow" class="external text" href="https://doi.org/10.1109%2FMSP.2012.2205597">10.1109/MSP.2012.2205597</a>. <a href="/wiki/S2CID_(identifier)" class="mw-redirect" title="S2CID (identifier)">S2CID</a>&#160;<a rel="nofollow" class="external text" href="https://api.semanticscholar.org/CorpusID:206485943">206485943</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=IEEE+Signal+Processing+Magazine&amp;rft.atitle=Deep+neural+networks+for+acoustic+modeling+in+speech+recognition&amp;rft.volume=Digital+Object+Identifier+10.1109%2FMSP.2012.2205597&amp;rft.date=2012-10-15&amp;rft_id=info%3Adoi%2F10.1109%2FMSP.2012.2205597&amp;rft_id=https%3A%2F%2Fapi.semanticscholar.org%2FCorpusID%3A206485943%23id-name%3DS2CID&amp;rft.aulast=Hinton+%28et-al%29&amp;rft.aufirst=Geoffrey&amp;rft_id=http%3A%2F%2Fcs224d.stanford.edu%2Fpapers%2Fmaas_paper.pdf&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-17"><span class="mw-cite-backlink"><b><a href="#cite_ref-17">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFDeng2014" class="citation journal cs1">Deng, Li (2014-01-22). <a rel="nofollow" class="external text" href="https://doi.org/10.1017%2Fatsip.2013.9">"A tutorial survey of architectures, algorithms, and applications for deep learning &#124; APSIPA Transactions on Signal and Information Processing &#124; Cambridge Core"</a>. <i>Apsipa Transactions on Signal and Information Processing</i>. <b>3</b>. Cambridge.org: e2. <a href="/wiki/Doi_(identifier)" class="mw-redirect" title="Doi (identifier)">doi</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://doi.org/10.1017%2Fatsip.2013.9">10.1017/atsip.2013.9</a></span>. <a href="/wiki/S2CID_(identifier)" class="mw-redirect" title="S2CID (identifier)">S2CID</a>&#160;<a rel="nofollow" class="external text" href="https://api.semanticscholar.org/CorpusID:9928823">9928823</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Apsipa+Transactions+on+Signal+and+Information+Processing&amp;rft.atitle=A+tutorial+survey+of+architectures%2C+algorithms%2C+and+applications+for+deep+learning+%26%23124%3B+APSIPA+Transactions+on+Signal+and+Information+Processing+%26%23124%3B+Cambridge+Core&amp;rft.volume=3&amp;rft.pages=e2&amp;rft.date=2014-01-22&amp;rft_id=info%3Adoi%2F10.1017%2Fatsip.2013.9&amp;rft_id=https%3A%2F%2Fapi.semanticscholar.org%2FCorpusID%3A9928823%23id-name%3DS2CID&amp;rft.aulast=Deng&amp;rft.aufirst=Li&amp;rft_id=https%3A%2F%2Fdoi.org%2F10.1017%252Fatsip.2013.9&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-18"><span class="mw-cite-backlink"><b><a href="#cite_ref-18">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFErhanCourvilleBengioVincent2010" class="citation journal cs1">Erhan, Dumitru; Courville, Aaron; Bengio, Yoshua; Vincent, Pascal (2010-03-31). <a rel="nofollow" class="external text" href="https://proceedings.mlr.press/v9/erhan10a.html">"Why Does Unsupervised Pre-training Help Deep Learning?"</a>. <i>Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics</i>. JMLR Workshop and Conference Proceedings: <span class="nowrap">201–</span>208.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Proceedings+of+the+Thirteenth+International+Conference+on+Artificial+Intelligence+and+Statistics&amp;rft.atitle=Why+Does+Unsupervised+Pre-training+Help+Deep+Learning%3F&amp;rft.pages=%3Cspan+class%3D%22nowrap%22%3E201-%3C%2Fspan%3E208&amp;rft.date=2010-03-31&amp;rft.aulast=Erhan&amp;rft.aufirst=Dumitru&amp;rft.au=Courville%2C+Aaron&amp;rft.au=Bengio%2C+Yoshua&amp;rft.au=Vincent%2C+Pascal&amp;rft_id=https%3A%2F%2Fproceedings.mlr.press%2Fv9%2Ferhan10a.html&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-19"><span class="mw-cite-backlink"><b><a href="#cite_ref-19">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="http://ethw.org/First-Hand:The_Hidden_Markov_Model">"First-Hand:The Hidden Markov Model – Engineering and Technology History Wiki"</a>. <i>ethw.org</i>. 12 January 2015. <a rel="nofollow" class="external text" href="https://web.archive.org/web/20180403191314/http://ethw.org/First-Hand:The_Hidden_Markov_Model">Archived</a> from the original on 3 April 2018<span class="reference-accessdate">. Retrieved <span class="nowrap">1 May</span> 2018</span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=ethw.org&amp;rft.atitle=First-Hand%3AThe+Hidden+Markov+Model+%E2%80%93+Engineering+and+Technology+History+Wiki&amp;rft.date=2015-01-12&amp;rft_id=http%3A%2F%2Fethw.org%2FFirst-Hand%3AThe_Hidden_Markov_Model&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-20"><span class="mw-cite-backlink"><b><a href="#cite_ref-20">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFJuangRabiner1991" class="citation journal cs1">Juang, B. H.; Rabiner, L. R. (1991). <a rel="nofollow" class="external text" href="https://www.jstor.org/stable/1268779">"Hidden Markov Models for Speech Recognition"</a>. <i>Technometrics</i>. <b>33</b> (3): <span class="nowrap">251–</span>272. <a href="/wiki/Doi_(identifier)" class="mw-redirect" title="Doi (identifier)">doi</a>:<a rel="nofollow" class="external text" href="https://doi.org/10.2307%2F1268779">10.2307/1268779</a>. <a href="/wiki/ISSN_(identifier)" class="mw-redirect" title="ISSN (identifier)">ISSN</a>&#160;<a rel="nofollow" class="external text" href="https://search.worldcat.org/issn/0040-1706">0040-1706</a>. <a href="/wiki/JSTOR_(identifier)" class="mw-redirect" title="JSTOR (identifier)">JSTOR</a>&#160;<a rel="nofollow" class="external text" href="https://www.jstor.org/stable/1268779">1268779</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Technometrics&amp;rft.atitle=Hidden+Markov+Models+for+Speech+Recognition&amp;rft.volume=33&amp;rft.issue=3&amp;rft.pages=%3Cspan+class%3D%22nowrap%22%3E251-%3C%2Fspan%3E272&amp;rft.date=1991&amp;rft.issn=0040-1706&amp;rft_id=https%3A%2F%2Fwww.jstor.org%2Fstable%2F1268779%23id-name%3DJSTOR&amp;rft_id=info%3Adoi%2F10.2307%2F1268779&amp;rft.aulast=Juang&amp;rft.aufirst=B.+H.&amp;rft.au=Rabiner%2C+L.+R.&amp;rft_id=https%3A%2F%2Fwww.jstor.org%2Fstable%2F1268779&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-21"><span class="mw-cite-backlink"><b><a href="#cite_ref-21">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFCottrellMunroZipser1987" class="citation journal cs1">Cottrell, Garrison W.; Munro, Paul; Zipser, David (1987). <a rel="nofollow" class="external text" href="https://escholarship.org/uc/item/2zs7w6z8">"Learning Internal Representation From Gray-Scale Images: An Example of Extensional Programming"</a>. <i>Proceedings of the Annual Meeting of the Cognitive Science Society</i>. <b>9</b>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Proceedings+of+the+Annual+Meeting+of+the+Cognitive+Science+Society&amp;rft.atitle=Learning+Internal+Representation+From+Gray-Scale+Images%3A+An+Example+of+Extensional+Programming&amp;rft.volume=9&amp;rft.date=1987&amp;rft.aulast=Cottrell&amp;rft.aufirst=Garrison+W.&amp;rft.au=Munro%2C+Paul&amp;rft.au=Zipser%2C+David&amp;rft_id=https%3A%2F%2Fescholarship.org%2Fuc%2Fitem%2F2zs7w6z8&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-22"><span class="mw-cite-backlink"><b><a href="#cite_ref-22">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFCottrell1991" class="citation cs2">Cottrell, Garrison W. (1991-01-01), Touretzky, David S.; Elman, Jeffrey L.; Sejnowski, Terrence J.; Hinton, Geoffrey E. (eds.), <a rel="nofollow" class="external text" href="https://www.sciencedirect.com/science/article/abs/pii/B9781483214481500391">"Extracting features from faces using compression networks: Face, identity, emotion, and gender recognition using holons"</a>, <i>Connectionist Models</i>, Morgan Kaufmann, pp.&#160;<span class="nowrap">328–</span>337, <a href="/wiki/ISBN_(identifier)" class="mw-redirect" title="ISBN (identifier)">ISBN</a>&#160;<a href="/wiki/Special:BookSources/978-1-4832-1448-1" title="Special:BookSources/978-1-4832-1448-1"><bdi>978-1-4832-1448-1</bdi></a><span class="reference-accessdate">, retrieved <span class="nowrap">2024-10-04</span></span></cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Connectionist+Models&amp;rft.atitle=Extracting+features+from+faces+using+compression+networks%3A+Face%2C+identity%2C+emotion%2C+and+gender+recognition+using+holons&amp;rft.pages=%3Cspan+class%3D%22nowrap%22%3E328-%3C%2Fspan%3E337&amp;rft.date=1991-01-01&amp;rft.isbn=978-1-4832-1448-1&amp;rft.aulast=Cottrell&amp;rft.aufirst=Garrison+W.&amp;rft_id=https%3A%2F%2Fwww.sciencedirect.com%2Fscience%2Farticle%2Fabs%2Fpii%2FB9781483214481500391&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-schmidhuber1992-23"><span class="mw-cite-backlink"><b><a href="#cite_ref-schmidhuber1992_23-0">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFSchmidhuber1992" class="citation journal cs1">Schmidhuber, Jürgen (1992). <a rel="nofollow" class="external text" href="https://gwern.net/doc/ai/nn/rnn/1992-schmidhuber.pdf">"Learning complex, extended sequences using the principle of history compression"</a> <span class="cs1-format">(PDF)</span>. <i>Neural Computation</i>. <b>4</b> (2): <span class="nowrap">234–</span>242. <a href="/wiki/Doi_(identifier)" class="mw-redirect" title="Doi (identifier)">doi</a>:<a rel="nofollow" class="external text" href="https://doi.org/10.1162%2Fneco.1992.4.2.234">10.1162/neco.1992.4.2.234</a>. <a href="/wiki/S2CID_(identifier)" class="mw-redirect" title="S2CID (identifier)">S2CID</a>&#160;<a rel="nofollow" class="external text" href="https://api.semanticscholar.org/CorpusID:18271205">18271205</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Neural+Computation&amp;rft.atitle=Learning+complex%2C+extended+sequences+using+the+principle+of+history+compression&amp;rft.volume=4&amp;rft.issue=2&amp;rft.pages=%3Cspan+class%3D%22nowrap%22%3E234-%3C%2Fspan%3E242&amp;rft.date=1992&amp;rft_id=info%3Adoi%2F10.1162%2Fneco.1992.4.2.234&amp;rft_id=https%3A%2F%2Fapi.semanticscholar.org%2FCorpusID%3A18271205%23id-name%3DS2CID&amp;rft.aulast=Schmidhuber&amp;rft.aufirst=J%C3%BCrgen&amp;rft_id=https%3A%2F%2Fgwern.net%2Fdoc%2Fai%2Fnn%2Frnn%2F1992-schmidhuber.pdf&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-24"><span class="mw-cite-backlink"><b><a href="#cite_ref-24">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFElmanZipser1988" class="citation journal cs1">Elman, Jeffrey L.; Zipser, David (1988-04-01). <a rel="nofollow" class="external text" href="https://pubs.aip.org/jasa/article/83/4/1615/826094/Learning-the-hidden-structure-of-speechLearning">"Learning the hidden structure of speech"</a>. <i>The Journal of the Acoustical Society of America</i>. <b>83</b> (4): <span class="nowrap">1615–</span>1626. <a href="/wiki/Bibcode_(identifier)" class="mw-redirect" title="Bibcode (identifier)">Bibcode</a>:<a rel="nofollow" class="external text" href="https://ui.adsabs.harvard.edu/abs/1988ASAJ...83.1615E">1988ASAJ...83.1615E</a>. <a href="/wiki/Doi_(identifier)" class="mw-redirect" title="Doi (identifier)">doi</a>:<a rel="nofollow" class="external text" href="https://doi.org/10.1121%2F1.395916">10.1121/1.395916</a>. <a href="/wiki/ISSN_(identifier)" class="mw-redirect" title="ISSN (identifier)">ISSN</a>&#160;<a rel="nofollow" class="external text" href="https://search.worldcat.org/issn/0001-4966">0001-4966</a>. <a href="/wiki/PMID_(identifier)" class="mw-redirect" title="PMID (identifier)">PMID</a>&#160;<a rel="nofollow" class="external text" href="https://pubmed.ncbi.nlm.nih.gov/3372872">3372872</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=The+Journal+of+the+Acoustical+Society+of+America&amp;rft.atitle=Learning+the+hidden+structure+of+speech&amp;rft.volume=83&amp;rft.issue=4&amp;rft.pages=%3Cspan+class%3D%22nowrap%22%3E1615-%3C%2Fspan%3E1626&amp;rft.date=1988-04-01&amp;rft_id=info%3Adoi%2F10.1121%2F1.395916&amp;rft.issn=0001-4966&amp;rft_id=info%3Apmid%2F3372872&amp;rft_id=info%3Abibcode%2F1988ASAJ...83.1615E&amp;rft.aulast=Elman&amp;rft.aufirst=Jeffrey+L.&amp;rft.au=Zipser%2C+David&amp;rft_id=https%3A%2F%2Fpubs.aip.org%2Fjasa%2Farticle%2F83%2F4%2F1615%2F826094%2FLearning-the-hidden-structure-of-speechLearning&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:14-25"><span class="mw-cite-backlink"><b><a href="#cite_ref-:14_25-0">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFBourlardKamp1988" class="citation journal cs1">Bourlard, H.; Kamp, Y. (1988). <a rel="nofollow" class="external text" href="http://infoscience.epfl.ch/record/82601">"Auto-association by multilayer perceptrons and singular value decomposition"</a>. <i>Biological Cybernetics</i>. <b>59</b> (<span class="nowrap">4–</span>5): <span class="nowrap">291–</span>294. <a href="/wiki/Doi_(identifier)" class="mw-redirect" title="Doi (identifier)">doi</a>:<a rel="nofollow" class="external text" href="https://doi.org/10.1007%2FBF00332918">10.1007/BF00332918</a>. <a href="/wiki/PMID_(identifier)" class="mw-redirect" title="PMID (identifier)">PMID</a>&#160;<a rel="nofollow" class="external text" href="https://pubmed.ncbi.nlm.nih.gov/3196773">3196773</a>. <a href="/wiki/S2CID_(identifier)" class="mw-redirect" title="S2CID (identifier)">S2CID</a>&#160;<a rel="nofollow" class="external text" href="https://api.semanticscholar.org/CorpusID:206775335">206775335</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Biological+Cybernetics&amp;rft.atitle=Auto-association+by+multilayer+perceptrons+and+singular+value+decomposition&amp;rft.volume=59&amp;rft.issue=%3Cspan+class%3D%22nowrap%22%3E4%E2%80%93%3C%2Fspan%3E5&amp;rft.pages=%3Cspan+class%3D%22nowrap%22%3E291-%3C%2Fspan%3E294&amp;rft.date=1988&amp;rft_id=https%3A%2F%2Fapi.semanticscholar.org%2FCorpusID%3A206775335%23id-name%3DS2CID&amp;rft_id=info%3Apmid%2F3196773&amp;rft_id=info%3Adoi%2F10.1007%2FBF00332918&amp;rft.aulast=Bourlard&amp;rft.aufirst=H.&amp;rft.au=Kamp%2C+Y.&amp;rft_id=http%3A%2F%2Finfoscience.epfl.ch%2Frecord%2F82601&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-26"><span class="mw-cite-backlink"><b><a href="#cite_ref-26">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFHintonZemel1993" class="citation journal cs1">Hinton, Geoffrey E; Zemel, Richard (1993). <a rel="nofollow" class="external text" href="https://proceedings.neurips.cc/paper/1993/hash/9e3cfc48eccf81a0d57663e129aef3cb-Abstract.html">"Autoencoders, Minimum Description Length and Helmholtz Free Energy"</a>. <i>Advances in Neural Information Processing Systems</i>. <b>6</b>. Morgan-Kaufmann.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Advances+in+Neural+Information+Processing+Systems&amp;rft.atitle=Autoencoders%2C+Minimum+Description+Length+and+Helmholtz+Free+Energy&amp;rft.volume=6&amp;rft.date=1993&amp;rft.aulast=Hinton&amp;rft.aufirst=Geoffrey+E&amp;rft.au=Zemel%2C+Richard&amp;rft_id=https%3A%2F%2Fproceedings.neurips.cc%2Fpaper%2F1993%2Fhash%2F9e3cfc48eccf81a0d57663e129aef3cb-Abstract.html&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-27"><span class="mw-cite-backlink"><b><a href="#cite_ref-27">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFVaswaniShazeerParmarUszkoreit2017" class="citation journal cs1"><a href="/wiki/Ashish_Vaswani" title="Ashish Vaswani">Vaswani, Ashish</a>; Shazeer, Noam; Parmar, Niki; Uszkoreit, Jakob; Jones, Llion; <a href="/wiki/Aidan_Gomez" title="Aidan Gomez">Gomez, Aidan N</a>; Kaiser, Łukasz; Polosukhin, Illia (2017). <a rel="nofollow" class="external text" href="https://proceedings.neurips.cc/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf">"Attention is All you Need"</a> <span class="cs1-format">(PDF)</span>. <i>Advances in Neural Information Processing Systems</i>. <b>30</b>. Curran Associates, Inc.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Advances+in+Neural+Information+Processing+Systems&amp;rft.atitle=Attention+is+All+you+Need&amp;rft.volume=30&amp;rft.date=2017&amp;rft.aulast=Vaswani&amp;rft.aufirst=Ashish&amp;rft.au=Shazeer%2C+Noam&amp;rft.au=Parmar%2C+Niki&amp;rft.au=Uszkoreit%2C+Jakob&amp;rft.au=Jones%2C+Llion&amp;rft.au=Gomez%2C+Aidan+N&amp;rft.au=Kaiser%2C+%C5%81ukasz&amp;rft.au=Polosukhin%2C+Illia&amp;rft_id=https%3A%2F%2Fproceedings.neurips.cc%2Fpaper%2F2017%2Ffile%2F3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-28"><span class="mw-cite-backlink"><b><a href="#cite_ref-28">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFDevlinChangLeeToutanova2019" class="citation journal cs1">Devlin, Jacob; Chang, Ming-Wei; Lee, Kenton; Toutanova, Kristina (May 24, 2019). "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding". <i>Association for Computational Linguistics</i>. <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/1810.04805">1810.04805</a></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Association+for+Computational+Linguistics&amp;rft.atitle=BERT%3A+Pre-training+of+Deep+Bidirectional+Transformers+for+Language+Understanding&amp;rft.date=2019-05-24&amp;rft_id=info%3Aarxiv%2F1810.04805&amp;rft.aulast=Devlin&amp;rft.aufirst=Jacob&amp;rft.au=Chang%2C+Ming-Wei&amp;rft.au=Lee%2C+Kenton&amp;rft.au=Toutanova%2C+Kristina&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-gpt1paper-29"><span class="mw-cite-backlink">^ <a href="#cite_ref-gpt1paper_29-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-gpt1paper_29-1"><sup><i><b>b</b></i></sup></a> <a href="#cite_ref-gpt1paper_29-2"><sup><i><b>c</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFRadfordNarasimhanSalimansSutskever2018" class="citation web cs1">Radford, Alec; Narasimhan, Karthik; Salimans, Tim; Sutskever, Ilya (11 June 2018). <a rel="nofollow" class="external text" href="https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf">"Improving Language Understanding by Generative Pre-Training"</a> <span class="cs1-format">(PDF)</span>. <a href="/wiki/OpenAI" title="OpenAI">OpenAI</a>. p.&#160;12. <a rel="nofollow" class="external text" href="https://web.archive.org/web/20210126024542/https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf">Archived</a> <span class="cs1-format">(PDF)</span> from the original on 26 January 2021<span class="reference-accessdate">. Retrieved <span class="nowrap">23 January</span> 2021</span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=Improving+Language+Understanding+by+Generative+Pre-Training&amp;rft.pages=12&amp;rft.pub=OpenAI&amp;rft.date=2018-06-11&amp;rft.aulast=Radford&amp;rft.aufirst=Alec&amp;rft.au=Narasimhan%2C+Karthik&amp;rft.au=Salimans%2C+Tim&amp;rft.au=Sutskever%2C+Ilya&amp;rft_id=https%3A%2F%2Fcdn.openai.com%2Fresearch-covers%2Flanguage-unsupervised%2Flanguage_understanding_paper.pdf&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-30"><span class="mw-cite-backlink"><b><a href="#cite_ref-30">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFRadfordJozefowiczSutskever2017" class="citation arxiv cs1">Radford, Alec; Jozefowicz, Rafal; Sutskever, Ilya (2017-04-06). "Learning to Generate Reviews and Discovering Sentiment". <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/1704.01444">1704.01444</a></span> [<a rel="nofollow" class="external text" href="https://arxiv.org/archive/cs.LG">cs.LG</a>].</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=preprint&amp;rft.jtitle=arXiv&amp;rft.atitle=Learning+to+Generate+Reviews+and+Discovering+Sentiment&amp;rft.date=2017-04-06&amp;rft_id=info%3Aarxiv%2F1704.01444&amp;rft.aulast=Radford&amp;rft.aufirst=Alec&amp;rft.au=Jozefowicz%2C+Rafal&amp;rft.au=Sutskever%2C+Ilya&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-31"><span class="mw-cite-backlink"><b><a href="#cite_ref-31">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFChenTworekJunYuan2021" class="citation journal cs1">Chen, Mark; Tworek, Jerry; Jun, Heewoo; Yuan, Qiming; Ponde de Oliveira Pinto, Henrique; Kaplan, Jared; Edwards, Harri; Burda, Yuri; Joseph, Nicholas; Brockman, Greg; Ray, Alex; Puri, Raul; Krueger, Gretchen; Petrov, Michael; Khlaaf, Heidy (2021-07-01). "Evaluating Large Language Models Trained on Code". <i>Association for Computational Linguistics</i>. <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/2107.03374">2107.03374</a></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Association+for+Computational+Linguistics&amp;rft.atitle=Evaluating+Large+Language+Models+Trained+on+Code&amp;rft.date=2021-07-01&amp;rft_id=info%3Aarxiv%2F2107.03374&amp;rft.aulast=Chen&amp;rft.aufirst=Mark&amp;rft.au=Tworek%2C+Jerry&amp;rft.au=Jun%2C+Heewoo&amp;rft.au=Yuan%2C+Qiming&amp;rft.au=Ponde+de+Oliveira+Pinto%2C+Henrique&amp;rft.au=Kaplan%2C+Jared&amp;rft.au=Edwards%2C+Harri&amp;rft.au=Burda%2C+Yuri&amp;rft.au=Joseph%2C+Nicholas&amp;rft.au=Brockman%2C+Greg&amp;rft.au=Ray%2C+Alex&amp;rft.au=Puri%2C+Raul&amp;rft.au=Krueger%2C+Gretchen&amp;rft.au=Petrov%2C+Michael&amp;rft.au=Khlaaf%2C+Heidy&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-32"><span class="mw-cite-backlink"><b><a href="#cite_ref-32">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFOuyangWuJiangAlmeida2022" class="citation journal cs1">Ouyang, Long; Wu, Jeffrey; Jiang, Xu; Almeida, Diogo; Wainwright, Carroll; Mishkin, Pamela; Zhang, Chong; Agarwal, Sandhini; Slama, Katarina; Ray, Alex; Schulman, John; Hilton, Jacob; Kelton, Fraser; Miller, Luke; Simens, Maddie (2022-12-06). <a rel="nofollow" class="external text" href="https://proceedings.neurips.cc/paper_files/paper/2022/hash/b1efde53be364a73914f58805a001731-Abstract-Conference.html">"Training language models to follow instructions with human feedback"</a>. <i>Advances in Neural Information Processing Systems</i>. <b>35</b>: <span class="nowrap">27730–</span>27744. <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/2203.02155">2203.02155</a></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Advances+in+Neural+Information+Processing+Systems&amp;rft.atitle=Training+language+models+to+follow+instructions+with+human+feedback&amp;rft.volume=35&amp;rft.pages=%3Cspan+class%3D%22nowrap%22%3E27730-%3C%2Fspan%3E27744&amp;rft.date=2022-12-06&amp;rft_id=info%3Aarxiv%2F2203.02155&amp;rft.aulast=Ouyang&amp;rft.aufirst=Long&amp;rft.au=Wu%2C+Jeffrey&amp;rft.au=Jiang%2C+Xu&amp;rft.au=Almeida%2C+Diogo&amp;rft.au=Wainwright%2C+Carroll&amp;rft.au=Mishkin%2C+Pamela&amp;rft.au=Zhang%2C+Chong&amp;rft.au=Agarwal%2C+Sandhini&amp;rft.au=Slama%2C+Katarina&amp;rft.au=Ray%2C+Alex&amp;rft.au=Schulman%2C+John&amp;rft.au=Hilton%2C+Jacob&amp;rft.au=Kelton%2C+Fraser&amp;rft.au=Miller%2C+Luke&amp;rft.au=Simens%2C+Maddie&amp;rft_id=https%3A%2F%2Fproceedings.neurips.cc%2Fpaper_files%2Fpaper%2F2022%2Fhash%2Fb1efde53be364a73914f58805a001731-Abstract-Conference.html&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-33"><span class="mw-cite-backlink"><b><a href="#cite_ref-33">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://openai.com/blog/gpt-3-edit-insert">"New GPT-3 capabilities: Edit &amp; insert"</a>. <i>openai.com</i><span class="reference-accessdate">. Retrieved <span class="nowrap">2023-06-24</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=openai.com&amp;rft.atitle=New+GPT-3+capabilities%3A+Edit+%26+insert&amp;rft_id=https%3A%2F%2Fopenai.com%2Fblog%2Fgpt-3-edit-insert&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-fu2022-34"><span class="mw-cite-backlink"><b><a href="#cite_ref-fu2022_34-0">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFFuPengKhot2022" class="citation journal cs1">Fu, Yao; Peng, Hao; Khot, Tushar (2022). <a rel="nofollow" class="external text" href="https://yaofu.notion.site/How-does-GPT-Obtain-its-Ability-Tracing-Emergent-Abilities-of-Language-Models-to-their-Sources-b9a57ac0fcf74f30a1ab9e3e36fa1dc1">"How does GPT Obtain its Ability? Tracing Emergent Abilities of Language Models to their Sources"</a>. <i>Yao Fu's Notion</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Yao+Fu%27s+Notion&amp;rft.atitle=How+does+GPT+Obtain+its+Ability%3F+Tracing+Emergent+Abilities+of+Language+Models+to+their+Sources&amp;rft.date=2022&amp;rft.aulast=Fu&amp;rft.aufirst=Yao&amp;rft.au=Peng%2C+Hao&amp;rft.au=Khot%2C+Tushar&amp;rft_id=https%3A%2F%2Fyaofu.notion.site%2FHow-does-GPT-Obtain-its-Ability-Tracing-Emergent-Abilities-of-Language-Models-to-their-Sources-b9a57ac0fcf74f30a1ab9e3e36fa1dc1&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-35"><span class="mw-cite-backlink"><b><a href="#cite_ref-35">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://platform.openai.com/docs/model-index-for-researchers">"Model index for researchers"</a>. <i>OpenAI API</i>. <a rel="nofollow" class="external text" href="https://archive.today/20230623231655/https://platform.openai.com/docs/model-index-for-researchers">Archived</a> from the original on 23 June 2023<span class="reference-accessdate">. Retrieved <span class="nowrap">2023-06-23</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=OpenAI+API&amp;rft.atitle=Model+index+for+researchers&amp;rft_id=https%3A%2F%2Fplatform.openai.com%2Fdocs%2Fmodel-index-for-researchers&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-36"><span class="mw-cite-backlink"><b><a href="#cite_ref-36">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://hai.stanford.edu/news/introducing-center-research-foundation-models-crfm">"Introducing the Center for Research on Foundation Models (CRFM)"</a>. <i>Stanford HAI</i>. 18 August 2021.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Stanford+HAI&amp;rft.atitle=Introducing+the+Center+for+Research+on+Foundation+Models+%28CRFM%29&amp;rft.date=2021-08-18&amp;rft_id=https%3A%2F%2Fhai.stanford.edu%2Fnews%2Fintroducing-center-research-foundation-models-crfm&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-37"><span class="mw-cite-backlink"><b><a href="#cite_ref-37">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://hai.stanford.edu/news/reflections-foundation-models">"Reflections on Foundation Models"</a>. <i>hai.stanford.edu</i>. 2021-10-18<span class="reference-accessdate">. Retrieved <span class="nowrap">2024-08-15</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=hai.stanford.edu&amp;rft.atitle=Reflections+on+Foundation+Models&amp;rft.date=2021-10-18&amp;rft_id=https%3A%2F%2Fhai.stanford.edu%2Fnews%2Freflections-foundation-models&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-gpt4-report-38"><span class="mw-cite-backlink">^ <a href="#cite_ref-gpt4-report_38-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-gpt4-report_38-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFOpenAI2023" class="citation web cs1">OpenAI (2023). <a rel="nofollow" class="external text" href="https://cdn.openai.com/papers/gpt-4.pdf">"GPT-4 Technical Report"</a> <span class="cs1-format">(PDF)</span>. <a rel="nofollow" class="external text" href="https://web.archive.org/web/20230314190904/https://cdn.openai.com/papers/gpt-4.pdf">Archived</a> <span class="cs1-format">(PDF)</span> from the original on 2023-03-14<span class="reference-accessdate">. Retrieved <span class="nowrap">2023-03-16</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=GPT-4+Technical+Report&amp;rft.date=2023&amp;rft.au=OpenAI&amp;rft_id=https%3A%2F%2Fcdn.openai.com%2Fpapers%2Fgpt-4.pdf&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-39"><span class="mw-cite-backlink"><b><a href="#cite_ref-39">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFZhuKirosZemelSalakhutdinov2015" class="citation conference cs1">Zhu, Yukun; Kiros, Ryan; Zemel, Rich; Salakhutdinov, Ruslan; Urtasun, Raquel; Torralba, Antonio; Fidler, Sanja (2015). <a rel="nofollow" class="external text" href="https://www.cv-foundation.org/openaccess/content_iccv_2015/html/Zhu_Aligning_Books_and_ICCV_2015_paper.html"><i>Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books</i></a>. IEEE International Conference on Computer Vision (ICCV) 2015. pp.&#160;<span class="nowrap">19–</span>27. <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/1506.06724">1506.06724</a></span>. <a rel="nofollow" class="external text" href="https://web.archive.org/web/20230205222219/https://www.cv-foundation.org/openaccess/content_iccv_2015/html/Zhu_Aligning_Books_and_ICCV_2015_paper.html">Archived</a> from the original on 2023-02-05<span class="reference-accessdate">. Retrieved <span class="nowrap">2023-02-07</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=conference&amp;rft.btitle=Aligning+Books+and+Movies%3A+Towards+Story-Like+Visual+Explanations+by+Watching+Movies+and+Reading+Books&amp;rft.pages=%3Cspan+class%3D%22nowrap%22%3E19-%3C%2Fspan%3E27&amp;rft.date=2015&amp;rft_id=info%3Aarxiv%2F1506.06724&amp;rft.aulast=Zhu&amp;rft.aufirst=Yukun&amp;rft.au=Kiros%2C+Ryan&amp;rft.au=Zemel%2C+Rich&amp;rft.au=Salakhutdinov%2C+Ruslan&amp;rft.au=Urtasun%2C+Raquel&amp;rft.au=Torralba%2C+Antonio&amp;rft.au=Fidler%2C+Sanja&amp;rft_id=https%3A%2F%2Fwww.cv-foundation.org%2Fopenaccess%2Fcontent_iccv_2015%2Fhtml%2FZhu_Aligning_Books_and_ICCV_2015_paper.html&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-40"><span class="mw-cite-backlink"><b><a href="#cite_ref-40">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFVincent2019" class="citation web cs1">Vincent, James (November 7, 2019). <a rel="nofollow" class="external text" href="https://www.theverge.com/2019/11/7/20953040/openai-text-generation-ai-gpt-2-full-model-release-1-5b-parameters">"OpenAI has published the text-generating AI it said was too dangerous to share"</a>. <i>The Verge</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=The+Verge&amp;rft.atitle=OpenAI+has+published+the+text-generating+AI+it+said+was+too+dangerous+to+share&amp;rft.date=2019-11-07&amp;rft.aulast=Vincent&amp;rft.aufirst=James&amp;rft_id=https%3A%2F%2Fwww.theverge.com%2F2019%2F11%2F7%2F20953040%2Fopenai-text-generation-ai-gpt-2-full-model-release-1-5b-parameters&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:2-41"><span class="mw-cite-backlink">^ <a href="#cite_ref-:2_41-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:2_41-1"><sup><i><b>b</b></i></sup></a> <a href="#cite_ref-:2_41-2"><sup><i><b>c</b></i></sup></a> <a href="#cite_ref-:2_41-3"><sup><i><b>d</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFBrownMannRyderSubbiah2020" class="citation journal cs1">Brown, Tom B.; Mann, Benjamin; Ryder, Nick; Subbiah, Melanie; Kaplan, Jared; Dhariwal, Prafulla; Neelakantan, Arvind; Shyam, Pranav; Sastry, Girish; Askell, Amanda; Agarwal, Sandhini; Herbert-Voss, Ariel; Krueger, Gretchen; Henighan, Tom; Child, Rewon; Ramesh, Aditya; Ziegler, Daniel M.; Wu, Jeffrey; Winter, Clemens; Hesse, Christopher; Chen, Mark; Sigler, Eric; Litwin, Mateusz; Gray, Scott; Chess, Benjamin; Clark, Jack; Berner, Christopher; McCandlish, Sam; Radford, Alec; Sutskever, Ilya; Amodei, Dario (May 28, 2020). "Language Models are Few-Shot Learners". <i>NeurIPS</i>. <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/2005.14165v4">2005.14165v4</a></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=NeurIPS&amp;rft.atitle=Language+Models+are+Few-Shot+Learners&amp;rft.date=2020-05-28&amp;rft_id=info%3Aarxiv%2F2005.14165v4&amp;rft.aulast=Brown&amp;rft.aufirst=Tom+B.&amp;rft.au=Mann%2C+Benjamin&amp;rft.au=Ryder%2C+Nick&amp;rft.au=Subbiah%2C+Melanie&amp;rft.au=Kaplan%2C+Jared&amp;rft.au=Dhariwal%2C+Prafulla&amp;rft.au=Neelakantan%2C+Arvind&amp;rft.au=Shyam%2C+Pranav&amp;rft.au=Sastry%2C+Girish&amp;rft.au=Askell%2C+Amanda&amp;rft.au=Agarwal%2C+Sandhini&amp;rft.au=Herbert-Voss%2C+Ariel&amp;rft.au=Krueger%2C+Gretchen&amp;rft.au=Henighan%2C+Tom&amp;rft.au=Child%2C+Rewon&amp;rft.au=Ramesh%2C+Aditya&amp;rft.au=Ziegler%2C+Daniel+M.&amp;rft.au=Wu%2C+Jeffrey&amp;rft.au=Winter%2C+Clemens&amp;rft.au=Hesse%2C+Christopher&amp;rft.au=Chen%2C+Mark&amp;rft.au=Sigler%2C+Eric&amp;rft.au=Litwin%2C+Mateusz&amp;rft.au=Gray%2C+Scott&amp;rft.au=Chess%2C+Benjamin&amp;rft.au=Clark%2C+Jack&amp;rft.au=Berner%2C+Christopher&amp;rft.au=McCandlish%2C+Sam&amp;rft.au=Radford%2C+Alec&amp;rft.au=Sutskever%2C+Ilya&amp;rft.au=Amodei%2C+Dario&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:3-42"><span class="mw-cite-backlink">^ <a href="#cite_ref-:3_42-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:3_42-1"><sup><i><b>b</b></i></sup></a> <a href="#cite_ref-:3_42-2"><sup><i><b>c</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://epochai.org/mlinputs/visualization">"ML input trends visualization"</a>. <i>Epoch</i><span class="reference-accessdate">. Retrieved <span class="nowrap">2023-05-02</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Epoch&amp;rft.atitle=ML+input+trends+visualization&amp;rft_id=https%3A%2F%2Fepochai.org%2Fmlinputs%2Fvisualization&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:8-43"><span class="mw-cite-backlink">^ <a href="#cite_ref-:8_43-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:8_43-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFVer_Meer2023" class="citation web cs1">Ver Meer, Dave (June 1, 2023). <a rel="nofollow" class="external text" href="https://www.namepepper.com/chatgpt-users">"ChatGPT Statistics"</a>. <i>NamePepper</i><span class="reference-accessdate">. Retrieved <span class="nowrap">2023-06-09</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=NamePepper&amp;rft.atitle=ChatGPT+Statistics&amp;rft.date=2023-06-01&amp;rft.aulast=Ver+Meer&amp;rft.aufirst=Dave&amp;rft_id=https%3A%2F%2Fwww.namepepper.com%2Fchatgpt-users&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-44"><span class="mw-cite-backlink"><b><a href="#cite_ref-44">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation news cs1"><a rel="nofollow" class="external text" href="https://the-decoder.com/gpt-4-has-a-trillion-parameters/">"GPT-4 has more than a trillion parameters – Report"</a>. March 25, 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.atitle=GPT-4+has+more+than+a+trillion+parameters+%E2%80%93+Report&amp;rft.date=2023-03-25&amp;rft_id=https%3A%2F%2Fthe-decoder.com%2Fgpt-4-has-a-trillion-parameters%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-45"><span class="mw-cite-backlink"><b><a href="#cite_ref-45">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFVincent2023" class="citation web cs1">Vincent, James (March 14, 2023). <a rel="nofollow" class="external text" href="https://www.theverge.com/2023/3/14/23639313/google-ai-language-model-palm-api-challenge-openai">"Google opens up its AI language model PaLM to challenge OpenAI and GPT-3"</a>. <i>The Verge</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=The+Verge&amp;rft.atitle=Google+opens+up+its+AI+language+model+PaLM+to+challenge+OpenAI+and+GPT-3&amp;rft.date=2023-03-14&amp;rft.aulast=Vincent&amp;rft.aufirst=James&amp;rft_id=https%3A%2F%2Fwww.theverge.com%2F2023%2F3%2F14%2F23639313%2Fgoogle-ai-language-model-palm-api-challenge-openai&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-46"><span class="mw-cite-backlink"><b><a href="#cite_ref-46">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://aibusiness.com/nlp/google-opens-access-to-palm-language-model">"Google Opens Access to PaLM Language Model"</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=Google+Opens+Access+to+PaLM+Language+Model&amp;rft_id=https%3A%2F%2Faibusiness.com%2Fnlp%2Fgoogle-opens-access-to-palm-language-model&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-47"><span class="mw-cite-backlink"><b><a href="#cite_ref-47">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFIyer2022" class="citation web cs1">Iyer, Aparna (November 30, 2022). <a rel="nofollow" class="external text" href="https://analyticsindiamag.com/meet-gpt-jt-the-closest-open-source-alternative-to-gpt-3/">"Meet GPT-JT, the Closest Open Source Alternative to GPT-3"</a>. <i>Analytics India Magazine</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Analytics+India+Magazine&amp;rft.atitle=Meet+GPT-JT%2C+the+Closest+Open+Source+Alternative+to+GPT-3&amp;rft.date=2022-11-30&amp;rft.aulast=Iyer&amp;rft.aufirst=Aparna&amp;rft_id=https%3A%2F%2Fanalyticsindiamag.com%2Fmeet-gpt-jt-the-closest-open-source-alternative-to-gpt-3%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-48"><span class="mw-cite-backlink"><b><a href="#cite_ref-48">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.pcmag.com/news/meta-debuts-ai-language-model-but-its-only-for-researchers">"Meta Debuts AI Language Model, But It's Only for Researchers"</a>. <i>PCMAG</i>. 24 February 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=PCMAG&amp;rft.atitle=Meta+Debuts+AI+Language+Model%2C+But+It%27s+Only+for+Researchers&amp;rft.date=2023-02-24&amp;rft_id=https%3A%2F%2Fwww.pcmag.com%2Fnews%2Fmeta-debuts-ai-language-model-but-its-only-for-researchers&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-49"><span class="mw-cite-backlink"><b><a href="#cite_ref-49">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFIslam2023" class="citation web cs1">Islam, Arham (March 27, 2023). <a rel="nofollow" class="external text" href="https://web.archive.org/web/20230515010932/https://www.marktechpost.com/2023/03/27/multimodal-language-models-the-future-of-artificial-intelligence-ai/">"Multimodal Language Models: The Future of Artificial Intelligence (AI)"</a>. Archived from <a rel="nofollow" class="external text" href="https://www.marktechpost.com/2023/03/27/multimodal-language-models-the-future-of-artificial-intelligence-ai/">the original</a> on May 15, 2023<span class="reference-accessdate">. Retrieved <span class="nowrap">May 15,</span> 2023</span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=Multimodal+Language+Models%3A+The+Future+of+Artificial+Intelligence+%28AI%29&amp;rft.date=2023-03-27&amp;rft.aulast=Islam&amp;rft.aufirst=Arham&amp;rft_id=https%3A%2F%2Fwww.marktechpost.com%2F2023%2F03%2F27%2Fmultimodal-language-models-the-future-of-artificial-intelligence-ai%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-50"><span class="mw-cite-backlink"><b><a href="#cite_ref-50">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFIslam2022" class="citation web cs1">Islam, Arham (November 14, 2022). <a rel="nofollow" class="external text" href="https://www.marktechpost.com/2022/11/14/how-do-dall%c2%b7e-2-stable-diffusion-and-midjourney-work/">"How Do DALL·E 2, Stable Diffusion, and Midjourney Work?"</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=How+Do+DALL%C2%B7E+2%2C+Stable+Diffusion%2C+and+Midjourney+Work%3F&amp;rft.date=2022-11-14&amp;rft.aulast=Islam&amp;rft.aufirst=Arham&amp;rft_id=https%3A%2F%2Fwww.marktechpost.com%2F2022%2F11%2F14%2Fhow-do-dall%25c2%25b7e-2-stable-diffusion-and-midjourney-work%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-51"><span class="mw-cite-backlink"><b><a href="#cite_ref-51">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFSaha2023" class="citation web cs1">Saha, Shritama (January 4, 2023). <a rel="nofollow" class="external text" href="https://analyticsindiamag.com/google-launches-muse-a-new-text-to-image-transformer-model/">"Google Launches Muse, A New Text-to-Image Transformer Model"</a>. <i>Analytics India Magazine</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Analytics+India+Magazine&amp;rft.atitle=Google+Launches+Muse%2C+A+New+Text-to-Image+Transformer+Model&amp;rft.date=2023-01-04&amp;rft.aulast=Saha&amp;rft.aufirst=Shritama&amp;rft_id=https%3A%2F%2Fanalyticsindiamag.com%2Fgoogle-launches-muse-a-new-text-to-image-transformer-model%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-52"><span class="mw-cite-backlink"><b><a href="#cite_ref-52">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFWu_(et-al)2023" class="citation arxiv cs1">Wu (et-al), Chenfei (March 8, 2023). "Visual ChatGPT". <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/2303.04671">2303.04671</a></span> [<a rel="nofollow" class="external text" href="https://arxiv.org/archive/cs.CV">cs.CV</a>].</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=preprint&amp;rft.jtitle=arXiv&amp;rft.atitle=Visual+ChatGPT&amp;rft.date=2023-03-08&amp;rft_id=info%3Aarxiv%2F2303.04671&amp;rft.aulast=Wu+%28et-al%29&amp;rft.aufirst=Chenfei&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-53"><span class="mw-cite-backlink"><b><a href="#cite_ref-53">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFBommasani_(et-al)2022" class="citation arxiv cs1">Bommasani (et-al), Rishi (July 12, 2022). "On the Opportunities and Risks of Foundation Models". <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/2108.07258">2108.07258</a></span> [<a rel="nofollow" class="external text" href="https://arxiv.org/archive/cs.LG">cs.LG</a>].</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=preprint&amp;rft.jtitle=arXiv&amp;rft.atitle=On+the+Opportunities+and+Risks+of+Foundation+Models&amp;rft.date=2022-07-12&amp;rft_id=info%3Aarxiv%2F2108.07258&amp;rft.aulast=Bommasani+%28et-al%29&amp;rft.aufirst=Rishi&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-instructgpt-blog-54"><span class="mw-cite-backlink">^ <a href="#cite_ref-instructgpt-blog_54-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-instructgpt-blog_54-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://openai.com/research/instruction-following">"Aligning language models to follow instructions"</a>. <i>openai.com</i>. <a rel="nofollow" class="external text" href="https://web.archive.org/web/20230323110040/https://openai.com/research/instruction-following">Archived</a> from the original on 23 March 2023<span class="reference-accessdate">. Retrieved <span class="nowrap">23 March</span> 2023</span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=openai.com&amp;rft.atitle=Aligning+language+models+to+follow+instructions&amp;rft_id=https%3A%2F%2Fopenai.com%2Fresearch%2Finstruction-following&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-instructgpt-paper-55"><span class="mw-cite-backlink">^ <a href="#cite_ref-instructgpt-paper_55-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-instructgpt-paper_55-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFOuyangWuJiangAlmeida2022" class="citation journal cs1">Ouyang, Long; Wu, Jeff; Jiang, Xu; et&#160;al. (4 November 2022). "Training language models to follow instructions with human feedback". <i>NeurIPS</i>. <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/2203.02155">2203.02155</a></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=NeurIPS&amp;rft.atitle=Training+language+models+to+follow+instructions+with+human+feedback&amp;rft.date=2022-11-04&amp;rft_id=info%3Aarxiv%2F2203.02155&amp;rft.aulast=Ouyang&amp;rft.aufirst=Long&amp;rft.au=Wu%2C+Jeff&amp;rft.au=Jiang%2C+Xu&amp;rft.au=Almeida%2C+Diogo&amp;rft.au=Wainwright%2C+Carroll+L.&amp;rft.au=Mishkin%2C+Pamela&amp;rft.au=Zhang%2C+Chong&amp;rft.au=Agarwal%2C+Sandhini&amp;rft.au=Slama%2C+Katarina&amp;rft.au=Ray%2C+Alex&amp;rft.au=Schulman%2C+John&amp;rft.au=Hilton%2C+Jacob&amp;rft.au=Kelton%2C+Fraser&amp;rft.au=Miller%2C+Luke&amp;rft.au=Simens%2C+Maddie&amp;rft.au=Askell%2C+Amanda&amp;rft.au=Welinder%2C+Peter&amp;rft.au=Christiano%2C+Paul&amp;rft.au=Leike%2C+Jan&amp;rft.au=Lowe%2C+Ryan&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-56"><span class="mw-cite-backlink"><b><a href="#cite_ref-56">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFRamnani2022" class="citation web cs1">Ramnani, Meeta (January 28, 2022). <a rel="nofollow" class="external text" href="https://analyticsindiamag.com/openai-dumps-its-own-gpt-3-for-something-called-instructgpt-and-for-right-reason/">"OpenAI dumps its own GPT-3 for something called InstructGPT, and for right reason"</a>. <i>Analytics India Magazine</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Analytics+India+Magazine&amp;rft.atitle=OpenAI+dumps+its+own+GPT-3+for+something+called+InstructGPT%2C+and+for+right+reason&amp;rft.date=2022-01-28&amp;rft.aulast=Ramnani&amp;rft.aufirst=Meeta&amp;rft_id=https%3A%2F%2Fanalyticsindiamag.com%2Fopenai-dumps-its-own-gpt-3-for-something-called-instructgpt-and-for-right-reason%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-57"><span class="mw-cite-backlink"><b><a href="#cite_ref-57">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://crfm.stanford.edu/2023/03/13/alpaca.html">"Stanford CRFM"</a>. <i>crfm.stanford.edu</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=crfm.stanford.edu&amp;rft.atitle=Stanford+CRFM&amp;rft_id=https%3A%2F%2Fcrfm.stanford.edu%2F2023%2F03%2F13%2Falpaca.html&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-58"><span class="mw-cite-backlink"><b><a href="#cite_ref-58">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.databricks.com/blog/2023/04/12/dolly-first-open-commercially-viable-instruction-tuned-llm">"Free Dolly: Introducing the World's First Truly Open Instruction-Tuned LLM"</a>. <i>Databricks</i>. April 12, 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Databricks&amp;rft.atitle=Free+Dolly%3A+Introducing+the+World%27s+First+Truly+Open+Instruction-Tuned+LLM&amp;rft.date=2023-04-12&amp;rft_id=https%3A%2F%2Fwww.databricks.com%2Fblog%2F2023%2F04%2F12%2Fdolly-first-open-commercially-viable-instruction-tuned-llm&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-chatgpt-blog-59"><span class="mw-cite-backlink">^ <a href="#cite_ref-chatgpt-blog_59-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-chatgpt-blog_59-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://openai.com/blog/chatgpt">"Introducing ChatGPT"</a>. <i>openai.com</i>. <a rel="nofollow" class="external text" href="https://web.archive.org/web/20230316001700/https://openai.com/blog/chatgpt/">Archived</a> from the original on 2023-03-16<span class="reference-accessdate">. Retrieved <span class="nowrap">2023-03-16</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=openai.com&amp;rft.atitle=Introducing+ChatGPT&amp;rft_id=https%3A%2F%2Fopenai.com%2Fblog%2Fchatgpt&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-60"><span class="mw-cite-backlink"><b><a href="#cite_ref-60">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFWiggers2023" class="citation web cs1">Wiggers, Kyle (May 4, 2023). <a rel="nofollow" class="external text" href="https://techcrunch.com/2023/05/04/microsoft-doubles-down-on-ai-with-new-bing-features/">"Microsoft doubles down on AI with new Bing features"</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=Microsoft+doubles+down+on+AI+with+new+Bing+features&amp;rft.date=2023-05-04&amp;rft.aulast=Wiggers&amp;rft.aufirst=Kyle&amp;rft_id=https%3A%2F%2Ftechcrunch.com%2F2023%2F05%2F04%2Fmicrosoft-doubles-down-on-ai-with-new-bing-features%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-61"><span class="mw-cite-backlink"><b><a href="#cite_ref-61">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.cnet.com/tech/services-and-software/chatgpt-vs-bing-vs-google-bard-which-ai-is-the-most-helpful/">"ChatGPT vs. Bing vs. Google Bard: Which AI Is the Most Helpful?"</a>. <i>CNET</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=CNET&amp;rft.atitle=ChatGPT+vs.+Bing+vs.+Google+Bard%3A+Which+AI+Is+the+Most+Helpful%3F&amp;rft_id=https%3A%2F%2Fwww.cnet.com%2Ftech%2Fservices-and-software%2Fchatgpt-vs-bing-vs-google-bard-which-ai-is-the-most-helpful%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-62"><span class="mw-cite-backlink"><b><a href="#cite_ref-62">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://mashable.com/article/autogpt-ai-agents-how-to-get-access">"Auto-GPT, BabyAGI, and AgentGPT: How to use AI agents"</a>. <i>Mashable</i>. April 19, 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Mashable&amp;rft.atitle=Auto-GPT%2C+BabyAGI%2C+and+AgentGPT%3A+How+to+use+AI+agents&amp;rft.date=2023-04-19&amp;rft_id=https%3A%2F%2Fmashable.com%2Farticle%2Fautogpt-ai-agents-how-to-get-access&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-63"><span class="mw-cite-backlink"><b><a href="#cite_ref-63">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFMarr" class="citation web cs1">Marr, Bernard. <a rel="nofollow" class="external text" href="https://www.forbes.com/sites/bernardmarr/2023/04/24/auto-gpt-may-be-the-strong-ai-tool-that-surpasses-chatgpt/">"Auto-GPT May Be The Strong AI Tool That Surpasses ChatGPT"</a>. <i>Forbes</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Forbes&amp;rft.atitle=Auto-GPT+May+Be+The+Strong+AI+Tool+That+Surpasses+ChatGPT&amp;rft.aulast=Marr&amp;rft.aufirst=Bernard&amp;rft_id=https%3A%2F%2Fwww.forbes.com%2Fsites%2Fbernardmarr%2F2023%2F04%2F24%2Fauto-gpt-may-be-the-strong-ai-tool-that-surpasses-chatgpt%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-64"><span class="mw-cite-backlink"><b><a href="#cite_ref-64">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.infoq.com/news/2023/04/microsoft-visual-chatgpt/">"Microsoft Open-Sources Multimodal Chatbot Visual ChatGPT"</a>. <i>InfoQ</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=InfoQ&amp;rft.atitle=Microsoft+Open-Sources+Multimodal+Chatbot+Visual+ChatGPT&amp;rft_id=https%3A%2F%2Fwww.infoq.com%2Fnews%2F2023%2F04%2Fmicrosoft-visual-chatgpt%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-65"><span class="mw-cite-backlink"><b><a href="#cite_ref-65">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFEdwards2023" class="citation web cs1">Edwards, Benj (January 9, 2023). <a rel="nofollow" class="external text" href="https://arstechnica.com/information-technology/2023/01/microsofts-new-ai-can-simulate-anyones-voice-with-3-seconds-of-audio/">"Microsoft's new AI can simulate anyone's voice with 3 seconds of audio"</a>. <i>Ars Technica</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Ars+Technica&amp;rft.atitle=Microsoft%27s+new+AI+can+simulate+anyone%27s+voice+with+3+seconds+of+audio&amp;rft.date=2023-01-09&amp;rft.aulast=Edwards&amp;rft.aufirst=Benj&amp;rft_id=https%3A%2F%2Farstechnica.com%2Finformation-technology%2F2023%2F01%2Fmicrosofts-new-ai-can-simulate-anyones-voice-with-3-seconds-of-audio%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-66"><span class="mw-cite-backlink"><b><a href="#cite_ref-66">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFMorrison2023" class="citation web cs1">Morrison, Ryan (March 7, 2023). <a rel="nofollow" class="external text" href="https://techmonitor.ai/technology/ai-and-automation/salesforce-einsteingpt-openai-chatgpt">"Salesforce launches EinsteinGPT built with OpenAI technology"</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=Salesforce+launches+EinsteinGPT+built+with+OpenAI+technology&amp;rft.date=2023-03-07&amp;rft.aulast=Morrison&amp;rft.aufirst=Ryan&amp;rft_id=https%3A%2F%2Ftechmonitor.ai%2Ftechnology%2Fai-and-automation%2Fsalesforce-einsteingpt-openai-chatgpt&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-67"><span class="mw-cite-backlink"><b><a href="#cite_ref-67">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFSharmaSharma2023" class="citation journal cs1">Sharma, Animesh K.; Sharma, Rahul (2023). <a rel="nofollow" class="external text" href="https://ideas.repec.org/s/aza/jcms00.html">"The role of generative pretrained transformers (GPTs) in revolutionising digital marketing: A conceptual model"</a>. <i>Journal of Cultural Marketing Strategy</i>. <b>8</b> (1): <span class="nowrap">80–</span>90. <a href="/wiki/Doi_(identifier)" class="mw-redirect" title="Doi (identifier)">doi</a>:<a rel="nofollow" class="external text" href="https://doi.org/10.69554%2FTLVQ2275">10.69554/TLVQ2275</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Journal+of+Cultural+Marketing+Strategy&amp;rft.atitle=The+role+of+generative+pretrained+transformers+%28GPTs%29+in+revolutionising+digital+marketing%3A+A+conceptual+model&amp;rft.volume=8&amp;rft.issue=1&amp;rft.pages=%3Cspan+class%3D%22nowrap%22%3E80-%3C%2Fspan%3E90&amp;rft.date=2023&amp;rft_id=info%3Adoi%2F10.69554%2FTLVQ2275&amp;rft.aulast=Sharma&amp;rft.aufirst=Animesh+K.&amp;rft.au=Sharma%2C+Rahul&amp;rft_id=https%3A%2F%2Fideas.repec.org%2Fs%2Faza%2Fjcms00.html&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-68"><span class="mw-cite-backlink"><b><a href="#cite_ref-68">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFLeswing2023" class="citation web cs1">Leswing, Kif (April 13, 2023). <a rel="nofollow" class="external text" href="https://www.cnbc.com/2023/04/13/bloomberg-plans-to-integrate-gpt-style-ai-into-its-terminal.html">"Bloomberg plans to integrate GPT-style A.I. into its terminal"</a>. <i>CNBC</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=CNBC&amp;rft.atitle=Bloomberg+plans+to+integrate+GPT-style+A.I.+into+its+terminal&amp;rft.date=2023-04-13&amp;rft.aulast=Leswing&amp;rft.aufirst=Kif&amp;rft_id=https%3A%2F%2Fwww.cnbc.com%2F2023%2F04%2F13%2Fbloomberg-plans-to-integrate-gpt-style-ai-into-its-terminal.html&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-69"><span class="mw-cite-backlink"><b><a href="#cite_ref-69">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFMelendez2023" class="citation news cs1">Melendez, Steven (May 4, 2023). <a rel="nofollow" class="external text" href="https://www.fastcompany.com/90891522/the-learning-nonprofit-khan-academy-piloting-a-version-of-gpt-called-khanmigo">"Learning nonprofit Khan Academy is piloting a version of GPT called Khanmigo"</a>. <i>Fast Company</i><span class="reference-accessdate">. Retrieved <span class="nowrap">May 22,</span> 2023</span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Fast+Company&amp;rft.atitle=Learning+nonprofit+Khan+Academy+is+piloting+a+version+of+GPT+called+Khanmigo&amp;rft.date=2023-05-04&amp;rft.aulast=Melendez&amp;rft.aufirst=Steven&amp;rft_id=https%3A%2F%2Fwww.fastcompany.com%2F90891522%2Fthe-learning-nonprofit-khan-academy-piloting-a-version-of-gpt-called-khanmigo&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-70"><span class="mw-cite-backlink"><b><a href="#cite_ref-70">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://thejournal.com/articles/2023/03/14/khan-academy-pilots-gpt-4-powered-tool-khanmigo-for-teachers.aspx">"Khan Academy Pilots GPT-4 Powered Tool Khanmigo for Teachers"</a>. <i>THE Journal</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=THE+Journal&amp;rft.atitle=Khan+Academy+Pilots+GPT-4+Powered+Tool+Khanmigo+for+Teachers&amp;rft_id=https%3A%2F%2Fthejournal.com%2Farticles%2F2023%2F03%2F14%2Fkhan-academy-pilots-gpt-4-powered-tool-khanmigo-for-teachers.aspx&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-71"><span class="mw-cite-backlink"><b><a href="#cite_ref-71">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFHachman2023" class="citation web cs1">Hachman, Mark (May 4, 2023). <a rel="nofollow" class="external text" href="https://www.pcworld.com/article/1807402/slack-gpt-will-bring-ai-chatbots-to-your-conversations.html">"Slack GPT will bring AI chatbots to your conversations"</a>. <i>PCWorld</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=PCWorld&amp;rft.atitle=Slack+GPT+will+bring+AI+chatbots+to+your+conversations&amp;rft.date=2023-05-04&amp;rft.aulast=Hachman&amp;rft.aufirst=Mark&amp;rft_id=https%3A%2F%2Fwww.pcworld.com%2Farticle%2F1807402%2Fslack-gpt-will-bring-ai-chatbots-to-your-conversations.html&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-72"><span class="mw-cite-backlink"><b><a href="#cite_ref-72">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFLuo_(et-al)2023" class="citation journal cs1">Luo (et-al), Renqian (April 3, 2023). "BioGPT: Generative pre-trained transformer for biomedical text generation and mining". <i>Briefings in Bioinformatics</i>. <b>23</b> (6). <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/2210.10341">2210.10341</a></span>. <a href="/wiki/Doi_(identifier)" class="mw-redirect" title="Doi (identifier)">doi</a>:<a rel="nofollow" class="external text" href="https://doi.org/10.1093%2Fbib%2Fbbac409">10.1093/bib/bbac409</a>. <a href="/wiki/PMID_(identifier)" class="mw-redirect" title="PMID (identifier)">PMID</a>&#160;<a rel="nofollow" class="external text" href="https://pubmed.ncbi.nlm.nih.gov/36156661">36156661</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Briefings+in+Bioinformatics&amp;rft.atitle=BioGPT%3A+Generative+pre-trained+transformer+for+biomedical+text+generation+and+mining&amp;rft.volume=23&amp;rft.issue=6&amp;rft.date=2023-04-03&amp;rft_id=info%3Aarxiv%2F2210.10341&amp;rft_id=info%3Apmid%2F36156661&amp;rft_id=info%3Adoi%2F10.1093%2Fbib%2Fbbac409&amp;rft.aulast=Luo+%28et-al%29&amp;rft.aufirst=Renqian&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-73"><span class="mw-cite-backlink"><b><a href="#cite_ref-73">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://wire19.com/chatgpt-plugins/">"Know about ChatGPT's 13 best plugins, designed to improve your overall user experience"</a>. <i>Latest Digital Transformation Trends &#124; Cloud News &#124; Wire19</i>. May 5, 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Latest+Digital+Transformation+Trends+%26%23124%3B+Cloud+News+%26%23124%3B+Wire19&amp;rft.atitle=Know+about+ChatGPT%27s+13+best+plugins%2C+designed+to+improve+your+overall+user+experience&amp;rft.date=2023-05-05&amp;rft_id=https%3A%2F%2Fwire19.com%2Fchatgpt-plugins%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-74"><span class="mw-cite-backlink"><b><a href="#cite_ref-74">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://openai.com/blog/chatgpt-plugins">"ChatGPT plugins"</a>. <i>openai.com</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=openai.com&amp;rft.atitle=ChatGPT+plugins&amp;rft_id=https%3A%2F%2Fopenai.com%2Fblog%2Fchatgpt-plugins&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-75"><span class="mw-cite-backlink"><b><a href="#cite_ref-75">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.makeuseof.com/how-use-chatgpt-google-sheets/">"How to Use ChatGPT on Google Sheets With GPT for Sheets and Docs"</a>. <i>MUO</i>. March 12, 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=MUO&amp;rft.atitle=How+to+Use+ChatGPT+on+Google+Sheets+With+GPT+for+Sheets+and+Docs&amp;rft.date=2023-03-12&amp;rft_id=https%3A%2F%2Fwww.makeuseof.com%2Fhow-use-chatgpt-google-sheets%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-76"><span class="mw-cite-backlink"><b><a href="#cite_ref-76">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFAsay2023" class="citation web cs1">Asay, Matt (February 27, 2023). <a rel="nofollow" class="external text" href="https://www.infoworld.com/article/3689175/embrace-and-extend-excel-for-ai-data-prep.html">"Embrace and extend Excel for AI data prep"</a>. <i>InfoWorld</i>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=InfoWorld&amp;rft.atitle=Embrace+and+extend+Excel+for+AI+data+prep&amp;rft.date=2023-02-27&amp;rft.aulast=Asay&amp;rft.aufirst=Matt&amp;rft_id=https%3A%2F%2Fwww.infoworld.com%2Farticle%2F3689175%2Fembrace-and-extend-excel-for-ai-data-prep.html&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-77"><span class="mw-cite-backlink"><b><a href="#cite_ref-77">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.techopedia.com/definition/openai-gpts">"OpenAI GPTS"</a>. 10 November 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=OpenAI+GPTS&amp;rft.date=2023-11-10&amp;rft_id=https%3A%2F%2Fwww.techopedia.com%2Fdefinition%2Fopenai-gpts&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:5-78"><span class="mw-cite-backlink">^ <a href="#cite_ref-:5_78-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:5_78-1"><sup><i><b>b</b></i></sup></a> <a href="#cite_ref-:5_78-2"><sup><i><b>c</b></i></sup></a> <a href="#cite_ref-:5_78-3"><sup><i><b>d</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFHicks2023" class="citation web cs1">Hicks, William (May 10, 2023). <a rel="nofollow" class="external text" href="https://www.bizjournals.com/sanfrancisco/inno/stories/news/2023/05/10/openai-startups-gpt.html">"ChatGPT creator OpenAI is asking startups to remove 'GPT' from their names"</a>. <i><a href="/wiki/American_City_Business_Journals" title="American City Business Journals">The Business Journal</a></i><span class="reference-accessdate">. Retrieved <span class="nowrap">2023-05-21</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=The+Business+Journal&amp;rft.atitle=ChatGPT+creator+OpenAI+is+asking+startups+to+remove+%27GPT%27+from+their+names&amp;rft.date=2023-05-10&amp;rft.aulast=Hicks&amp;rft.aufirst=William&amp;rft_id=https%3A%2F%2Fwww.bizjournals.com%2Fsanfrancisco%2Finno%2Fstories%2Fnews%2F2023%2F05%2F10%2Fopenai-startups-gpt.html&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-79"><span class="mw-cite-backlink"><b><a href="#cite_ref-79">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFOpenAI2023" class="citation web cs1">OpenAI (April 24, 2023). <a rel="nofollow" class="external text" href="https://openai.com/brand">"Brand Guidelines"</a><span class="reference-accessdate">. Retrieved <span class="nowrap">21 May</span> 2023</span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=Brand+Guidelines&amp;rft.date=2023-04-24&amp;rft.au=OpenAI&amp;rft_id=https%3A%2F%2Fopenai.com%2Fbrand&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:11-80"><span class="mw-cite-backlink">^ <a href="#cite_ref-:11_80-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:11_80-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://openai.com/brand#models">"Brand guidelines"</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=Brand+guidelines&amp;rft_id=https%3A%2F%2Fopenai.com%2Fbrand%23models&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-81"><span class="mw-cite-backlink"><b><a href="#cite_ref-81">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://openai.com/blog/introducing-gpts">"Introducing GPTS"</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=Introducing+GPTS&amp;rft_id=https%3A%2F%2Fopenai.com%2Fblog%2Fintroducing-gpts&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:6-82"><span class="mw-cite-backlink">^ <a href="#cite_ref-:6_82-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:6_82-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFHeah2023" class="citation news cs1">Heah, Alexa (April 26, 2023). <a rel="nofollow" class="external text" href="https://designtaxi.com/news/423211/OpenAI-Unsuccessful-At-Speeding-Up-Its-Attempt-To-Trademark-GPT/">"OpenAI Unsuccessful At Speeding Up Its Attempt To Trademark 'GPT'<span class="cs1-kern-right"></span>"</a>. <i>DesignTAXI</i><span class="reference-accessdate">. Retrieved <span class="nowrap">May 21,</span> 2023</span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=DesignTAXI&amp;rft.atitle=OpenAI+Unsuccessful+At+Speeding+Up+Its+Attempt+To+Trademark+%27GPT%27&amp;rft.date=2023-04-26&amp;rft.aulast=Heah&amp;rft.aufirst=Alexa&amp;rft_id=https%3A%2F%2Fdesigntaxi.com%2Fnews%2F423211%2FOpenAI-Unsuccessful-At-Speeding-Up-Its-Attempt-To-Trademark-GPT%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-83"><span class="mw-cite-backlink"><b><a href="#cite_ref-83">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://tsdr.uspto.gov/documentviewer?caseId=sn97733259&amp;docId=NFIN20230525093517#docIndex=4&amp;page=1">"NONFINAL OFFICE ACTION"</a>. <i>USPTO</i>. May 25, 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=USPTO&amp;rft.atitle=NONFINAL+OFFICE+ACTION&amp;rft.date=2023-05-25&amp;rft_id=https%3A%2F%2Ftsdr.uspto.gov%2Fdocumentviewer%3FcaseId%3Dsn97733259%26docId%3DNFIN20230525093517%23docIndex%3D4%26page%3D1&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-84"><span class="mw-cite-backlink"><b><a href="#cite_ref-84">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://digital.gov/resources/u-s-trademark-law/">"U.S. Trademark Law"</a>. December 2015.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=U.S.+Trademark+Law&amp;rft.date=2015-12&amp;rft_id=https%3A%2F%2Fdigital.gov%2Fresources%2Fu-s-trademark-law%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-85"><span class="mw-cite-backlink"><b><a href="#cite_ref-85">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.inta.org/fact-sheets/international-trademark-rights/">"International Trademark Rights"</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=International+Trademark+Rights&amp;rft_id=https%3A%2F%2Fwww.inta.org%2Ffact-sheets%2Finternational-trademark-rights%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-86"><span class="mw-cite-backlink"><b><a href="#cite_ref-86">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.techtimes.com/articles/290766/20230425/openai-trademark-gpt-chatgpt-rise-ai-chatbots.htm">"OpenAI Wants to Trademark 'GPT' Amid Rise of AI Chatbots"</a>. Tech Times. 25 April 2023<span class="reference-accessdate">. Retrieved <span class="nowrap">2023-05-21</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=OpenAI+Wants+to+Trademark+%27GPT%27+Amid+Rise+of+AI+Chatbots&amp;rft.pub=Tech+Times&amp;rft.date=2023-04-25&amp;rft_id=https%3A%2F%2Fwww.techtimes.com%2Farticles%2F290766%2F20230425%2Fopenai-trademark-gpt-chatgpt-rise-ai-chatbots.htm&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-87"><span class="mw-cite-backlink"><b><a href="#cite_ref-87">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFLouise2023" class="citation web cs1">Louise, Nickie (April 3, 2023). <a rel="nofollow" class="external text" href="https://techstartups.com/2023/04/03/openai-files-a-udrp-case-against-the-current-owner-of-chatgpt-com/">"OpenAI files a UDRP case against the current owner of ChatGPT.com"</a><span class="reference-accessdate">. Retrieved <span class="nowrap">2023-05-21</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=OpenAI+files+a+UDRP+case+against+the+current+owner+of+ChatGPT.com&amp;rft.date=2023-04-03&amp;rft.aulast=Louise&amp;rft.aufirst=Nickie&amp;rft_id=https%3A%2F%2Ftechstartups.com%2F2023%2F04%2F03%2Fopenai-files-a-udrp-case-against-the-current-owner-of-chatgpt-com%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-:7-88"><span class="mw-cite-backlink">^ <a href="#cite_ref-:7_88-0"><sup><i><b>a</b></i></sup></a> <a href="#cite_ref-:7_88-1"><sup><i><b>b</b></i></sup></a></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFDemcak2023" class="citation web cs1">Demcak, Tramatm-Igor (2023-04-26). <a rel="nofollow" class="external text" href="https://web.archive.org/web/20230505162827/https://www.lexology.com/library/detail.aspx?g=763049f7-7ef8-4a68-bdb1-2e4fa194b7ad">"OpenAI's Battle for Brand Protection: Can GPT be trademarked?"</a>. <i>Lexology</i>. Archived from <a rel="nofollow" class="external text" href="https://www.lexology.com/library/detail.aspx?g=763049f7-7ef8-4a68-bdb1-2e4fa194b7ad">the original</a> on May 5, 2023<span class="reference-accessdate">. Retrieved <span class="nowrap">2023-05-22</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Lexology&amp;rft.atitle=OpenAI%27s+Battle+for+Brand+Protection%3A+Can+GPT+be+trademarked%3F&amp;rft.date=2023-04-26&amp;rft.aulast=Demcak&amp;rft.aufirst=Tramatm-Igor&amp;rft_id=https%3A%2F%2Fwww.lexology.com%2Flibrary%2Fdetail.aspx%3Fg%3D763049f7-7ef8-4a68-bdb1-2e4fa194b7ad&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-89"><span class="mw-cite-backlink"><b><a href="#cite_ref-89">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFLawton2023" class="citation web cs1">Lawton, George (April 20, 2023). <a rel="nofollow" class="external text" href="https://web.archive.org/web/20230509150052/https://www.techtarget.com/searchenterpriseai/feature/ChatGPT-vs-GPT-How-are-they-different">"ChatGPT vs. GPT: How are they different? | TechTarget"</a>. <i>Enterprise AI</i>. Archived from <a rel="nofollow" class="external text" href="https://www.techtarget.com/searchenterpriseai/feature/ChatGPT-vs-GPT-How-are-they-different">the original</a> on May 9, 2023<span class="reference-accessdate">. Retrieved <span class="nowrap">2023-05-21</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Enterprise+AI&amp;rft.atitle=ChatGPT+vs.+GPT%3A+How+are+they+different%3F+%7C+TechTarget&amp;rft.date=2023-04-20&amp;rft.aulast=Lawton&amp;rft.aufirst=George&amp;rft_id=https%3A%2F%2Fwww.techtarget.com%2Fsearchenterpriseai%2Ffeature%2FChatGPT-vs-GPT-How-are-they-different&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-90"><span class="mw-cite-backlink"><b><a href="#cite_ref-90">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFRobb2023" class="citation web cs1">Robb, Drew (2023-04-12). <a rel="nofollow" class="external text" href="https://www.eweek.com/artificial-intelligence/gpt-4-vs-chatgpt/">"GPT-4 vs. ChatGPT: AI Chatbot Comparison"</a>. <i>eWEEK</i><span class="reference-accessdate">. Retrieved <span class="nowrap">2023-05-21</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=eWEEK&amp;rft.atitle=GPT-4+vs.+ChatGPT%3A+AI+Chatbot+Comparison&amp;rft.date=2023-04-12&amp;rft.aulast=Robb&amp;rft.aufirst=Drew&amp;rft_id=https%3A%2F%2Fwww.eweek.com%2Fartificial-intelligence%2Fgpt-4-vs-chatgpt%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-91"><span class="mw-cite-backlink"><b><a href="#cite_ref-91">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFRusso2023" class="citation news cs1">Russo, Philip (August 22, 2023). <a rel="nofollow" class="external text" href="https://commercialobserver.com/2023/08/jll-ai-gpt-proptech/">"The Genesis of Generative AI for Everything Everywhere All at Once in CRE"</a>. <i>Commercial Observer</i>. <a rel="nofollow" class="external text" href="https://web.archive.org/web/20230824103201/https://commercialobserver.com/2023/08/jll-ai-gpt-proptech/">Archived</a> from the original on August 24, 2023.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=Commercial+Observer&amp;rft.atitle=The+Genesis+of+Generative+AI+for+Everything+Everywhere+All+at+Once+in+CRE&amp;rft.date=2023-08-22&amp;rft.aulast=Russo&amp;rft.aufirst=Philip&amp;rft_id=https%3A%2F%2Fcommercialobserver.com%2F2023%2F08%2Fjll-ai-gpt-proptech%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-92"><span class="mw-cite-backlink"><b><a href="#cite_ref-92">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://www.law.cornell.edu/wex/trademark_infringement">"Trademark infringement"</a>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=unknown&amp;rft.btitle=Trademark+infringement&amp;rft_id=https%3A%2F%2Fwww.law.cornell.edu%2Fwex%2Ftrademark_infringement&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-93"><span class="mw-cite-backlink"><b><a href="#cite_ref-93">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFRheintgen2013" class="citation web cs1">Rheintgen, Husch Blackwell LLP-Kathleen A. (2013-08-16). <a rel="nofollow" class="external text" href="https://www.lexology.com/library/detail.aspx?g=4f7fc6dd-1d5f-41a1-beac-2638750faa75">"Branding 101: trademark descriptive fair use"</a>. <i>Lexology</i><span class="reference-accessdate">. Retrieved <span class="nowrap">2023-05-21</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=Lexology&amp;rft.atitle=Branding+101%3A+trademark+descriptive+fair+use&amp;rft.date=2013-08-16&amp;rft.aulast=Rheintgen&amp;rft.aufirst=Husch+Blackwell+LLP-Kathleen+A.&amp;rft_id=https%3A%2F%2Fwww.lexology.com%2Flibrary%2Fdetail.aspx%3Fg%3D4f7fc6dd-1d5f-41a1-beac-2638750faa75&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-94"><span class="mw-cite-backlink"><b><a href="#cite_ref-94">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation cs2"><a rel="nofollow" class="external text" href="https://github.com/openai/finetune-transformer-lm"><i>finetune-transformer-lm</i></a>, OpenAI, June 11, 2018<span class="reference-accessdate">, retrieved <span class="nowrap">2023-05-01</span></span></cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=book&amp;rft.btitle=finetune-transformer-lm&amp;rft.pub=OpenAI&amp;rft.date=2018-06-11&amp;rft_id=https%3A%2F%2Fgithub.com%2Fopenai%2Ffinetune-transformer-lm&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-95"><span class="mw-cite-backlink"><b><a href="#cite_ref-95">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://openai.com/research/gpt-2-1-5b-release">"GPT-2: 1.5B release"</a>. <i>openai.com</i><span class="reference-accessdate">. Retrieved <span class="nowrap">2023-05-01</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=openai.com&amp;rft.atitle=GPT-2%3A+1.5B+release&amp;rft_id=https%3A%2F%2Fopenai.com%2Fresearch%2Fgpt-2-1-5b-release&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-96"><span class="mw-cite-backlink"><b><a href="#cite_ref-96">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFSolaimanBrundageClarkAskell2019" class="citation arxiv cs1"><a href="/wiki/Irene_Solaiman" title="Irene Solaiman">Solaiman, Irene</a>; Brundage, Miles; Clark, Jack; Askell, Amanda; Herbert-Voss, Ariel; Wu, Jeff; Radford, Alec; Krueger, Gretchen; Kim, Jong Wook; Kreps, Sarah; McCain, Miles; Newhouse, Alex; Blazakis, Jason; McGuffie, Kris; Wang, Jasmine (2019-11-12). "Release Strategies and the Social Impacts of Language Models". <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/1908.09203">1908.09203</a></span> [<a rel="nofollow" class="external text" href="https://arxiv.org/archive/cs.CL">cs.CL</a>].</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=preprint&amp;rft.jtitle=arXiv&amp;rft.atitle=Release+Strategies+and+the+Social+Impacts+of+Language+Models&amp;rft.date=2019-11-12&amp;rft_id=info%3Aarxiv%2F1908.09203&amp;rft.aulast=Solaiman&amp;rft.aufirst=Irene&amp;rft.au=Brundage%2C+Miles&amp;rft.au=Clark%2C+Jack&amp;rft.au=Askell%2C+Amanda&amp;rft.au=Herbert-Voss%2C+Ariel&amp;rft.au=Wu%2C+Jeff&amp;rft.au=Radford%2C+Alec&amp;rft.au=Krueger%2C+Gretchen&amp;rft.au=Kim%2C+Jong+Wook&amp;rft.au=Kreps%2C+Sarah&amp;rft.au=McCain%2C+Miles&amp;rft.au=Newhouse%2C+Alex&amp;rft.au=Blazakis%2C+Jason&amp;rft.au=McGuffie%2C+Kris&amp;rft.au=Wang%2C+Jasmine&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-97"><span class="mw-cite-backlink"><b><a href="#cite_ref-97">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation cs2"><a rel="nofollow" class="external text" href="https://github.com/openai/gpt-2"><i>gpt-2</i></a>, OpenAI, 2023-05-01<span class="reference-accessdate">, retrieved <span class="nowrap">2023-05-01</span></span></cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&amp;rft.genre=book&amp;rft.btitle=gpt-2&amp;rft.pub=OpenAI&amp;rft.date=2023-05-01&amp;rft_id=https%3A%2F%2Fgithub.com%2Fopenai%2Fgpt-2&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-98"><span class="mw-cite-backlink"><b><a href="#cite_ref-98">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://web.archive.org/web/20230621182942/https://openai.com/research/webgpt">"WebGPT: Improving the factual accuracy of language models through web browsing"</a>. <i>openai.com</i>. Archived from <a rel="nofollow" class="external text" href="https://openai.com/research/webgpt">the original</a> on 21 Jun 2023<span class="reference-accessdate">. Retrieved <span class="nowrap">2023-07-02</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=openai.com&amp;rft.atitle=WebGPT%3A+Improving+the+factual+accuracy+of+language+models+through+web+browsing&amp;rft_id=https%3A%2F%2Fopenai.com%2Fresearch%2Fwebgpt&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-99"><span class="mw-cite-backlink"><b><a href="#cite_ref-99">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFNakanoHiltonBalajiWu2021" class="citation journal cs1">Nakano, Reiichiro; Hilton, Jacob; <a href="/wiki/Suchir_Balaji" title="Suchir Balaji">Balaji, Suchir</a>; Wu, Jeff; Ouyang, Long; Kim, Christina; Hesse, Christopher; Jain, Shantanu; Kosaraju, Vineet; Saunders, William; Jiang, Xu; Cobbe, Karl; Eloundou, Tyna; Krueger, Gretchen; Button, Kevin (2021-12-01). <a rel="nofollow" class="external text" href="https://ui.adsabs.harvard.edu/abs/2021arXiv211209332N">"WebGPT: Browser-assisted question-answering with human feedback"</a>. <i>CoRR</i>. <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/2112.09332">2112.09332</a></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=article&amp;rft.jtitle=CoRR&amp;rft.atitle=WebGPT%3A+Browser-assisted+question-answering+with+human+feedback&amp;rft.date=2021-12-01&amp;rft_id=info%3Aarxiv%2F2112.09332&amp;rft.aulast=Nakano&amp;rft.aufirst=Reiichiro&amp;rft.au=Hilton%2C+Jacob&amp;rft.au=Balaji%2C+Suchir&amp;rft.au=Wu%2C+Jeff&amp;rft.au=Ouyang%2C+Long&amp;rft.au=Kim%2C+Christina&amp;rft.au=Hesse%2C+Christopher&amp;rft.au=Jain%2C+Shantanu&amp;rft.au=Kosaraju%2C+Vineet&amp;rft.au=Saunders%2C+William&amp;rft.au=Jiang%2C+Xu&amp;rft.au=Cobbe%2C+Karl&amp;rft.au=Eloundou%2C+Tyna&amp;rft.au=Krueger%2C+Gretchen&amp;rft.au=Button%2C+Kevin&amp;rft_id=https%3A%2F%2Fui.adsabs.harvard.edu%2Fabs%2F2021arXiv211209332N&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-100"><span class="mw-cite-backlink"><b><a href="#cite_ref-100">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://openai.com/research/gpt-4">"GPT-4"</a>. <i>openai.com</i><span class="reference-accessdate">. Retrieved <span class="nowrap">2023-05-01</span></span>.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=openai.com&amp;rft.atitle=GPT-4&amp;rft_id=https%3A%2F%2Fopenai.com%2Fresearch%2Fgpt-4&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-101"><span class="mw-cite-backlink"><b><a href="#cite_ref-101">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFOpenAI2023" class="citation arxiv cs1">OpenAI (2023-03-27). "GPT-4 Technical Report". <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/2303.08774">2303.08774</a></span> [<a rel="nofollow" class="external text" href="https://arxiv.org/archive/cs.CL">cs.CL</a>].</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=preprint&amp;rft.jtitle=arXiv&amp;rft.atitle=GPT-4+Technical+Report&amp;rft.date=2023-03-27&amp;rft_id=info%3Aarxiv%2F2303.08774&amp;rft.au=OpenAI&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-102"><span class="mw-cite-backlink"><b><a href="#cite_ref-102">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite id="CITEREFBubeckChandrasekaranEldanGehrke2023" class="citation arxiv cs1">Bubeck, Sébastien; Chandrasekaran, Varun; Eldan, Ronen; Gehrke, Johannes; Horvitz, Eric; Kamar, Ece; Lee, Peter; Lee, Yin Tat; Li, Yuanzhi; Lundberg, Scott; Nori, Harsha; Palangi, Hamid; Ribeiro, Marco Tulio; Zhang, Yi (2023-04-13). "Sparks of Artificial General Intelligence: Early experiments with GPT-4". <a href="/wiki/ArXiv_(identifier)" class="mw-redirect" title="ArXiv (identifier)">arXiv</a>:<span class="id-lock-free" title="Freely accessible"><a rel="nofollow" class="external text" href="https://arxiv.org/abs/2303.12712">2303.12712</a></span> [<a rel="nofollow" class="external text" href="https://arxiv.org/archive/cs.CL">cs.CL</a>].</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=preprint&amp;rft.jtitle=arXiv&amp;rft.atitle=Sparks+of+Artificial+General+Intelligence%3A+Early+experiments+with+GPT-4&amp;rft.date=2023-04-13&amp;rft_id=info%3Aarxiv%2F2303.12712&amp;rft.aulast=Bubeck&amp;rft.aufirst=S%C3%A9bastien&amp;rft.au=Chandrasekaran%2C+Varun&amp;rft.au=Eldan%2C+Ronen&amp;rft.au=Gehrke%2C+Johannes&amp;rft.au=Horvitz%2C+Eric&amp;rft.au=Kamar%2C+Ece&amp;rft.au=Lee%2C+Peter&amp;rft.au=Lee%2C+Yin+Tat&amp;rft.au=Li%2C+Yuanzhi&amp;rft.au=Lundberg%2C+Scott&amp;rft.au=Nori%2C+Harsha&amp;rft.au=Palangi%2C+Hamid&amp;rft.au=Ribeiro%2C+Marco+Tulio&amp;rft.au=Zhang%2C+Yi&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> <li id="cite_note-103"><span class="mw-cite-backlink"><b><a href="#cite_ref-103">^</a></b></span> <span class="reference-text"><a rel="nofollow" class="external text" href="https://cdn.openai.com/papers/gpt-4-system-card.pdf">GPT-4 System Card</a>, OpenAI, March 23 2023 (Accessed May 22 2023).</span> </li> <li id="cite_note-104"><span class="mw-cite-backlink"><b><a href="#cite_ref-104">^</a></b></span> <span class="reference-text"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1238218222"><cite class="citation web cs1"><a rel="nofollow" class="external text" href="https://openai.com/index/hello-gpt-4o/">"Hello GPT-4o"</a>. <i>OpenAI</i>. May 13, 2024.</cite><span title="ctx_ver=Z39.88-2004&amp;rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&amp;rft.genre=unknown&amp;rft.jtitle=OpenAI&amp;rft.atitle=Hello+GPT-4o&amp;rft.date=2024-05-13&amp;rft_id=https%3A%2F%2Fopenai.com%2Findex%2Fhello-gpt-4o%2F&amp;rfr_id=info%3Asid%2Fen.wikipedia.org%3AGenerative+pre-trained+transformer" class="Z3988"></span></span> </li> </ol></div></div> <div class="navbox-styles"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1129693374"><style data-mw-deduplicate="TemplateStyles:r1236075235">.mw-parser-output .navbox{box-sizing:border-box;border:1px solid #a2a9b1;width:100%;clear:both;font-size:88%;text-align:center;padding:1px;margin:1em auto 0}.mw-parser-output .navbox .navbox{margin-top:0}.mw-parser-output .navbox+.navbox,.mw-parser-output .navbox+.navbox-styles+.navbox{margin-top:-1px}.mw-parser-output .navbox-inner,.mw-parser-output .navbox-subgroup{width:100%}.mw-parser-output .navbox-group,.mw-parser-output .navbox-title,.mw-parser-output .navbox-abovebelow{padding:0.25em 1em;line-height:1.5em;text-align:center}.mw-parser-output .navbox-group{white-space:nowrap;text-align:right}.mw-parser-output .navbox,.mw-parser-output .navbox-subgroup{background-color:#fdfdfd}.mw-parser-output .navbox-list{line-height:1.5em;border-color:#fdfdfd}.mw-parser-output .navbox-list-with-group{text-align:left;border-left-width:2px;border-left-style:solid}.mw-parser-output tr+tr>.navbox-abovebelow,.mw-parser-output tr+tr>.navbox-group,.mw-parser-output tr+tr>.navbox-image,.mw-parser-output tr+tr>.navbox-list{border-top:2px solid #fdfdfd}.mw-parser-output .navbox-title{background-color:#ccf}.mw-parser-output .navbox-abovebelow,.mw-parser-output .navbox-group,.mw-parser-output .navbox-subgroup .navbox-title{background-color:#ddf}.mw-parser-output .navbox-subgroup .navbox-group,.mw-parser-output .navbox-subgroup .navbox-abovebelow{background-color:#e6e6ff}.mw-parser-output .navbox-even{background-color:#f7f7f7}.mw-parser-output .navbox-odd{background-color:transparent}.mw-parser-output .navbox .hlist td dl,.mw-parser-output .navbox .hlist td ol,.mw-parser-output .navbox .hlist td ul,.mw-parser-output .navbox td.hlist dl,.mw-parser-output .navbox td.hlist ol,.mw-parser-output .navbox td.hlist ul{padding:0.125em 0}.mw-parser-output .navbox .navbar{display:block;font-size:100%}.mw-parser-output .navbox-title .navbar{float:left;text-align:left;margin-right:0.5em}body.skin--responsive .mw-parser-output .navbox-image img{max-width:none!important}@media print{body.ns-0 .mw-parser-output .navbox{display:none!important}}</style></div><div role="navigation" class="navbox" aria-labelledby="OpenAI158" style="padding:3px"><table class="nowraplinks hlist mw-collapsible autocollapse navbox-inner" style="border-spacing:0;background:transparent;color:inherit"><tbody><tr><th scope="col" class="navbox-title" colspan="3"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1129693374"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1239400231"><div class="navbar plainlinks hlist navbar-mini"><ul><li class="nv-view"><a href="/wiki/Template:OpenAI" title="Template:OpenAI"><abbr title="View this template">v</abbr></a></li><li class="nv-talk"><a href="/wiki/Template_talk:OpenAI" title="Template talk:OpenAI"><abbr title="Discuss this template">t</abbr></a></li><li class="nv-edit"><a href="/wiki/Special:EditPage/Template:OpenAI" title="Special:EditPage/Template:OpenAI"><abbr title="Edit this template">e</abbr></a></li></ul></div><div id="OpenAI158" style="font-size:114%;margin:0 4em"><a href="/wiki/OpenAI" title="OpenAI">OpenAI</a></div></th></tr><tr><th scope="row" class="navbox-group" style="width:1%">Products</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"></div><table class="nowraplinks navbox-subgroup" style="border-spacing:0"><tbody><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Chatbot" title="Chatbot">Chatbots</a></th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/ChatGPT" title="ChatGPT">ChatGPT</a> <ul><li><a href="/wiki/ChatGPT_in_education" title="ChatGPT in education">in education</a></li> <li><a href="/wiki/GPT_Store" title="GPT Store">GPT Store</a></li> <li><a href="/wiki/DALL-E" title="DALL-E">DALL-E</a></li> <li><a href="/wiki/SearchGPT" class="mw-redirect" title="SearchGPT">SearchGPT</a></li> <li><a href="/wiki/Sora_(text-to-video_model)" title="Sora (text-to-video model)">Sora</a></li> <li><a href="/wiki/Whisper_(speech_recognition_system)" title="Whisper (speech recognition system)">Whisper</a></li></ul></li> <li><a href="/wiki/GitHub_Copilot" title="GitHub Copilot">GitHub Copilot</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Foundation_model" title="Foundation model">Foundation models</a></th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/OpenAI_Codex" title="OpenAI Codex">OpenAI Codex</a></li> <li><a class="mw-selflink selflink">Generative pre-trained transformer</a> <ul><li><a href="/wiki/GPT-1" title="GPT-1">GPT-1</a></li> <li><a href="/wiki/GPT-2" title="GPT-2">GPT-2</a></li> <li><a href="/wiki/GPT-3" title="GPT-3">GPT-3</a></li> <li><a href="/wiki/GPT-4" title="GPT-4">GPT-4</a></li> <li><a href="/wiki/GPT-4o" title="GPT-4o">GPT-4o</a></li> <li><a href="/wiki/OpenAI_o1" title="OpenAI o1">o1</a></li> <li><a href="/wiki/OpenAI_o3" title="OpenAI o3">o3</a></li></ul></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/AI_agent" class="mw-redirect" title="AI agent">AI agents</a></th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Deep_Research" class="mw-redirect" title="Deep Research">Deep Research</a></li> <li><a href="/w/index.php?title=OpenAI_Operator&amp;action=edit&amp;redlink=1" class="new" title="OpenAI Operator (page does not exist)">Operator</a></li></ul> </div></td></tr></tbody></table><div></div></td><td class="noviewer navbox-image" rowspan="4" style="width:1px;padding:0 0 0 2px"><div><span typeof="mw:File"><a href="/wiki/File:OpenAI_logo_2025_(wordmark).svg" class="mw-file-description"><img src="//upload.wikimedia.org/wikipedia/commons/thumb/a/af/OpenAI_logo_2025_%28wordmark%29.svg/150px-OpenAI_logo_2025_%28wordmark%29.svg.png" decoding="async" width="150" height="40" class="mw-file-element" srcset="//upload.wikimedia.org/wikipedia/commons/thumb/a/af/OpenAI_logo_2025_%28wordmark%29.svg/225px-OpenAI_logo_2025_%28wordmark%29.svg.png 1.5x, //upload.wikimedia.org/wikipedia/commons/thumb/a/af/OpenAI_logo_2025_%28wordmark%29.svg/300px-OpenAI_logo_2025_%28wordmark%29.svg.png 2x" data-file-width="512" data-file-height="138" /></a></span></div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">People</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"></div><table class="nowraplinks navbox-subgroup" style="border-spacing:0"><tbody><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Senior_management" title="Senior management">Senior management</a></th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"></div><table class="nowraplinks navbox-subgroup" style="border-spacing:0"><tbody><tr><th scope="row" class="navbox-group" style="width:1%">Current</th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Sam_Altman" title="Sam Altman">Sam Altman</a> <ul><li><a href="/wiki/Removal_of_Sam_Altman_from_OpenAI" title="Removal of Sam Altman from OpenAI">removal</a></li></ul></li> <li><a href="/wiki/Greg_Brockman" title="Greg Brockman">Greg Brockman</a></li> <li><a href="/wiki/Sarah_Friar" title="Sarah Friar">Sarah Friar</a></li> <li><a href="/wiki/Scott_Schools" title="Scott Schools">Scott Schools</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Former</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Mira_Murati" title="Mira Murati">Mira Murati</a></li> <li><a href="/wiki/Emmett_Shear" title="Emmett Shear">Emmett Shear</a></li></ul> </div></td></tr></tbody></table><div></div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Board_of_directors" title="Board of directors">Board of directors</a></th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"></div><table class="nowraplinks navbox-subgroup" style="border-spacing:0"><tbody><tr><th scope="row" class="navbox-group" style="width:1%">Current</th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Sam_Altman" title="Sam Altman">Sam Altman</a></li> <li><a href="/wiki/Adam_D%27Angelo" title="Adam D&#39;Angelo">Adam D'Angelo</a></li> <li><a href="/wiki/Sue_Desmond-Hellmann" title="Sue Desmond-Hellmann">Sue Desmond-Hellmann</a></li> <li><a href="/wiki/Paul_Nakasone" title="Paul Nakasone">Paul Nakasone</a></li> <li><a href="/wiki/Adebayo_Ogunlesi" title="Adebayo Ogunlesi">Adebayo Ogunlesi</a></li> <li><a href="/wiki/Nicole_Seligman" title="Nicole Seligman">Nicole Seligman</a></li> <li><a href="/wiki/Fidji_Simo" title="Fidji Simo">Fidji Simo</a></li> <li><a href="/wiki/Lawrence_Summers" title="Lawrence Summers">Lawrence Summers</a></li> <li><a href="/wiki/Bret_Taylor" title="Bret Taylor">Bret Taylor</a> (chair)</li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Former</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Greg_Brockman" title="Greg Brockman">Greg Brockman</a> (2017–2023)</li> <li><a href="/wiki/Reid_Hoffman" title="Reid Hoffman">Reid Hoffman</a> (2019–2023)</li> <li><a href="/wiki/Will_Hurd" title="Will Hurd">Will Hurd</a> (2021–2023)</li> <li><a href="/wiki/Holden_Karnofsky" title="Holden Karnofsky">Holden Karnofsky</a> (2017–2021)</li> <li><a href="/wiki/Elon_Musk" title="Elon Musk">Elon Musk</a> (2015–2018)</li> <li><a href="/wiki/Ilya_Sutskever" title="Ilya Sutskever">Ilya Sutskever</a> (2017–2023)</li> <li><a href="/wiki/Helen_Toner" title="Helen Toner">Helen Toner</a> (2021–2023)</li> <li><a href="/wiki/Shivon_Zilis" title="Shivon Zilis">Shivon Zilis</a> (2019–2023)</li></ul> </div></td></tr></tbody></table><div></div></td></tr></tbody></table><div></div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Joint_venture" title="Joint venture">Joint ventures</a></th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Stargate_LLC" title="Stargate LLC">Stargate LLC</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Related</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Apple_Intelligence" title="Apple Intelligence">Apple Intelligence</a></li> <li><i><a href="/wiki/AI_Dungeon" title="AI Dungeon">AI Dungeon</a></i></li> <li><a href="/wiki/AutoGPT" title="AutoGPT">AutoGPT</a></li> <li>"<a href="/wiki/Deep_Learning_(South_Park)" title="Deep Learning (South Park)">Deep Learning</a>"</li> <li><a href="/wiki/LangChain" title="LangChain">LangChain</a></li> <li><a href="/wiki/Microsoft_Copilot" title="Microsoft Copilot">Microsoft Copilot</a></li> <li><a href="/wiki/OpenAI_Five" title="OpenAI Five">OpenAI Five</a></li> <li><a href="/wiki/Transformer_(deep_learning_architecture)" title="Transformer (deep learning architecture)">Transformer</a></li></ul> </div></td></tr><tr><td class="navbox-abovebelow" colspan="3"><div> <ul><li><span class="noviewer" typeof="mw:File"><span title="Category"><img alt="" src="//upload.wikimedia.org/wikipedia/en/thumb/9/96/Symbol_category_class.svg/16px-Symbol_category_class.svg.png" decoding="async" width="16" height="16" class="mw-file-element" srcset="//upload.wikimedia.org/wikipedia/en/thumb/9/96/Symbol_category_class.svg/23px-Symbol_category_class.svg.png 1.5x, //upload.wikimedia.org/wikipedia/en/thumb/9/96/Symbol_category_class.svg/31px-Symbol_category_class.svg.png 2x" data-file-width="180" data-file-height="185" /></span></span> <a href="/wiki/Category:OpenAI" title="Category:OpenAI">Category</a></li></ul> </div></td></tr></tbody></table></div> <div class="navbox-styles"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1129693374"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1236075235"></div><div role="navigation" class="navbox" aria-labelledby="Artificial_intelligence_(AI)776" style="padding:3px"><table class="nowraplinks hlist mw-collapsible autocollapse navbox-inner" style="border-spacing:0;background:transparent;color:inherit"><tbody><tr><th scope="col" class="navbox-title" colspan="2"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1129693374"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1239400231"><div class="navbar plainlinks hlist navbar-mini"><ul><li class="nv-view"><a href="/wiki/Template:Artificial_intelligence_navbox" title="Template:Artificial intelligence navbox"><abbr title="View this template">v</abbr></a></li><li class="nv-talk"><a href="/wiki/Template_talk:Artificial_intelligence_navbox" title="Template talk:Artificial intelligence navbox"><abbr title="Discuss this template">t</abbr></a></li><li class="nv-edit"><a href="/wiki/Special:EditPage/Template:Artificial_intelligence_navbox" title="Special:EditPage/Template:Artificial intelligence navbox"><abbr title="Edit this template">e</abbr></a></li></ul></div><div id="Artificial_intelligence_(AI)776" style="font-size:114%;margin:0 4em"><a href="/wiki/Artificial_intelligence" title="Artificial intelligence">Artificial intelligence (AI)</a></div></th></tr><tr><td class="navbox-abovebelow" colspan="2"><div><a href="/wiki/History_of_artificial_intelligence" title="History of artificial intelligence">History</a> (<a href="/wiki/Timeline_of_artificial_intelligence" title="Timeline of artificial intelligence">timeline</a>)</div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Concepts</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Parameter" title="Parameter">Parameter</a> <ul><li><a href="/wiki/Hyperparameter_(machine_learning)" title="Hyperparameter (machine learning)">Hyperparameter</a></li></ul></li> <li><a href="/wiki/Loss_functions_for_classification" title="Loss functions for classification">Loss functions</a></li> <li><a href="/wiki/Regression_analysis" title="Regression analysis">Regression</a> <ul><li><a href="/wiki/Bias%E2%80%93variance_tradeoff" title="Bias–variance tradeoff">Bias–variance tradeoff</a></li> <li><a href="/wiki/Double_descent" title="Double descent">Double descent</a></li> <li><a href="/wiki/Overfitting" title="Overfitting">Overfitting</a></li></ul></li> <li><a href="/wiki/Cluster_analysis" title="Cluster analysis">Clustering</a></li> <li><a href="/wiki/Gradient_descent" title="Gradient descent">Gradient descent</a> <ul><li><a href="/wiki/Stochastic_gradient_descent" title="Stochastic gradient descent">SGD</a></li> <li><a href="/wiki/Quasi-Newton_method" title="Quasi-Newton method">Quasi-Newton method</a></li> <li><a href="/wiki/Conjugate_gradient_method" title="Conjugate gradient method">Conjugate gradient method</a></li></ul></li> <li><a href="/wiki/Backpropagation" title="Backpropagation">Backpropagation</a></li> <li><a href="/wiki/Attention_(machine_learning)" title="Attention (machine learning)">Attention</a></li> <li><a href="/wiki/Convolution" title="Convolution">Convolution</a></li> <li><a href="/wiki/Normalization_(machine_learning)" title="Normalization (machine learning)">Normalization</a> <ul><li><a href="/wiki/Batch_normalization" title="Batch normalization">Batchnorm</a></li></ul></li> <li><a href="/wiki/Activation_function" title="Activation function">Activation</a> <ul><li><a href="/wiki/Softmax_function" title="Softmax function">Softmax</a></li> <li><a href="/wiki/Sigmoid_function" title="Sigmoid function">Sigmoid</a></li> <li><a href="/wiki/Rectifier_(neural_networks)" title="Rectifier (neural networks)">Rectifier</a></li></ul></li> <li><a href="/wiki/Gating_mechanism" title="Gating mechanism">Gating</a></li> <li><a href="/wiki/Weight_initialization" title="Weight initialization">Weight initialization</a></li> <li><a href="/wiki/Regularization_(mathematics)" title="Regularization (mathematics)">Regularization</a></li> <li><a href="/wiki/Training,_validation,_and_test_data_sets" title="Training, validation, and test data sets">Datasets</a> <ul><li><a href="/wiki/Data_augmentation" title="Data augmentation">Augmentation</a></li></ul></li> <li><a href="/wiki/Prompt_engineering" title="Prompt engineering">Prompt engineering</a></li> <li><a href="/wiki/Reinforcement_learning" title="Reinforcement learning">Reinforcement learning</a> <ul><li><a href="/wiki/Q-learning" title="Q-learning">Q-learning</a></li> <li><a href="/wiki/State%E2%80%93action%E2%80%93reward%E2%80%93state%E2%80%93action" title="State–action–reward–state–action">SARSA</a></li> <li><a href="/wiki/Imitation_learning" title="Imitation learning">Imitation</a></li> <li><a href="/wiki/Policy_gradient_method" title="Policy gradient method">Policy gradient</a></li></ul></li> <li><a href="/wiki/Diffusion_process" title="Diffusion process">Diffusion</a></li> <li><a href="/wiki/Latent_diffusion_model" title="Latent diffusion model">Latent diffusion model</a></li> <li><a href="/wiki/Autoregressive_model" title="Autoregressive model">Autoregression</a></li> <li><a href="/wiki/Adversarial_machine_learning" title="Adversarial machine learning">Adversary</a></li> <li><a href="/wiki/Retrieval-augmented_generation" title="Retrieval-augmented generation">RAG</a></li> <li><a href="/wiki/Uncanny_valley" title="Uncanny valley">Uncanny valley</a></li> <li><a href="/wiki/Reinforcement_learning_from_human_feedback" title="Reinforcement learning from human feedback">RLHF</a></li> <li><a href="/wiki/Self-supervised_learning" title="Self-supervised learning">Self-supervised learning</a></li> <li><a href="/wiki/Recursive_self-improvement" title="Recursive self-improvement">Recursive self-improvement</a></li> <li><a href="/wiki/Word_embedding" title="Word embedding">Word embedding</a></li> <li><a href="/wiki/Hallucination_(artificial_intelligence)" title="Hallucination (artificial intelligence)">Hallucination</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Applications</th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Machine_learning" title="Machine learning">Machine learning</a> <ul><li><a href="/wiki/Prompt_engineering#In-context_learning" title="Prompt engineering">In-context learning</a></li></ul></li> <li><a href="/wiki/Neural_network_(machine_learning)" title="Neural network (machine learning)">Artificial neural network</a> <ul><li><a href="/wiki/Deep_learning" title="Deep learning">Deep learning</a></li></ul></li> <li><a href="/wiki/Language_model" title="Language model">Language model</a> <ul><li><a href="/wiki/Large_language_model" title="Large language model">Large language model</a></li> <li><a href="/wiki/Neural_machine_translation" title="Neural machine translation">NMT</a></li></ul></li> <li><a href="/wiki/Artificial_general_intelligence" title="Artificial general intelligence">Artificial general intelligence</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Implementations</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"></div><table class="nowraplinks navbox-subgroup" style="border-spacing:0"><tbody><tr><th scope="row" class="navbox-group" style="width:1%">Audio–visual</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/AlexNet" title="AlexNet">AlexNet</a></li> <li><a href="/wiki/WaveNet" title="WaveNet">WaveNet</a></li> <li><a href="/wiki/Human_image_synthesis" title="Human image synthesis">Human image synthesis</a></li> <li><a href="/wiki/Handwriting_recognition" title="Handwriting recognition">HWR</a></li> <li><a href="/wiki/Optical_character_recognition" title="Optical character recognition">OCR</a></li> <li><a href="/wiki/Deep_learning_speech_synthesis" title="Deep learning speech synthesis">Speech synthesis</a> <ul><li><a href="/wiki/15.ai" title="15.ai">15.ai</a></li> <li><a href="/wiki/ElevenLabs" title="ElevenLabs">ElevenLabs</a></li></ul></li> <li><a href="/wiki/Speech_recognition" title="Speech recognition">Speech recognition</a> <ul><li><a href="/wiki/Whisper_(speech_recognition_system)" title="Whisper (speech recognition system)">Whisper</a></li></ul></li> <li><a href="/wiki/Facial_recognition_system" title="Facial recognition system">Facial recognition</a></li> <li><a href="/wiki/AlphaFold" title="AlphaFold">AlphaFold</a></li> <li><a href="/wiki/Text-to-image_model" title="Text-to-image model">Text-to-image models</a> <ul><li><a href="/wiki/Aurora_(text-to-image_model)" class="mw-redirect" title="Aurora (text-to-image model)">Aurora</a></li> <li><a href="/wiki/DALL-E" title="DALL-E">DALL-E</a></li> <li><a href="/wiki/Adobe_Firefly" title="Adobe Firefly">Firefly</a></li> <li><a href="/wiki/Flux_(text-to-image_model)" title="Flux (text-to-image model)">Flux</a></li> <li><a href="/wiki/Ideogram_(text-to-image_model)" title="Ideogram (text-to-image model)">Ideogram</a></li> <li><a href="/wiki/Google_Brain#Text-to-image_model" title="Google Brain">Imagen</a></li> <li><a href="/wiki/Midjourney" title="Midjourney">Midjourney</a></li> <li><a href="/wiki/Stable_Diffusion" title="Stable Diffusion">Stable Diffusion</a></li></ul></li> <li><a href="/wiki/Text-to-video_model" title="Text-to-video model">Text-to-video models</a> <ul><li><a href="/wiki/Dream_Machine_(text-to-video_model)" title="Dream Machine (text-to-video model)">Dream Machine</a></li> <li><a href="/wiki/Runway_(company)#Gen-3_Alpha" title="Runway (company)">Gen-3 Alpha</a></li> <li><a href="/wiki/MiniMax_(company)#Hailuo_AI" title="MiniMax (company)">Hailuo AI</a></li> <li><a href="/wiki/Kling_(text-to-video_model)" class="mw-redirect" title="Kling (text-to-video model)">Kling</a></li> <li><a href="/wiki/Sora_(text-to-video_model)" title="Sora (text-to-video model)">Sora</a></li> <li><a href="/wiki/Google_DeepMind#Video_model" title="Google DeepMind">Veo</a></li></ul></li> <li><a href="/wiki/Music_and_artificial_intelligence" title="Music and artificial intelligence">Music generation</a> <ul><li><a href="/wiki/Suno_AI" title="Suno AI">Suno AI</a></li> <li><a href="/wiki/Udio" title="Udio">Udio</a></li></ul></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Text</th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Word2vec" title="Word2vec">Word2vec</a></li> <li><a href="/wiki/Seq2seq" title="Seq2seq">Seq2seq</a></li> <li><a href="/wiki/GloVe" title="GloVe">GloVe</a></li> <li><a href="/wiki/BERT_(language_model)" title="BERT (language model)">BERT</a></li> <li><a href="/wiki/T5_(language_model)" title="T5 (language model)">T5</a></li> <li><a href="/wiki/Llama_(language_model)" title="Llama (language model)">Llama</a></li> <li><a href="/wiki/Chinchilla_(language_model)" title="Chinchilla (language model)">Chinchilla AI</a></li> <li><a href="/wiki/PaLM" title="PaLM">PaLM</a></li> <li><a class="mw-selflink selflink">GPT</a> <ul><li><a href="/wiki/GPT-1" title="GPT-1">1</a></li> <li><a href="/wiki/GPT-2" title="GPT-2">2</a></li> <li><a href="/wiki/GPT-3" title="GPT-3">3</a></li> <li><a href="/wiki/GPT-J" title="GPT-J">J</a></li> <li><a href="/wiki/ChatGPT" title="ChatGPT">ChatGPT</a></li> <li><a href="/wiki/GPT-4" title="GPT-4">4</a></li> <li><a href="/wiki/GPT-4o" title="GPT-4o">4o</a></li> <li><a href="/wiki/OpenAI_o1" title="OpenAI o1">o1</a></li> <li><a href="/wiki/OpenAI_o3" title="OpenAI o3">o3</a></li></ul></li> <li><a href="/wiki/Claude_(language_model)" title="Claude (language model)">Claude</a></li> <li><a href="/wiki/Gemini_(language_model)" title="Gemini (language model)">Gemini</a> <ul><li><a href="/wiki/Gemini_(chatbot)" title="Gemini (chatbot)">chatbot</a></li></ul></li> <li><a href="/wiki/Grok_(chatbot)" title="Grok (chatbot)">Grok</a></li> <li><a href="/wiki/LaMDA" title="LaMDA">LaMDA</a></li> <li><a href="/wiki/BLOOM_(language_model)" title="BLOOM (language model)">BLOOM</a></li> <li><a href="/wiki/Project_Debater" title="Project Debater">Project Debater</a></li> <li><a href="/wiki/IBM_Watson" title="IBM Watson">IBM Watson</a></li> <li><a href="/wiki/IBM_Watsonx" title="IBM Watsonx">IBM Watsonx</a></li> <li><a href="/wiki/IBM_Granite" title="IBM Granite">Granite</a></li> <li><a href="/wiki/Huawei_PanGu" title="Huawei PanGu">PanGu-Σ</a></li> <li><a href="/wiki/DeepSeek_(chatbot)" title="DeepSeek (chatbot)">DeepSeek</a></li> <li><a href="/wiki/Qwen" title="Qwen">Qwen</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Decisional</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/AlphaGo" title="AlphaGo">AlphaGo</a></li> <li><a href="/wiki/AlphaZero" title="AlphaZero">AlphaZero</a></li> <li><a href="/wiki/OpenAI_Five" title="OpenAI Five">OpenAI Five</a></li> <li><a href="/wiki/Self-driving_car" title="Self-driving car">Self-driving car</a></li> <li><a href="/wiki/MuZero" title="MuZero">MuZero</a></li> <li><a href="/wiki/Action_selection" title="Action selection">Action selection</a> <ul><li><a href="/wiki/AutoGPT" title="AutoGPT">AutoGPT</a></li></ul></li> <li><a href="/wiki/Robot_control" title="Robot control">Robot control</a></li></ul> </div></td></tr></tbody></table><div></div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">People</th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Alan_Turing" title="Alan Turing">Alan Turing</a></li> <li><a href="/wiki/Warren_Sturgis_McCulloch" title="Warren Sturgis McCulloch">Warren Sturgis McCulloch</a></li> <li><a href="/wiki/Walter_Pitts" title="Walter Pitts">Walter Pitts</a></li> <li><a href="/wiki/John_von_Neumann" title="John von Neumann">John von Neumann</a></li> <li><a href="/wiki/Claude_Shannon" title="Claude Shannon">Claude Shannon</a></li> <li><a href="/wiki/Marvin_Minsky" title="Marvin Minsky">Marvin Minsky</a></li> <li><a href="/wiki/John_McCarthy_(computer_scientist)" title="John McCarthy (computer scientist)">John McCarthy</a></li> <li><a href="/wiki/Nathaniel_Rochester_(computer_scientist)" title="Nathaniel Rochester (computer scientist)">Nathaniel Rochester</a></li> <li><a href="/wiki/Allen_Newell" title="Allen Newell">Allen Newell</a></li> <li><a href="/wiki/Cliff_Shaw" title="Cliff Shaw">Cliff Shaw</a></li> <li><a href="/wiki/Herbert_A._Simon" title="Herbert A. Simon">Herbert A. Simon</a></li> <li><a href="/wiki/Oliver_Selfridge" title="Oliver Selfridge">Oliver Selfridge</a></li> <li><a href="/wiki/Frank_Rosenblatt" title="Frank Rosenblatt">Frank Rosenblatt</a></li> <li><a href="/wiki/Bernard_Widrow" title="Bernard Widrow">Bernard Widrow</a></li> <li><a href="/wiki/Joseph_Weizenbaum" title="Joseph Weizenbaum">Joseph Weizenbaum</a></li> <li><a href="/wiki/Seymour_Papert" title="Seymour Papert">Seymour Papert</a></li> <li><a href="/wiki/Seppo_Linnainmaa" title="Seppo Linnainmaa">Seppo Linnainmaa</a></li> <li><a href="/wiki/Paul_Werbos" title="Paul Werbos">Paul Werbos</a></li> <li><a href="/wiki/J%C3%BCrgen_Schmidhuber" title="Jürgen Schmidhuber">Jürgen Schmidhuber</a></li> <li><a href="/wiki/Yann_LeCun" title="Yann LeCun">Yann LeCun</a></li> <li><a href="/wiki/Geoffrey_Hinton" title="Geoffrey Hinton">Geoffrey Hinton</a></li> <li><a href="/wiki/John_Hopfield" title="John Hopfield">John Hopfield</a></li> <li><a href="/wiki/Yoshua_Bengio" title="Yoshua Bengio">Yoshua Bengio</a></li> <li><a href="/wiki/Lotfi_A._Zadeh" title="Lotfi A. Zadeh">Lotfi A. Zadeh</a></li> <li><a href="/wiki/Stephen_Grossberg" title="Stephen Grossberg">Stephen Grossberg</a></li> <li><a href="/wiki/Alex_Graves_(computer_scientist)" title="Alex Graves (computer scientist)">Alex Graves</a></li> <li><a href="/wiki/Andrew_Ng" title="Andrew Ng">Andrew Ng</a></li> <li><a href="/wiki/Fei-Fei_Li" title="Fei-Fei Li">Fei-Fei Li</a></li> <li><a href="/wiki/Alex_Krizhevsky" title="Alex Krizhevsky">Alex Krizhevsky</a></li> <li><a href="/wiki/Ilya_Sutskever" title="Ilya Sutskever">Ilya Sutskever</a></li> <li><a href="/wiki/Demis_Hassabis" title="Demis Hassabis">Demis Hassabis</a></li> <li><a href="/wiki/David_Silver_(computer_scientist)" title="David Silver (computer scientist)">David Silver</a></li> <li><a href="/wiki/Ian_Goodfellow" title="Ian Goodfellow">Ian Goodfellow</a></li> <li><a href="/wiki/Andrej_Karpathy" title="Andrej Karpathy">Andrej Karpathy</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Architectures</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Neural_Turing_machine" title="Neural Turing machine">Neural Turing machine</a></li> <li><a href="/wiki/Differentiable_neural_computer" title="Differentiable neural computer">Differentiable neural computer</a></li> <li><a href="/wiki/Transformer_(deep_learning_architecture)" title="Transformer (deep learning architecture)">Transformer</a> <ul><li><a href="/wiki/Vision_transformer" title="Vision transformer">Vision transformer (ViT)</a></li></ul></li> <li><a href="/wiki/Recurrent_neural_network" title="Recurrent neural network">Recurrent neural network (RNN)</a></li> <li><a href="/wiki/Long_short-term_memory" title="Long short-term memory">Long short-term memory (LSTM)</a></li> <li><a href="/wiki/Gated_recurrent_unit" title="Gated recurrent unit">Gated recurrent unit (GRU)</a></li> <li><a href="/wiki/Echo_state_network" title="Echo state network">Echo state network</a></li> <li><a href="/wiki/Multilayer_perceptron" title="Multilayer perceptron">Multilayer perceptron (MLP)</a></li> <li><a href="/wiki/Convolutional_neural_network" title="Convolutional neural network">Convolutional neural network (CNN)</a></li> <li><a href="/wiki/Residual_neural_network" title="Residual neural network">Residual neural network (RNN)</a></li> <li><a href="/wiki/Highway_network" title="Highway network">Highway network</a></li> <li><a href="/wiki/Mamba_(deep_learning_architecture)" title="Mamba (deep learning architecture)">Mamba</a></li> <li><a href="/wiki/Autoencoder" title="Autoencoder">Autoencoder</a></li> <li><a href="/wiki/Variational_autoencoder" title="Variational autoencoder">Variational autoencoder (VAE)</a></li> <li><a href="/wiki/Generative_adversarial_network" title="Generative adversarial network">Generative adversarial network (GAN)</a></li> <li><a href="/wiki/Graph_neural_network" title="Graph neural network">Graph neural network (GNN)</a></li></ul> </div></td></tr><tr><td class="navbox-abovebelow" colspan="2"><div> <ul><li><span class="noviewer" typeof="mw:File"><a href="/wiki/File:Symbol_portal_class.svg" class="mw-file-description" title="Portal"><img alt="" src="//upload.wikimedia.org/wikipedia/en/thumb/e/e2/Symbol_portal_class.svg/16px-Symbol_portal_class.svg.png" decoding="async" width="16" height="16" class="mw-file-element" srcset="//upload.wikimedia.org/wikipedia/en/thumb/e/e2/Symbol_portal_class.svg/23px-Symbol_portal_class.svg.png 1.5x, //upload.wikimedia.org/wikipedia/en/thumb/e/e2/Symbol_portal_class.svg/31px-Symbol_portal_class.svg.png 2x" data-file-width="180" data-file-height="185" /></a></span> Portals <ul><li><a href="/wiki/Portal:Technology" title="Portal:Technology">Technology</a></li></ul></li> <li><span class="noviewer" typeof="mw:File"><span title="Category"><img alt="" src="//upload.wikimedia.org/wikipedia/en/thumb/9/96/Symbol_category_class.svg/16px-Symbol_category_class.svg.png" decoding="async" width="16" height="16" class="mw-file-element" srcset="//upload.wikimedia.org/wikipedia/en/thumb/9/96/Symbol_category_class.svg/23px-Symbol_category_class.svg.png 1.5x, //upload.wikimedia.org/wikipedia/en/thumb/9/96/Symbol_category_class.svg/31px-Symbol_category_class.svg.png 2x" data-file-width="180" data-file-height="185" /></span></span> <a href="/wiki/Category:Artificial_intelligence" title="Category:Artificial intelligence">Category</a> <ul><li><a href="/wiki/Category:Artificial_neural_networks" title="Category:Artificial neural networks">Artificial neural networks</a></li> <li><a href="/wiki/Category:Machine_learning" title="Category:Machine learning">Machine learning</a></li></ul></li> <li><span class="noviewer" typeof="mw:File"><span title="List-Class article"><img alt="" src="//upload.wikimedia.org/wikipedia/en/thumb/d/db/Symbol_list_class.svg/16px-Symbol_list_class.svg.png" decoding="async" width="16" height="16" class="mw-file-element" srcset="//upload.wikimedia.org/wikipedia/en/thumb/d/db/Symbol_list_class.svg/23px-Symbol_list_class.svg.png 1.5x, //upload.wikimedia.org/wikipedia/en/thumb/d/db/Symbol_list_class.svg/31px-Symbol_list_class.svg.png 2x" data-file-width="180" data-file-height="185" /></span></span> List <ul><li><a href="/wiki/List_of_artificial_intelligence_companies" title="List of artificial intelligence companies">Companies</a></li> <li><a href="/wiki/List_of_artificial_intelligence_projects" title="List of artificial intelligence projects">Projects</a></li></ul></li></ul> </div></td></tr></tbody></table></div> <div class="navbox-styles"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1129693374"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1236075235"></div><div role="navigation" class="navbox" aria-labelledby="Generative_AI409" style="padding:3px"><table class="nowraplinks hlist mw-collapsible autocollapse navbox-inner" style="border-spacing:0;background:transparent;color:inherit"><tbody><tr><th scope="col" class="navbox-title" colspan="2"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1129693374"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1239400231"><div class="navbar plainlinks hlist navbar-mini"><ul><li class="nv-view"><a href="/wiki/Template:Generative_AI" title="Template:Generative AI"><abbr title="View this template">v</abbr></a></li><li class="nv-talk"><a href="/wiki/Template_talk:Generative_AI" title="Template talk:Generative AI"><abbr title="Discuss this template">t</abbr></a></li><li class="nv-edit"><a href="/wiki/Special:EditPage/Template:Generative_AI" title="Special:EditPage/Template:Generative AI"><abbr title="Edit this template">e</abbr></a></li></ul></div><div id="Generative_AI409" style="font-size:114%;margin:0 4em"><a href="/wiki/Generative_artificial_intelligence" title="Generative artificial intelligence">Generative AI</a></div></th></tr><tr><th scope="row" class="navbox-group" style="width:1%">Concepts</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Autoencoder" title="Autoencoder">Autoencoder</a></li> <li><a href="/wiki/Deep_learning" title="Deep learning">Deep learning</a></li> <li><a href="/wiki/Generative_adversarial_network" title="Generative adversarial network">Generative adversarial network</a></li> <li><a class="mw-selflink selflink">Generative pre-trained transformer</a></li> <li><a href="/wiki/Large_language_model" title="Large language model">Large language model</a></li> <li><a href="/wiki/Neural_network_(machine_learning)" title="Neural network (machine learning)">Neural network</a></li> <li><a href="/wiki/Prompt_engineering" title="Prompt engineering">Prompt engineering</a></li> <li><a href="/wiki/Retrieval-augmented_generation" title="Retrieval-augmented generation">Retrieval-augmented generation</a></li> <li><a href="/wiki/Reinforcement_learning_from_human_feedback" title="Reinforcement learning from human feedback">Reinforcement learning from human feedback</a></li> <li><a href="/wiki/Self-supervised_learning" title="Self-supervised learning">Self-supervised learning</a></li> <li><a href="/wiki/Transformer_(deep_learning_architecture)" title="Transformer (deep learning architecture)">Transformer</a></li> <li><a href="/wiki/Variational_autoencoder" title="Variational autoencoder">Variational autoencoder</a></li> <li><a href="/wiki/Vision_transformer" title="Vision transformer">Vision transformer</a></li> <li><a href="/wiki/Word_embedding" title="Word embedding">Word embedding</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Models</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"></div><table class="nowraplinks navbox-subgroup" style="border-spacing:0"><tbody><tr><th scope="row" class="navbox-group" style="width:1%">Text</th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Claude_(language_model)" title="Claude (language model)">Claude</a></li> <li><a href="/wiki/DBRX" title="DBRX">DBRX</a></li> <li><a href="/wiki/DeepSeek_(chatbot)" title="DeepSeek (chatbot)">DeepSeek</a></li> <li><a href="/wiki/Gemini_(chatbot)" title="Gemini (chatbot)">Gemini</a></li> <li><a class="mw-selflink selflink">GPT</a> <ul><li><a href="/wiki/GPT-1" title="GPT-1">1</a></li> <li><a href="/wiki/GPT-2" title="GPT-2">2</a></li> <li><a href="/wiki/GPT-3" title="GPT-3">3</a></li> <li><a href="/wiki/GPT-J" title="GPT-J">J</a></li> <li><a href="/wiki/ChatGPT" title="ChatGPT">ChatGPT</a></li> <li><a href="/wiki/GPT-4" title="GPT-4">4</a></li> <li><a href="/wiki/GPT-4o" title="GPT-4o">4o</a></li> <li><a href="/wiki/OpenAI_o1" title="OpenAI o1">o1</a></li> <li><a href="/wiki/OpenAI_o3" title="OpenAI o3">o3</a></li></ul></li> <li><a href="/wiki/Grok_(chatbot)" title="Grok (chatbot)">Grok</a></li> <li><a href="/wiki/IBM_Granite" title="IBM Granite">Granite</a></li> <li><a href="/wiki/Llama_(language_model)" title="Llama (language model)">Llama</a></li> <li><a href="/wiki/Mistral_AI#Mistral_Large" title="Mistral AI">Mistral Large</a></li> <li><a href="/wiki/Huawei_PanGu" title="Huawei PanGu">PanGu-Σ</a></li> <li><a href="/wiki/Qwen" title="Qwen">Qwen</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Text-to-image_model" title="Text-to-image model">Image</a></th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Aurora_(text-to-image_model)" class="mw-redirect" title="Aurora (text-to-image model)">Aurora</a></li> <li><a href="/wiki/DALL-E" title="DALL-E">DALL-E</a></li> <li><a href="/wiki/Adobe_Firefly" title="Adobe Firefly">Firefly</a></li> <li><a href="/wiki/Flux_(text-to-image_model)" title="Flux (text-to-image model)">Flux</a></li> <li><a href="/wiki/Ideogram_(text-to-image_model)" title="Ideogram (text-to-image model)">Ideogram</a></li> <li><a href="/wiki/Midjourney" title="Midjourney">Midjourney</a></li> <li><a href="/wiki/Stable_Diffusion" title="Stable Diffusion">Stable Diffusion</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Speech</th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/15.ai" title="15.ai">15.ai</a></li> <li><a href="/wiki/WaveNet" title="WaveNet">WaveNet</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Text-to-video_model" title="Text-to-video model">Video</a></th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Dream_Machine_(text-to-video_model)" title="Dream Machine (text-to-video model)">Dream Machine</a></li> <li><a href="/wiki/Runway_(company)#Gen-3_Alpha" title="Runway (company)">Gen-3 Alpha</a></li> <li><a href="/wiki/MiniMax_(company)#Hailuo_AI" title="MiniMax (company)">Hailuo AI</a></li> <li><a href="/wiki/Kling_(text-to-video_model)" class="mw-redirect" title="Kling (text-to-video model)">Kling</a></li> <li><a href="/wiki/Sora_(text-to-video_model)" title="Sora (text-to-video model)">Sora</a></li> <li><a href="/wiki/Google_DeepMind#Video_model" title="Google DeepMind">Veo</a></li> <li><a href="/wiki/VideoPoet" title="VideoPoet">VideoPoet</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Music</th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Udio" title="Udio">Udio</a></li> <li><a href="/wiki/Suno_AI" title="Suno AI">Suno AI</a></li></ul> </div></td></tr></tbody></table><div></div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/List_of_artificial_intelligence_companies" title="List of artificial intelligence companies">Companies</a></th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/01.AI" title="01.AI">01.AI</a></li> <li><a href="/wiki/Alibaba_Group" title="Alibaba Group">Alibaba</a></li> <li><a href="/wiki/Anthropic" title="Anthropic">Anthropic</a></li> <li><a href="/wiki/Baichuan" title="Baichuan">Baichuan</a></li> <li><a href="/wiki/DeepSeek" title="DeepSeek">DeepSeek</a></li> <li><a href="/wiki/ElevenLabs" title="ElevenLabs">ElevenLabs</a></li> <li><a href="/wiki/Google_DeepMind" title="Google DeepMind">Google DeepMind</a></li> <li><a href="/wiki/Hugging_Face" title="Hugging Face">Hugging Face</a></li> <li><a href="/wiki/Kuaishou" title="Kuaishou">Kuaishou</a></li> <li><a href="/wiki/Meta_AI" title="Meta AI">Meta AI</a></li> <li><a href="/wiki/MiniMax_(company)" title="MiniMax (company)">MiniMax</a></li> <li><a href="/wiki/Mistral_AI" title="Mistral AI">Mistral AI</a></li> <li><a href="/wiki/Moonshot_AI" title="Moonshot AI">Moonshot AI</a></li> <li><a href="/wiki/OpenAI" title="OpenAI">OpenAI</a></li> <li><a href="/wiki/Runway_(company)" title="Runway (company)">Runway</a></li> <li><a href="/wiki/Stability_AI" title="Stability AI">Stability AI</a></li> <li><a href="/wiki/Synthesia_(company)" title="Synthesia (company)">Synthesia</a></li> <li><a href="/wiki/XAI_(company)" title="XAI (company)">xAI</a></li> <li><a href="/wiki/Zhipu_AI" title="Zhipu AI">Zhipu AI</a></li></ul> </div></td></tr><tr><td class="navbox-abovebelow" colspan="2"><div> <ul><li><span class="noviewer" typeof="mw:File"><span title="Category"><img alt="" src="//upload.wikimedia.org/wikipedia/en/thumb/9/96/Symbol_category_class.svg/16px-Symbol_category_class.svg.png" decoding="async" width="16" height="16" class="mw-file-element" srcset="//upload.wikimedia.org/wikipedia/en/thumb/9/96/Symbol_category_class.svg/23px-Symbol_category_class.svg.png 1.5x, //upload.wikimedia.org/wikipedia/en/thumb/9/96/Symbol_category_class.svg/31px-Symbol_category_class.svg.png 2x" data-file-width="180" data-file-height="185" /></span></span> <b><a href="/wiki/Category:Generative_artificial_intelligence" title="Category:Generative artificial intelligence">Category</a></b></li> <li><span class="noviewer" typeof="mw:File"><span title="Commons page"><img alt="" src="//upload.wikimedia.org/wikipedia/en/thumb/4/4a/Commons-logo.svg/12px-Commons-logo.svg.png" decoding="async" width="12" height="16" class="mw-file-element" srcset="//upload.wikimedia.org/wikipedia/en/thumb/4/4a/Commons-logo.svg/18px-Commons-logo.svg.png 1.5x, //upload.wikimedia.org/wikipedia/en/thumb/4/4a/Commons-logo.svg/24px-Commons-logo.svg.png 2x" data-file-width="1024" data-file-height="1376" /></span></span> <b><a href="https://commons.wikimedia.org/wiki/Category:Generative_artificial_intelligence" class="extiw" title="commons:Category:Generative artificial intelligence">Commons</a></b></li></ul> </div></td></tr></tbody></table></div> <div class="navbox-styles"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1129693374"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1236075235"></div><div role="navigation" class="navbox" aria-labelledby="Natural_language_processing454" style="padding:3px"><table class="nowraplinks hlist mw-collapsible autocollapse navbox-inner" style="border-spacing:0;background:transparent;color:inherit"><tbody><tr><th scope="col" class="navbox-title" colspan="2"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1129693374"><link rel="mw-deduplicated-inline-style" href="mw-data:TemplateStyles:r1239400231"><div class="navbar plainlinks hlist navbar-mini"><ul><li class="nv-view"><a href="/wiki/Template:Natural_language_processing" title="Template:Natural language processing"><abbr title="View this template">v</abbr></a></li><li class="nv-talk"><a href="/wiki/Template_talk:Natural_language_processing" title="Template talk:Natural language processing"><abbr title="Discuss this template">t</abbr></a></li><li class="nv-edit"><a href="/wiki/Special:EditPage/Template:Natural_language_processing" title="Special:EditPage/Template:Natural language processing"><abbr title="Edit this template">e</abbr></a></li></ul></div><div id="Natural_language_processing454" style="font-size:114%;margin:0 4em"><a href="/wiki/Natural_language_processing" title="Natural language processing">Natural language processing</a></div></th></tr><tr><th scope="row" class="navbox-group" style="width:1%">General terms</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/AI-complete" title="AI-complete">AI-complete</a></li> <li><a href="/wiki/Bag-of-words_model" title="Bag-of-words model">Bag-of-words</a></li> <li><a href="/wiki/N-gram" title="N-gram">n-gram</a> <ul><li><a href="/wiki/Bigram" title="Bigram">Bigram</a></li> <li><a href="/wiki/Trigram" title="Trigram">Trigram</a></li></ul></li> <li><a href="/wiki/Computational_linguistics" title="Computational linguistics">Computational linguistics</a></li> <li><a href="/wiki/Natural_language_understanding" title="Natural language understanding">Natural language understanding</a></li> <li><a href="/wiki/Stop_word" title="Stop word">Stop words</a></li> <li><a href="/wiki/Text_processing" title="Text processing">Text processing</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Text_mining" title="Text mining">Text analysis</a></th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Argument_mining" title="Argument mining">Argument mining</a></li> <li><a href="/wiki/Collocation_extraction" title="Collocation extraction">Collocation extraction</a></li> <li><a href="/wiki/Concept_mining" title="Concept mining">Concept mining</a></li> <li><a href="/wiki/Coreference#Coreference_resolution" title="Coreference">Coreference resolution</a></li> <li><a href="/wiki/Deep_linguistic_processing" title="Deep linguistic processing">Deep linguistic processing</a></li> <li><a href="/wiki/Distant_reading" title="Distant reading">Distant reading</a></li> <li><a href="/wiki/Information_extraction" title="Information extraction">Information extraction</a></li> <li><a href="/wiki/Named-entity_recognition" title="Named-entity recognition">Named-entity recognition</a></li> <li><a href="/wiki/Ontology_learning" title="Ontology learning">Ontology learning</a></li> <li><a href="/wiki/Parsing" title="Parsing">Parsing</a> <ul><li><a href="/wiki/Semantic_parsing" title="Semantic parsing">Semantic parsing</a></li> <li><a href="/wiki/Syntactic_parsing_(computational_linguistics)" title="Syntactic parsing (computational linguistics)">Syntactic parsing</a></li></ul></li> <li><a href="/wiki/Part-of-speech_tagging" title="Part-of-speech tagging">Part-of-speech tagging</a></li> <li><a href="/wiki/Semantic_analysis_(machine_learning)" title="Semantic analysis (machine learning)">Semantic analysis</a></li> <li><a href="/wiki/Semantic_role_labeling" title="Semantic role labeling">Semantic role labeling</a></li> <li><a href="/wiki/Semantic_decomposition_(natural_language_processing)" title="Semantic decomposition (natural language processing)">Semantic decomposition</a></li> <li><a href="/wiki/Semantic_similarity" title="Semantic similarity">Semantic similarity</a></li> <li><a href="/wiki/Sentiment_analysis" title="Sentiment analysis">Sentiment analysis</a></li></ul> <ul><li><a href="/wiki/Terminology_extraction" title="Terminology extraction">Terminology extraction</a></li> <li><a href="/wiki/Text_mining" title="Text mining">Text mining</a></li> <li><a href="/wiki/Textual_entailment" title="Textual entailment">Textual entailment</a></li> <li><a href="/wiki/Truecasing" title="Truecasing">Truecasing</a></li> <li><a href="/wiki/Word-sense_disambiguation" title="Word-sense disambiguation">Word-sense disambiguation</a></li> <li><a href="/wiki/Word-sense_induction" title="Word-sense induction">Word-sense induction</a></li></ul> </div><table class="nowraplinks navbox-subgroup" style="border-spacing:0"><tbody><tr><th id="Text_segmentation21" scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Text_segmentation" title="Text segmentation">Text segmentation</a></th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Compound-term_processing" title="Compound-term processing">Compound-term processing</a></li> <li><a href="/wiki/Lemmatisation" class="mw-redirect" title="Lemmatisation">Lemmatisation</a></li> <li><a href="/wiki/Lexical_analysis" title="Lexical analysis">Lexical analysis</a></li> <li><a href="/wiki/Shallow_parsing" title="Shallow parsing">Text chunking</a></li> <li><a href="/wiki/Stemming" title="Stemming">Stemming</a></li> <li><a href="/wiki/Sentence_boundary_disambiguation" title="Sentence boundary disambiguation">Sentence segmentation</a></li> <li><a href="/wiki/Word#Word_boundaries" title="Word">Word segmentation</a></li></ul> </div></td></tr></tbody></table><div> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Automatic_summarization" title="Automatic summarization">Automatic summarization</a></th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Multi-document_summarization" title="Multi-document summarization">Multi-document summarization</a></li> <li><a href="/wiki/Sentence_extraction" title="Sentence extraction">Sentence extraction</a></li> <li><a href="/wiki/Text_simplification" title="Text simplification">Text simplification</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Machine_translation" title="Machine translation">Machine translation</a></th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Computer-assisted_translation" title="Computer-assisted translation">Computer-assisted</a></li> <li><a href="/wiki/Example-based_machine_translation" title="Example-based machine translation">Example-based</a></li> <li><a href="/wiki/Rule-based_machine_translation" title="Rule-based machine translation">Rule-based</a></li> <li><a href="/wiki/Statistical_machine_translation" title="Statistical machine translation">Statistical</a></li> <li><a href="/wiki/Transfer-based_machine_translation" title="Transfer-based machine translation">Transfer-based</a></li> <li><a href="/wiki/Neural_machine_translation" title="Neural machine translation">Neural</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Distributional_semantics" title="Distributional semantics">Distributional semantics</a> models</th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/BERT_(language_model)" title="BERT (language model)">BERT</a></li> <li><a href="/wiki/Document-term_matrix" title="Document-term matrix">Document-term matrix</a></li> <li><a href="/wiki/Explicit_semantic_analysis" title="Explicit semantic analysis">Explicit semantic analysis</a></li> <li><a href="/wiki/FastText" title="FastText">fastText</a></li> <li><a href="/wiki/GloVe" title="GloVe">GloVe</a></li> <li><a href="/wiki/Language_model" title="Language model">Language model</a> (<a href="/wiki/Large_language_model" title="Large language model">large</a>)</li> <li><a href="/wiki/Latent_semantic_analysis" title="Latent semantic analysis">Latent semantic analysis</a></li> <li><a href="/wiki/Seq2seq" title="Seq2seq">Seq2seq</a></li> <li><a href="/wiki/Word_embedding" title="Word embedding">Word embedding</a></li> <li><a href="/wiki/Word2vec" title="Word2vec">Word2vec</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Language_resource" title="Language resource">Language resources</a>,<br />datasets and corpora</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"></div><table class="nowraplinks navbox-subgroup" style="border-spacing:0"><tbody><tr><th scope="row" class="navbox-group" style="width:1%">Types and<br />standards</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Corpus_linguistics" title="Corpus linguistics">Corpus linguistics</a></li> <li><a href="/wiki/Lexical_resource" title="Lexical resource">Lexical resource</a></li> <li><a href="/wiki/Linguistic_Linked_Open_Data" title="Linguistic Linked Open Data">Linguistic Linked Open Data</a></li> <li><a href="/wiki/Machine-readable_dictionary" title="Machine-readable dictionary">Machine-readable dictionary</a></li> <li><a href="/wiki/Parallel_text" title="Parallel text">Parallel text</a></li> <li><a href="/wiki/PropBank" title="PropBank">PropBank</a></li> <li><a href="/wiki/Semantic_network" title="Semantic network">Semantic network</a></li> <li><a href="/wiki/Simple_Knowledge_Organization_System" title="Simple Knowledge Organization System">Simple Knowledge Organization System</a></li> <li><a href="/wiki/Speech_corpus" title="Speech corpus">Speech corpus</a></li> <li><a href="/wiki/Text_corpus" title="Text corpus">Text corpus</a></li> <li><a href="/wiki/Thesaurus_(information_retrieval)" title="Thesaurus (information retrieval)">Thesaurus (information retrieval)</a></li> <li><a href="/wiki/Treebank" title="Treebank">Treebank</a></li> <li><a href="/wiki/Universal_Dependencies" title="Universal Dependencies">Universal Dependencies</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Data</th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/BabelNet" title="BabelNet">BabelNet</a></li> <li><a href="/wiki/Bank_of_English" title="Bank of English">Bank of English</a></li> <li><a href="/wiki/DBpedia" title="DBpedia">DBpedia</a></li> <li><a href="/wiki/FrameNet" title="FrameNet">FrameNet</a></li> <li><a href="/wiki/Google_Ngram_Viewer" class="mw-redirect" title="Google Ngram Viewer">Google Ngram Viewer</a></li> <li><a href="/wiki/UBY" title="UBY">UBY</a></li> <li><a href="/wiki/WordNet" title="WordNet">WordNet</a></li> <li><a href="/wiki/Wikidata" title="Wikidata">Wikidata</a></li></ul> </div></td></tr></tbody></table><div></div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Automatic_identification_and_data_capture" title="Automatic identification and data capture">Automatic identification<br />and data capture</a></th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Speech_recognition" title="Speech recognition">Speech recognition</a></li> <li><a href="/wiki/Speech_segmentation" title="Speech segmentation">Speech segmentation</a></li> <li><a href="/wiki/Speech_synthesis" title="Speech synthesis">Speech synthesis</a></li> <li><a href="/wiki/Natural_language_generation" title="Natural language generation">Natural language generation</a></li> <li><a href="/wiki/Optical_character_recognition" title="Optical character recognition">Optical character recognition</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Topic_model" title="Topic model">Topic model</a></th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Document_classification" title="Document classification">Document classification</a></li> <li><a href="/wiki/Latent_Dirichlet_allocation" title="Latent Dirichlet allocation">Latent Dirichlet allocation</a></li> <li><a href="/wiki/Pachinko_allocation" title="Pachinko allocation">Pachinko allocation</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Computer-assisted_reviewing" title="Computer-assisted reviewing">Computer-assisted<br />reviewing</a></th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Automated_essay_scoring" title="Automated essay scoring">Automated essay scoring</a></li> <li><a href="/wiki/Concordancer" title="Concordancer">Concordancer</a></li> <li><a href="/wiki/Grammar_checker" title="Grammar checker">Grammar checker</a></li> <li><a href="/wiki/Predictive_text" title="Predictive text">Predictive text</a></li> <li><a href="/wiki/Pronunciation_assessment" title="Pronunciation assessment">Pronunciation assessment</a></li> <li><a href="/wiki/Spell_checker" title="Spell checker">Spell checker</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%"><a href="/wiki/Natural-language_user_interface" title="Natural-language user interface">Natural language<br />user interface</a></th><td class="navbox-list-with-group navbox-list navbox-even" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Chatbot" title="Chatbot">Chatbot</a></li> <li><a href="/wiki/Interactive_fiction" title="Interactive fiction">Interactive fiction</a> (c.f. <a href="/wiki/Syntax_guessing" class="mw-redirect" title="Syntax guessing">Syntax guessing</a>)</li> <li><a href="/wiki/Question_answering" title="Question answering">Question answering</a></li> <li><a href="/wiki/Virtual_assistant" title="Virtual assistant">Virtual assistant</a></li> <li><a href="/wiki/Voice_user_interface" title="Voice user interface">Voice user interface</a></li></ul> </div></td></tr><tr><th scope="row" class="navbox-group" style="width:1%">Related</th><td class="navbox-list-with-group navbox-list navbox-odd" style="width:100%;padding:0"><div style="padding:0 0.25em"> <ul><li><a href="/wiki/Formal_semantics_(natural_language)" title="Formal semantics (natural language)">Formal semantics</a></li> <li><a href="/wiki/Hallucination_(artificial_intelligence)" title="Hallucination (artificial intelligence)">Hallucination</a></li> <li><a href="/wiki/Natural_Language_Toolkit" title="Natural Language Toolkit">Natural Language Toolkit</a></li> <li><a href="/wiki/SpaCy" title="SpaCy">spaCy</a></li></ul> </div></td></tr></tbody></table></div> <!-- NewPP limit report Parsed by mw‐web.codfw.main‐6ff899887‐7tkmj Cached time: 20250220223108 Cache expiry: 2592000 Reduced expiry: false Complications: [vary‐revision‐sha1, show‐toc] CPU time usage: 1.001 seconds Real time usage: 1.103 seconds Preprocessor visited node count: 6524/1000000 Post‐expand include size: 325534/2097152 bytes Template argument size: 3751/2097152 bytes Highest expansion depth: 12/100 Expensive parser function count: 5/500 Unstrip recursion depth: 1/20 Unstrip post‐expand size: 407229/5000000 bytes Lua time usage: 0.631/10.000 seconds Lua memory usage: 6911860/52428800 bytes Number of Wikibase entities loaded: 0/400 --> <!-- Transclusion expansion time report (%,ms,calls,template) 100.00% 919.245 1 -total 61.86% 568.627 1 Template:Reflist 31.96% 293.830 68 Template:Cite_web 13.98% 128.535 19 Template:Cite_journal 10.48% 96.303 1 Template:Machine_learning 9.38% 86.200 1 Template:Sidebar_with_collapsible_lists 7.59% 69.736 12 Template:Navbox 7.23% 66.471 1 Template:Short_description 4.45% 40.891 2 Template:Pagetype 4.37% 40.204 1 Template:OpenAI_navbox --> <!-- Saved in parser cache with key enwiki:pcache:72970020:|#|:idhash:canonical and timestamp 20250220223108 and revision id 1275868082. Rendering was triggered because: page-view --> </div><!--esi <esi:include src="/esitest-fa8a495983347898/content" /> --><noscript><img src="https://login.wikimedia.org/wiki/Special:CentralAutoLogin/start?useformat=desktop&amp;type=1x1&amp;usesul3=0" alt="" width="1" height="1" style="border: none; position: absolute;"></noscript> <div class="printfooter" data-nosnippet="">Retrieved from "<a dir="ltr" href="https://en.wikipedia.org/w/index.php?title=Generative_pre-trained_transformer&amp;oldid=1275868082">https://en.wikipedia.org/w/index.php?title=Generative_pre-trained_transformer&amp;oldid=1275868082</a>"</div></div> <div id="catlinks" class="catlinks" data-mw="interface"><div id="mw-normal-catlinks" class="mw-normal-catlinks"><a href="/wiki/Help:Category" title="Help:Category">Categories</a>: <ul><li><a href="/wiki/Category:Large_language_models" title="Category:Large language models">Large language models</a></li><li><a href="/wiki/Category:Generative_artificial_intelligence" title="Category:Generative artificial intelligence">Generative artificial intelligence</a></li><li><a href="/wiki/Category:Artificial_neural_networks" title="Category:Artificial neural networks">Artificial neural networks</a></li><li><a href="/wiki/Category:Generative_pre-trained_transformers" title="Category:Generative pre-trained transformers">Generative pre-trained transformers</a></li><li><a href="/wiki/Category:OpenAI" title="Category:OpenAI">OpenAI</a></li></ul></div><div id="mw-hidden-catlinks" class="mw-hidden-catlinks mw-hidden-cats-hidden">Hidden categories: <ul><li><a href="/wiki/Category:CS1:_long_volume_value" title="Category:CS1: long volume value">CS1: long volume value</a></li><li><a href="/wiki/Category:Articles_with_short_description" title="Category:Articles with short description">Articles with short description</a></li><li><a href="/wiki/Category:Short_description_matches_Wikidata" title="Category:Short description matches Wikidata">Short description matches Wikidata</a></li><li><a href="/wiki/Category:All_articles_with_unsourced_statements" title="Category:All articles with unsourced statements">All articles with unsourced statements</a></li><li><a href="/wiki/Category:Articles_with_unsourced_statements_from_December_2024" title="Category:Articles with unsourced statements from December 2024">Articles with unsourced statements from December 2024</a></li><li><a href="/wiki/Category:Articles_with_unsourced_statements_from_November_2023" title="Category:Articles with unsourced statements from November 2023">Articles with unsourced statements from November 2023</a></li></ul></div></div> </div> </main> </div> <div class="mw-footer-container"> <footer id="footer" class="mw-footer" > <ul id="footer-info"> <li id="footer-info-lastmod"> This page was last edited on 15 February 2025, at 15:13<span class="anonymous-show">&#160;(UTC)</span>.</li> <li id="footer-info-copyright">Text is available under the <a href="/wiki/Wikipedia:Text_of_the_Creative_Commons_Attribution-ShareAlike_4.0_International_License" title="Wikipedia:Text of the Creative Commons Attribution-ShareAlike 4.0 International License">Creative Commons Attribution-ShareAlike 4.0 License</a>; additional terms may apply. By using this site, you agree to the <a href="https://foundation.wikimedia.org/wiki/Special:MyLanguage/Policy:Terms_of_Use" class="extiw" title="foundation:Special:MyLanguage/Policy:Terms of Use">Terms of Use</a> and <a href="https://foundation.wikimedia.org/wiki/Special:MyLanguage/Policy:Privacy_policy" class="extiw" title="foundation:Special:MyLanguage/Policy:Privacy policy">Privacy Policy</a>. Wikipedia® is a registered trademark of the <a rel="nofollow" class="external text" href="https://wikimediafoundation.org/">Wikimedia Foundation, Inc.</a>, a non-profit organization.</li> </ul> <ul id="footer-places"> <li id="footer-places-privacy"><a href="https://foundation.wikimedia.org/wiki/Special:MyLanguage/Policy:Privacy_policy">Privacy policy</a></li> <li id="footer-places-about"><a href="/wiki/Wikipedia:About">About Wikipedia</a></li> <li id="footer-places-disclaimers"><a href="/wiki/Wikipedia:General_disclaimer">Disclaimers</a></li> <li id="footer-places-contact"><a href="//en.wikipedia.org/wiki/Wikipedia:Contact_us">Contact Wikipedia</a></li> <li id="footer-places-wm-codeofconduct"><a href="https://foundation.wikimedia.org/wiki/Special:MyLanguage/Policy:Universal_Code_of_Conduct">Code of Conduct</a></li> <li id="footer-places-developers"><a href="https://developer.wikimedia.org">Developers</a></li> <li id="footer-places-statslink"><a href="https://stats.wikimedia.org/#/en.wikipedia.org">Statistics</a></li> <li id="footer-places-cookiestatement"><a href="https://foundation.wikimedia.org/wiki/Special:MyLanguage/Policy:Cookie_statement">Cookie statement</a></li> <li id="footer-places-mobileview"><a href="//en.m.wikipedia.org/w/index.php?title=Generative_pre-trained_transformer&amp;mobileaction=toggle_view_mobile" class="noprint stopMobileRedirectToggle">Mobile view</a></li> </ul> <ul id="footer-icons" class="noprint"> <li id="footer-copyrightico"><a href="https://wikimediafoundation.org/" class="cdx-button cdx-button--fake-button cdx-button--size-large cdx-button--fake-button--enabled"><picture><source media="(min-width: 500px)" srcset="/static/images/footer/wikimedia-button.svg" width="84" height="29"><img src="/static/images/footer/wikimedia.svg" width="25" height="25" alt="Wikimedia Foundation" lang="en" loading="lazy"></picture></a></li> <li id="footer-poweredbyico"><a href="https://www.mediawiki.org/" class="cdx-button cdx-button--fake-button cdx-button--size-large cdx-button--fake-button--enabled"><picture><source media="(min-width: 500px)" srcset="/w/resources/assets/poweredby_mediawiki.svg" width="88" height="31"><img src="/w/resources/assets/mediawiki_compact.svg" alt="Powered by MediaWiki" width="25" height="25" loading="lazy"></picture></a></li> </ul> </footer> </div> </div> </div> <div class="vector-header-container vector-sticky-header-container"> <div id="vector-sticky-header" class="vector-sticky-header"> <div class="vector-sticky-header-start"> <div class="vector-sticky-header-icon-start vector-button-flush-left vector-button-flush-right" aria-hidden="true"> <button class="cdx-button cdx-button--weight-quiet cdx-button--icon-only vector-sticky-header-search-toggle" tabindex="-1" data-event-name="ui.vector-sticky-search-form.icon"><span class="vector-icon mw-ui-icon-search mw-ui-icon-wikimedia-search"></span> <span>Search</span> </button> </div> <div role="search" class="vector-search-box-vue vector-search-box-show-thumbnail vector-search-box"> <div class="vector-typeahead-search-container"> <div class="cdx-typeahead-search cdx-typeahead-search--show-thumbnail"> <form action="/w/index.php" id="vector-sticky-search-form" class="cdx-search-input cdx-search-input--has-end-button"> <div class="cdx-search-input__input-wrapper" data-search-loc="header-moved"> <div class="cdx-text-input cdx-text-input--has-start-icon"> <input class="cdx-text-input__input" type="search" name="search" placeholder="Search Wikipedia"> <span class="cdx-text-input__icon cdx-text-input__start-icon"></span> </div> <input type="hidden" name="title" value="Special:Search"> </div> <button class="cdx-button cdx-search-input__end-button">Search</button> </form> </div> </div> </div> <div class="vector-sticky-header-context-bar"> <nav aria-label="Contents" class="vector-toc-landmark"> <div id="vector-sticky-header-toc" class="vector-dropdown mw-portlet mw-portlet-sticky-header-toc vector-sticky-header-toc vector-button-flush-left" > <input type="checkbox" id="vector-sticky-header-toc-checkbox" role="button" aria-haspopup="true" data-event-name="ui.dropdown-vector-sticky-header-toc" class="vector-dropdown-checkbox " aria-label="Toggle the table of contents" > <label id="vector-sticky-header-toc-label" for="vector-sticky-header-toc-checkbox" class="vector-dropdown-label cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only " aria-hidden="true" ><span class="vector-icon mw-ui-icon-listBullet mw-ui-icon-wikimedia-listBullet"></span> <span class="vector-dropdown-label-text">Toggle the table of contents</span> </label> <div class="vector-dropdown-content"> <div id="vector-sticky-header-toc-unpinned-container" class="vector-unpinned-container"> </div> </div> </div> </nav> <div class="vector-sticky-header-context-bar-primary" aria-hidden="true" ><span class="mw-page-title-main">Generative pre-trained transformer</span></div> </div> </div> <div class="vector-sticky-header-end" aria-hidden="true"> <div class="vector-sticky-header-icons"> <a href="#" class="cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only" id="ca-talk-sticky-header" tabindex="-1" data-event-name="talk-sticky-header"><span class="vector-icon mw-ui-icon-speechBubbles mw-ui-icon-wikimedia-speechBubbles"></span> <span></span> </a> <a href="#" class="cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only" id="ca-subject-sticky-header" tabindex="-1" data-event-name="subject-sticky-header"><span class="vector-icon mw-ui-icon-article mw-ui-icon-wikimedia-article"></span> <span></span> </a> <a href="#" class="cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only" id="ca-history-sticky-header" tabindex="-1" data-event-name="history-sticky-header"><span class="vector-icon mw-ui-icon-wikimedia-history mw-ui-icon-wikimedia-wikimedia-history"></span> <span></span> </a> <a href="#" class="cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only mw-watchlink" id="ca-watchstar-sticky-header" tabindex="-1" data-event-name="watch-sticky-header"><span class="vector-icon mw-ui-icon-wikimedia-star mw-ui-icon-wikimedia-wikimedia-star"></span> <span></span> </a> <a href="#" class="cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only" id="ca-edit-sticky-header" tabindex="-1" data-event-name="wikitext-edit-sticky-header"><span class="vector-icon mw-ui-icon-wikimedia-wikiText mw-ui-icon-wikimedia-wikimedia-wikiText"></span> <span></span> </a> <a href="#" class="cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only" id="ca-ve-edit-sticky-header" tabindex="-1" data-event-name="ve-edit-sticky-header"><span class="vector-icon mw-ui-icon-wikimedia-edit mw-ui-icon-wikimedia-wikimedia-edit"></span> <span></span> </a> <a href="#" class="cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--icon-only" id="ca-viewsource-sticky-header" tabindex="-1" data-event-name="ve-edit-protected-sticky-header"><span class="vector-icon mw-ui-icon-wikimedia-editLock mw-ui-icon-wikimedia-wikimedia-editLock"></span> <span></span> </a> </div> <div class="vector-sticky-header-buttons"> <button class="cdx-button cdx-button--weight-quiet mw-interlanguage-selector" id="p-lang-btn-sticky-header" tabindex="-1" data-event-name="ui.dropdown-p-lang-btn-sticky-header"><span class="vector-icon mw-ui-icon-wikimedia-language mw-ui-icon-wikimedia-wikimedia-language"></span> <span>26 languages</span> </button> <a href="#" class="cdx-button cdx-button--fake-button cdx-button--fake-button--enabled cdx-button--weight-quiet cdx-button--action-progressive" id="ca-addsection-sticky-header" tabindex="-1" data-event-name="addsection-sticky-header"><span class="vector-icon mw-ui-icon-speechBubbleAdd-progressive mw-ui-icon-wikimedia-speechBubbleAdd-progressive"></span> <span>Add topic</span> </a> </div> <div class="vector-sticky-header-icon-end"> <div class="vector-user-links"> </div> </div> </div> </div> </div> <div class="vector-settings" id="p-dock-bottom"> <ul></ul> </div><script>(RLQ=window.RLQ||[]).push(function(){mw.config.set({"wgHostname":"mw-web.codfw.main-d8647bfd6-h9rjt","wgBackendResponseTime":128,"wgPageParseReport":{"limitreport":{"cputime":"1.001","walltime":"1.103","ppvisitednodes":{"value":6524,"limit":1000000},"postexpandincludesize":{"value":325534,"limit":2097152},"templateargumentsize":{"value":3751,"limit":2097152},"expansiondepth":{"value":12,"limit":100},"expensivefunctioncount":{"value":5,"limit":500},"unstrip-depth":{"value":1,"limit":20},"unstrip-size":{"value":407229,"limit":5000000},"entityaccesscount":{"value":0,"limit":400},"timingprofile":["100.00% 919.245 1 -total"," 61.86% 568.627 1 Template:Reflist"," 31.96% 293.830 68 Template:Cite_web"," 13.98% 128.535 19 Template:Cite_journal"," 10.48% 96.303 1 Template:Machine_learning"," 9.38% 86.200 1 Template:Sidebar_with_collapsible_lists"," 7.59% 69.736 12 Template:Navbox"," 7.23% 66.471 1 Template:Short_description"," 4.45% 40.891 2 Template:Pagetype"," 4.37% 40.204 1 Template:OpenAI_navbox"]},"scribunto":{"limitreport-timeusage":{"value":"0.631","limit":"10.000"},"limitreport-memusage":{"value":6911860,"limit":52428800}},"cachereport":{"origin":"mw-web.codfw.main-6ff899887-7tkmj","timestamp":"20250220223108","ttl":2592000,"transientcontent":false}}});});</script> <script type="application/ld+json">{"@context":"https:\/\/schema.org","@type":"Article","name":"Generative pre-trained transformer","url":"https:\/\/en.wikipedia.org\/wiki\/Generative_pre-trained_transformer","sameAs":"http:\/\/www.wikidata.org\/entity\/Q116777014","mainEntity":"http:\/\/www.wikidata.org\/entity\/Q116777014","author":{"@type":"Organization","name":"Contributors to Wikimedia projects"},"publisher":{"@type":"Organization","name":"Wikimedia Foundation, Inc.","logo":{"@type":"ImageObject","url":"https:\/\/www.wikimedia.org\/static\/images\/wmf-hor-googpub.png"}},"datePublished":"2023-02-07T15:42:35Z","dateModified":"2025-02-15T15:13:58Z","image":"https:\/\/upload.wikimedia.org\/wikipedia\/commons\/5\/51\/Full_GPT_architecture.svg","headline":"type of large language model"}</script> </body> </html>

Pages: 1 2 3 4 5 6 7 8 9 10