학술논문

GPT-4 Technical Report
Document Type
Working Paper
Author
OpenAIAchiam, JoshAdler, StevenAgarwal, SandhiniAhmad, LamaAkkaya, IlgeAleman, Florencia LeoniAlmeida, DiogoAltenschmidt, JankoAltman, SamAnadkat, ShyamalAvila, RedBabuschkin, IgorBalaji, SuchirBalcom, ValerieBaltescu, PaulBao, HaimingBavarian, MohammadBelgum, JeffBello, IrwanBerdine, JakeBernadett-Shapiro, GabrielBerner, ChristopherBogdonoff, LennyBoiko, OlegBoyd, MadelaineBrakman, Anna-LuisaBrockman, GregBrooks, TimBrundage, MilesButton, KevinCai, TrevorCampbell, RosieCann, AndrewCarey, BrittanyCarlson, ChelseaCarmichael, RoryChan, BrookeChang, CheChantzis, FotisChen, DerekChen, SullyChen, RubyChen, JasonChen, MarkChess, BenCho, ChesterChu, CaseyChung, Hyung WonCummings, DaveCurrier, JeremiahDai, YunxingDecareaux, CoryDegry, ThomasDeutsch, NoahDeville, DamienDhar, ArkaDohan, DavidDowling, SteveDunning, SheilaEcoffet, AdrienEleti, AttyEloundou, TynaFarhi, DavidFedus, LiamFelix, NikoFishman, Simón PosadaForte, JustonFulford, IsabellaGao, LeoGeorges, ElieGibson, ChristianGoel, VikGogineni, TarunGoh, GabrielGontijo-Lopes, RaphaGordon, JonathanGrafstein, MorganGray, ScottGreene, RyanGross, JoshuaGu, Shixiang ShaneGuo, YufeiHallacy, ChrisHan, JesseHarris, JeffHe, YuchenHeaton, MikeHeidecke, JohannesHesse, ChrisHickey, AlanHickey, WadeHoeschele, PeterHoughton, BrandonHsu, KennyHu, ShengliHu, XinHuizinga, JoostJain, ShantanuJain, ShawnJang, JoanneJiang, AngelaJiang, RogerJin, HaozhunJin, DennyJomoto, ShinoJonn, BillieJun, HeewooKaftan, TomerKaiser, ŁukaszKamali, AliKanitscheider, IngmarKeskar, Nitish ShirishKhan, TabarakKilpatrick, LoganKim, Jong WookKim, ChristinaKim, YongjikKirchner, Jan HendrikKiros, JamieKnight, MattKokotajlo, DanielKondraciuk, ŁukaszKondrich, AndrewKonstantinidis, ArisKosic, KyleKrueger, GretchenKuo, VishalLampe, MichaelLan, IkaiLee, TeddyLeike, JanLeung, JadeLevy, DanielLi, Chak MingLim, RachelLin, MollyLin, StephanieLitwin, MateuszLopez, TheresaLowe, RyanLue, PatriciaMakanju, AnnaMalfacini, KimManning, SamMarkov, TodorMarkovski, YanivMartin, BiancaMayer, KatieMayne, AndrewMcGrew, BobMcKinney, Scott MayerMcLeavey, ChristineMcMillan, PaulMcNeil, JakeMedina, DavidMehta, AalokMenick, JacobMetz, LukeMishchenko, AndreyMishkin, PamelaMonaco, VinnieMorikawa, EvanMossing, DanielMu, TongMurati, MiraMurk, OlegMély, DavidNair, AshvinNakano, ReiichiroNayak, RajeevNeelakantan, ArvindNgo, RichardNoh, HyeonwooOuyang, LongO'Keefe, CullenPachocki, JakubPaino, AlexPalermo, JoePantuliano, AshleyParascandolo, GiambattistaParish, JoelParparita, EmyPassos, AlexPavlov, MikhailPeng, AndrewPerelman, AdamPeres, Filipe de Avila BelbutePetrov, MichaelPinto, Henrique Ponde de OliveiraMichaelPokornyPokrass, MichellePong, Vitchyr H.Powell, TollyPower, AletheaPower, BorisProehl, ElizabethPuri, RaulRadford, AlecRae, JackRamesh, AdityaRaymond, CameronReal, FrancisRimbach, KendraRoss, CarlRotsted, BobRoussez, HenriRyder, NickSaltarelli, MarioSanders, TedSanturkar, ShibaniSastry, GirishSchmidt, HeatherSchnurr, DavidSchulman, JohnSelsam, DanielSheppard, KylaSherbakov, TokiShieh, JessicaShoker, SarahShyam, PranavSidor, SzymonSigler, EricSimens, MaddieSitkin, JordanSlama, KatarinaSohl, IanSokolowsky, BenjaminSong, YangStaudacher, NatalieSuch, Felipe PetroskiSummers, NatalieSutskever, IlyaTang, JieTezak, NikolasThompson, Madeleine B.Tillet, PhilTootoonchian, AminTseng, ElizabethTuggle, PrestonTurley, NickTworek, JerryUribe, Juan Felipe CerónVallone, AndreaVijayvergiya, ArunVoss, ChelseaWainwright, CarrollWang, Justin JayWang, AlvinWang, BenWard, JonathanWei, JasonWeinmann, CJWelihinda, AkilaWelinder, PeterWeng, JiayiWeng, LilianWiethoff, MattWillner, DaveWinter, ClemensWolrich, SamuelWong, HannahWorkman, LaurenWu, SherwinWu, JeffWu, MichaelXiao, KaiXu, TaoYoo, SarahYu, KevinYuan, QimingZaremba, WojciechZellers, RowanZhang, ChongZhang, MarvinZhao, ShengjiaZheng, TianhaoZhuang, JuntangZhuk, WilliamZoph, Barret
Source
Subject
Computer Science - Computation and Language
Computer Science - Artificial Intelligence
Language
Abstract
We report the development of GPT-4, a large-scale, multimodal model which can accept image and text inputs and produce text outputs. While less capable than humans in many real-world scenarios, GPT-4 exhibits human-level performance on various professional and academic benchmarks, including passing a simulated bar exam with a score around the top 10% of test takers. GPT-4 is a Transformer-based model pre-trained to predict the next token in a document. The post-training alignment process results in improved performance on measures of factuality and adherence to desired behavior. A core component of this project was developing infrastructure and optimization methods that behave predictably across a wide range of scales. This allowed us to accurately predict some aspects of GPT-4's performance based on models trained with no more than 1/1,000th the compute of GPT-4.
Comment: 100 pages; updated authors list; fixed author names and added citation