Back

generative pre trained transformer (GPT)

GPT stands for Generative Pre-trained Transformer, which is an artificial intelligence language model developed by OpenAI. The “generative” aspect refers to the model’s ability to generate text, “pre-trained” indicates that the model has been trained on a large dataset before it is fine-tuned for specific tasks, and “Transformer” is the name of the architecture on which GPT is based[1][2][4][6].


GPT models are designed to perform various language-related tasks, such as answering questions, translating languages, generating extemporaneous text, and creating computer code. They are trained on vast amounts of text data, allowing them to predict the most likely next word in a sequence, thereby generating coherent and contextually relevant text[2][3].


The GPT series has evolved rapidly, with GPT-3 being one of the most well-known versions, featuring a neural network with over 175 billion parameters. GPT-4, introduced in March 2023, is a multimodal model that accepts both text and images, and there are plans for future versions like GPT-5 to support videos[2][7].


Citations:

[1] https://en.wikipedia.org/wiki/GPT

[2] https://www.pcmag.com/encyclopedia/term/gpt

[3] https://aws.amazon.com/what-is/gpt/

[4] https://www.iotforall.com/what-does-gpt-stand-for

[5] https://www.dictionary.com/browse/gpt

[6] https://zapier.com/blog/what-is-gpt/

[7] https://www.techtarget.com/searchenterpriseai/definition/GPT-3

[8] https://dictionary.cambridge.org/us/dictionary/english/gpt

Share: