computers to mimic human intelligence & behaviour A subset of ML exposing multilayered neural networks to vast amount of data A subset of AI that includes statistical techniques to solve the tasks using experience
new outputs from scratch. So, given a prompt, it can imagine the rest. In the case of GPT, it can only generate text. Other models can also generate images, sounds, etc. (DALL-E, Stable Diffusion, Riffusion, ...)
Think ~all of the publically available text on the internet. * Models trained on LOTS of data are called Large Language Models (LLM). GPT is an example of an LLM, but there exist many others too (Flan-T5, BLOOM, ...)
Transformers were first published in 2017 at NeurIPS by researchers at Google. This was a watershed moment for LLMs. It allowed researchers to train models efficiently on huge datasets. https://arxiv.org/pdf/1706.03762.pdf