The development of increasingly
large language models in the late 2010s and early 2020s created substantial computational challenges.
GPT-1, released in 2018 with 117 million parameters, cost less than $50,000 to train.
GPT-2, released in 2019 with 1.5 billion parameters, required $40,000 to train. Training consumed approximately 1,287 megawatt-hours of electricity.
GPT-4, released in 2023, required over $100 million to train and consumed approximately 50 gigawatt-hours of energy using 25,000 Nvidia A100
GPUs running for 90 to 100 days.
GPT-5, released in August 2025, required individual training runs costing over $500 million each, with total training costs estimated between $1.25 billion and $2.5 billion. This created a barrier where adapting such models to specific tasks through traditional fine-tuning became prohibitively expensive for most researchers and organizations. ==Purpose==