The history of open-source artificial intelligence is intertwined with both the development of AI technologies and the growth of the
open-source software movement.
1990s: Early development of AI and open-source software The concept of AI dates back to the mid-20th century, when computer scientists like
Alan Turing and
John McCarthy laid the groundwork for modern AI theories and algorithms. An early form of AI, the
natural language processing "doctor"
ELIZA, was re-implemented and shared in 1977 by Jeff Shrager as a BASIC program, and soon translated to many other languages. Early AI research focused on developing
symbolic reasoning systems and
rule-based expert systems. During this period, the idea of open-source software was beginning to take shape, with pioneers like
Richard Stallman advocating for free software as a means to promote collaboration and innovation in programming. The
Free Software Foundation, founded in 1985 by Stallman, was one of the first major organizations to promote the idea of software that could be freely used, modified, and distributed. The ideas from this movement eventually influenced the development of open-source AI, as more developers began to see the potential benefits of open collaboration in software creation, including AI models and algorithms. In the 1990s, open-source software began to gain more traction, the rise of machine learning and statistical methods also led to the development of more practical AI tools. In 1993, the CMU Artificial Intelligence Repository was initiated, with a variety of openly shared software.
2000s: Emergence of open-source AI In the early 2000s open-source AI began to take off, with the release of more user-friendly foundational libraries and frameworks that were available for anyone to use and contribute to.
OpenCV was released in 2000 with a variety of traditional AI algorithms like
decision trees,
k-Nearest Neighbors (kNN),
Naive Bayes and
Support Vector Machines (SVM).
2010s: Rise of open-source AI frameworks Open-source deep learning framework as
Torch was released in 2002 and made open-source with Torch7 in 2011, and was later augmented by
PyTorch, and
TensorFlow.
AlexNet was released in 2012.
OpenAI was founded in 2015 with a mission to create open-source artificial intelligence that benefited humanity, at least in part to help with recruitment in the early phases of the organization.
GPT-1 was released in 2018.
2020s: Open-weight and open-source generative AI With the announcement of
GPT-2 in 2019, OpenAI originally planned to keep the source code of their models private citing concerns about malicious applications. After OpenAI faced public backlash, however, it released the source code for GPT-2 to GitHub three months after its release. At the time of GPT-3's release GPT-2 was still the most powerful open source language model in the world. 2022 also saw the rise of larger and more powerful models under licenses of varying openness including Meta's OPT. The
Open Source Initiative consulted experts over two years to create a definition of "open-source" that would fit the needs of AI software and models. The most controversial aspect relates to data access, since some models are trained on sensitive data which can't be released. In 2024, they published the Open Source AI Definition 1.0 (OSAID 1.0). along with
MosaicML's smaller open-source models. The release of the Llama models was a milestone in generating interest in open-weight and open-source models. In 2024, Meta released a collection of large AI models, including
Llama 3.1 405B, which was competitive with less open models. Meta's description of Llama as open-source has been disputed due to Llama's software license, which prohibits it from being used for some purposes, and due to Meta not disclosing the origin of the data used to train the models.
DeepSeek released their V3 LLM in December 2024, and their R1
reasoning model on 20 January 2025, both as open-weights models under the MIT license. This release made widely known how China had been embracing using and building more open AI systems as a way to reduce reliance on western software and gatekeeping as well as to help give its industries access to higher-powered AI more quickly. Projects based in China have since become more widely used around the world as well as they have closed at least some of the gap with leading proprietary American models. Since the release of OpenAI's proprietary ChatGPT model in late 2022, there have been only a few fully open (weights, data, code, etc.) large language models released. In September 2025, a Swiss consortium added to this short list by releasing a fully open model named
Apertus. In December 2025, the
Linux Foundation created the
Agentic AI Foundation, which assumed control of some open-source
agentic AI protocols and other technologies created by OpenAI,
Anthropic and
Block. Starting in November 2024,
Lightricks began releasing the
LTX video models as open weights. == Significance ==