Frontier models Certain highly advanced foundation models are termed "frontier models", which have the potential to "possess dangerous capabilities sufficient to pose severe risks to public safety." These "dangerous capabilities" stem from the accidental or intentional misuse of such models, which in conjunction with their powerful nature can lead to severe harms. As foundation models continue to improve, some AI researchers speculate that almost all next-generation foundation models will be considered frontier models. Since the concept of dangerous capabilities is inherently subjective, there is no strict designation for what foundation models qualify as frontier models. However, some generally held ideas for sufficiently dangerous capabilities include: • Designing and synthesizing new biological or chemical weapons • Producing and propagating convincing, tailored disinformation with minimal user instruction • Harnessing unprecedented offensive cyber capabilities • Evading human control through deceptive means The U.S. definition (via the
AI Safety Institute at
NIST) tied "frontier" status also to the compute technical threshold to create objective legal requirements. Due to frontier models' unique capabilities, it is difficult to effectively regulate their development and deployment. Because of their emergent nature, new dangerous capabilities can appear on their own in frontier models, both in the development stage and after being deployed. A global network of governmental and intergovernmental bodies has emerged to create formal safety standards, including the Global Network of AI Safety Institutes (AISIs) and the G7"Hiroshima AI Process" (HAIP).
General-purpose AI Due to their adaptability to a wide range of use-cases, foundation models are sometimes considered to be examples of general-purpose AI. In designing the EU AI Act, the European Parliament has stated that a new wave of general-purpose AI technologies shapes the overall AI ecosystem. The fuller structure of the ecosystem, in addition to the properties of specific general-purpose AI systems, influences the design of AI policy and research. General-purpose AI systems also often appear in people's everyday lives through applications and tools like
ChatGPT or
DALL-E. Government agencies like EU Parliament have identified regulation of general-purpose AI, such as foundation models, to be a high priority. General-purpose AI systems are often characterized by large size, opacity, and potential for emergence, all of which can create unintended harms. Such systems also heavily influence downstream applications, which further exacerbates the need for regulation. In regards to prominent legislation, a number of stakeholders have pushed for the
EU AI Act to include restrictions on general-purpose AI systems, all of which would also apply to foundation models.
World models World models are sometimes described as foundation models. as well as to implicitly model physical concepts such as
gravity. as well as videos or 3D scenes, and the resulting 3D environments can be exported. World models do not have a fully agreed definition, but have been divided into two scopes: one for representing and understanding the current environment, and another for predicting the future state of that environment. In the former view, world models are developed using model-based
reinforcement learning and a
Markov decision process, using
model predictive control or
Monte Carlo tree search to create policies. With the latter, (
multimodal) large language models or video generation models can be used. In addition, these environments can be immersive simulations for training AI agents that can interact in the real world.
History Quanta Magazine traced world models back to a 1943 publication by
Kenneth Craik on mental models and the
blocks world of
SHRDLU in the 1960s.
Business Insider traced world models to a 1971 paper by
Jay Wright Forrester. In 2022,
Yann LeCun saw a world model (defined by him as a
neural network that acts as a
mental model for aspects of the world that are seen as relevant) as part of a larger system of
cognitive architecture – other neural networks that are analogous to different regions of the
brain. In his view, this framework could lead to
commonsense reasoning. LeCun has estimated that world models would be fully functional by the late 2020s to mid 2030s.
Training World models are trained on a variety of data modalities, including text, images, audio and video, and have been applied to
video generation. One open source dataset for world models includes 1 billion data points across multiple modalities (text, images, audio, video and
point clouds), including 1 million
manual annotations. The
South China Morning Post wrote that
Manycore Tech was another example of companies aiming to build a world model, viewing their work as an example of
spatial intelligence. In May 2025,
Mohamed bin Zayed University of Artificial Intelligence released a world model for building simulations to test
AI agents.
Google DeepMind has also released two world models in
two-dimensional space and
three-dimensional space, respectively, that were trained on video data, with Google claiming that the latter can be a training environment for AI agents.
Meta released a world model in June 2025,
Tencent released an open source world model in July 2025.
Niantic, Inc. spinoff,
Niantic Spatial, is developing a world model using anonymized player scans from
Pokémon GO. Other companies that are planning as of 2025 to build world models include
ByteDance Applications Fei-Fei Li views world models as applying to
robotics and
creative works. Due to the complexity of these models, she advocates for more complex strategies in
data acquisition,
data engineering,
data processing, and
synthesizing data. She co-founded a startup on building world models, which, as of 2024, planned to do so in three phases: incorporating an understanding of three-dimensional space along with time; support for
augmented reality; and support for robotics. Her startup, World Labs, released its commercial world model, Marble, in November 2025. World models are intended for use in interactive media (such as video games and movies) and environment simulation. Proposed use cases for world models include action planning and outcome prediction. As of October 2025, research has shown mixed results in the spatial reasoning capabilities of
text-to-video models (in particular,
Veo 3). In Ophthalmology Foundation Models have been applied to retinal imaging including volumetric optical coherence tomography
Concerns TechCrunch noted that world models could use more data than
large language models and would require significantly more computational power (including the use of thousands of
GPUs for training and inference). and
copyright. == Technical details ==