Hubbry Logo
Foundation modelFoundation modelMain
Open search
Foundation model
Community hub
Foundation model
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Foundation model
Foundation model
from Wikipedia

In artificial intelligence, a foundation model (FM), also known as large x model (LxM, where "x" is a variable representing any text, image, sound, etc.), is a machine learning or deep learning model trained on vast datasets so that it can be applied across a wide range of use cases.[1] Generative AI applications like large language models (LLM) are common examples of foundation models.[1]

Building foundation models is often highly resource-intensive, with the most advanced models costing hundreds of millions of dollars to cover the expenses of acquiring, curating, and processing massive datasets, as well as the compute power required for training.[2] These costs stem from the need for sophisticated infrastructure, extended training times, and advanced hardware, such as GPUs. In contrast, adapting an existing foundation model for a specific task or using it directly is far less costly, as it leverages pre-trained capabilities and typically requires only fine-tuning on smaller, task-specific datasets.

Early examples of foundation models are language models like OpenAI's GPT series and Google's BERT.[3][4] Beyond text, foundation models have been developed across a range of modalities—including DALL-E, Stable diffusion, and Flamingo[5] for images, MusicGen[6] and LLark[7] for music, and RT-2[8] for robotic control. Foundation models are also being developed for fields like astronomy,[9] radiology,[10] genomics,[11] coding,[12] times-series forecasting,[13] mathematics,[14] and chemistry.[15]

Definitions

[edit]

The Stanford Institute for Human-Centered Artificial Intelligence's (HAI) Center for Research on Foundation Models (CRFM) coined the term "foundation model" in August 2021[16] to mean "any model that is trained on broad data (generally using self-supervision at scale) that can be adapted (e.g., fine-tuned) to a wide range of downstream tasks".[17] This was based on their observation that preexisting terms, while overlapping, were not adequate, stating that "'large language model' was too narrow given the focus is not only language; 'self-supervised model' was too specific to the training objective; and 'pretrained model' suggested that the noteworthy action all happened after 'pretraining."[18] The term "foundation model" was chosen over "foundational model"[19] because "foundational" implies that these models provide fundamental principles in a way that "foundation" does not.[20] The term vision-language model (VLM) is also used as a near-synonym.

As governments regulate foundation models, new legal definitions have emerged.

  • In the United States, the Executive Order on the Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence defines a foundation model as "an AI model that is trained on broad data; generally uses self-supervision; contains at least tens of billions of parameters; is applicable across a wide range of contexts".[21]
  • In the United States, the proposed AI Foundation Model Transparency Act of 2023[22] by House Representatives Don Beyer (DVA) and Anna Eshoo (D, CA) defines a foundation model as "an artificial intelligence model trained on broad data, generally uses self supervision, generally contains at least 1,000,000,000 parameters, is applicable across a wide range of contexts, and exhibits, or could be easily modified to exhibit, high levels of performance at tasks that could pose a serious risk to security, national economic security, national public health or safety, or any combination of those matters."
  • In the European Union, the European Parliament's negotiated position on the E.U. AI Act defines a foundation model as an "AI model that is trained on broad data at scale, is designed for generality of output, and can be adapted to a wide range of distinctive tasks".
  • In the United Kingdom, the Competition and Markets Authority's AI Foundation Models: Initial Report[1] defines foundations model as "a type of AI technology that are trained on vast amounts of data that can be adapted to a wide range of tasks and operations."

The United States's definitions are the only ones to make reference to the size of a foundation model, and differ on magnitude. Beyer and Eshoo's definition also specifies that foundation models must achieve a level of performance as to be a potential danger. In contrast, the E.U. definition requires the model to be designed for generality of output. All definitions agree that foundation models must be trained on a broad range of data with potential applications in many domains.

History

[edit]

Technologically, foundation models are built using established machine learning techniques like deep neural networks, transfer learning, and self-supervised learning. Foundation models differ from previous techniques as they are general purpose models that function as a reusable infrastructure, instead of bespoke and one-off task-specific models.

Advances in computer parallelism (e.g., CUDA GPUs) and new developments in neural network architecture (e.g., Transformers), and the increased use of training data with minimal supervision all contributed to the rise of foundation models. Foundation models began to materialize as the latest wave of deep learning models in the late 2010s.[23] Relative to most prior work on deep learning, these language models demonstrated the potential of training on much larger web-sourced datasets using self-supervised objectives (e.g. predicting the next word in a large corpus of text). These approaches, which draw upon earlier works like word2vec and GloVe, deviated from prior supervised approaches that required annotated data (e.g. crowd-sourced labels).

The 2022 releases of Stable Diffusion and ChatGPT (initially powered by the GPT-3.5 model) led to foundation models and generative AI entering widespread public discourse. Further, releases of LLaMA, Llama 2, and Mistral in 2023 contributed to a greater emphasis placed on how foundation models are released with open foundation models garnering a lot of support[24] and scrutiny.[25]

[edit]

Frontier models

[edit]

Certain highly advanced foundation models are termed "frontier models", which have the potential to "possess dangerous capabilities sufficient to pose severe risks to public safety."[26] These "dangerous capabilities" stem from the accidental or intentional misuse of such models, which in conjunction with their powerful nature can lead to severe harms. As foundation models continue to improve, some AI researchers speculate that almost all next-generation foundation models will be considered frontier models.

Since the concept of dangerous capabilities is inherently subjective, there is no strict designation for what foundation models qualify as frontier models. However, some generally held ideas for sufficiently dangerous capabilities include:

  • Designing and synthesizing new biological or chemical weapons[27]
  • Producing and propagating convincing, tailored disinformation with minimal user instruction[28]
  • Harnessing unprecedented offensive cyber capabilities[29]
  • Evading human control through deceptive means[30]

Due to frontier models' unique capabilities, it is difficult to effectively regulate their development and deployment. Because of their emergent nature, new dangerous capabilities can appear on their own in frontier models, both in the development stage and after being deployed.[26] Additionally, since frontier models continue to adapt after deployment, it remains difficult to mitigate all harms that arise from already-deployed models. If a frontier model happens to be open-source or is released online, the model can also disseminate rapidly, further hampering regulators by creating a lack of accountability.

General-purpose AI

[edit]

Due to their adaptability to a wide range of use-cases, foundation models are sometimes considered to be examples of general-purpose AI. In designing the EU AI Act, the European Parliament has stated that a new wave of general-purpose AI technologies shapes the overall AI ecosystem.[31] The fuller structure of the ecosystem, in addition to the properties of specific general-purpose AI systems, influences the design of AI policy and research.[32] General-purpose AI systems also often appear in people's everyday lives through applications and tools like ChatGPT or DALL-E.

Government agencies like EU Parliament have identified regulation of general-purpose AI, such as foundation models, to be a high priority. General-purpose AI systems are often characterized by large size, opacity, and potential for emergence, all of which can create unintended harms. Such systems also heavily influence downstream applications, which further exacerbates the need for regulation. In regards to prominent legislation, a number of stakeholders have pushed for the EU AI Act to include restrictions on general-purpose AI systems, all of which would also apply to foundation models.

World models

[edit]

World models are sometimes described as foundation models.[33][34] World models are a representation of an environment intended to predict the state of that environment after taking a set of actions,[35][36] as well as to implicitly model physical concepts such as gravity.[36] Input prompts for world models can include text or images,[37][38] as well as videos or 3D scenes,[39] and the resulting 3D environments can be exported.[39] World models, alongside embodied AI, multi-agent models, and neuroscience models of the brain, are seen as alternatives to large language models for achieving general artificial intelligence.[40]

World models do not have a fully agreed definition, but have been divided into two scopes: one for representing and understanding the current environment, and another for predicting the future state of that environment. In the former view, world models are developed using model-based reinforcement learning and a Markov decision process, using model predictive control or Monte Carlo tree search to create policies. With the latter, (multimodal) large language models or video generation models can be used. In addition, these environments can be immersive simulations for training AI agents that can interact in the real world.[41]

History

[edit]

Quanta Magazine traced world models back to a 1943 publication by Kenneth Craik on mental models and the blocks world of SHRDLU in the 1960s.[42] Business Insider traced world models to a 1971 paper by Jay Wright Forrester.[40] A related idea of organizing world knowledge, the frame representation, was proposed by Marvin Minsky in 1974.[41]

In 2018, researchers David Ha and Jürgen Schmidhuber defined world models in the context of reinforcement learning: an agent with a variational autoencoder model V for representing visual observations, a recurrent neural network model M for representing memory, and a linear model C for making decisions. They suggested that agents trained on world models in environments that simulate reality could be applied to real world settings.[43]

In 2022, Yann LeCun saw a world model (defined by him as a neural network that acts as a mental model for aspects of the world that are seen as relevant) as part of a larger system of cognitive architecture – other neural networks that are analogous to different regions of the brain. In his view, this framework could lead to commonsense reasoning.[44][45] LeCun has estimated that world models would be fully functional by the late 2020s[46] to mid 2030s.[47]

Training

[edit]

World models are trained on a variety of data modalities, including text, images, audio and video, and have been applied to video generation.[48] One open source dataset for world models includes 1 billion data points across multiple modalities (text, images, audio, video and point clouds), including 1 million manual annotations.[36]

Examples

[edit]

TechCrunch saw Sora as an example of a world model,[48] while in January 2025, Nvidia released its own set of world models.[49][34] The South China Morning Post wrote that Manycore Tech was another example of companies aiming to build a world model, viewing their work as an example of spatial intelligence.[50] In May 2025, Mohamed bin Zayed University of Artificial Intelligence released a world model for building simulations to test AI agents.[51]

Google DeepMind has also released two world models in two-dimensional space and three-dimensional space, respectively, that were trained on video data, with Google claiming that the latter can be a training environment for AI agents.[52][53] Meta released a world model in June 2025,[54] Tencent released an open source world model in July 2025.[55] Niantic, Inc. spinoff, Niantic Spatial, is developing a world model using anonymized player scans from Pokémon GO.[56][57] Other companies that are planning as of 2025 to build world models include ByteDance[55] and xAI.[58]

Applications

[edit]

Fei-Fei Li views world models as applying to robotics and creative works. Due to the complexity of these models, she advocates for more complex strategies in data acquisition, data engineering, data processing, and synthesizing data.[59] She co-founded a startup on building world models, which, as of 2024, planned to do so in three phases: incorporating an understanding of three-dimensional space along with time; support for augmented reality; and support for robotics.[60] Her startup, World Labs, released its commercial world model, Marble, in November 2025.[61]

World models are intended for use in interactive media (such as video games and movies[62]) and environment simulation.[63] Proposed use cases for world models include action planning and outcome prediction.[61] Other applications include social simulacra to simulate social systems.[41] Wired compared world models to the metaverse,[60] while Business Insider noted possible military applications.[59]

In 2025, world models are being applied to drone warfare, robotics, and self-driving vehicles. The Wall Street Journal speculated that world models could improve spatial reasoning of artificial intelligence models and successfully automate both blue-collar and white-collar jobs.[64] As of October 2025, research has shown mixed results in the spatial reasoning capabilities of text-to-video models (in particular, Veo 3).[65]

Concerns

[edit]

TechCrunch noted that world models could use more data than large language models and would require significantly more computational power (including the use of thousands of GPUs for training and inference).[45][48] It also noted the risk of hallucinations, coverage bias and algorithmic bias.[48] Similarly, The Financial Times noted the difficulty and expense in collecting data to simulate the world and training models to use that data.[58]

Creative professionals have expressed concern that world models could disrupt jobs in their industries.[63]

Other concerns include data privacy,[41] simulation of harmful situations,[41] misinformation and disinformation,[41] emergent behaviors,[66] and copyright.[62]

Technical details

[edit]

Modeling

[edit]

For a foundation model to effectively generalize, it must acquire rich representations of the training data. As a result, expressive model architectures that efficiently process large-scale data are often preferred in building foundation models.[17] Currently, the Transformer architecture is the de facto choice for building foundation models across a range of modalities.[67]

Training

[edit]

Foundation models are built by optimizing a training objective(s), which is a mathematical function that determines how model parameters are updated based on model predictions on training data.[68] Language models are often trained with a next-tokens prediction objective, which refers to the extent at which the model is able to predict the next token in a sequence. Image models are commonly trained with contrastive learning or diffusion training objectives. For contrastive learning, images are randomly augmented before being evaluated on the resulting similarity of the model's representations. For diffusion models, images are noised and the model learns to gradually de-noise via the objective. Multimodal training objectives also exist, with some separating images and text during training, while others examine them concurrently.[69] In general, the training objectives for foundation models promote the learning of broadly useful representations of data.

With the rise of foundation models and the larger datasets that power them, a training objective must be able to parse through internet-scale data for meaningful data points. Additionally, since foundation models are designed to solve a general range of tasks, training objectives ought to be domain complete, or able to solve a broad set of downstream capabilities within the given domain. Lastly, foundation model training objectives should seek to scale well and be computationally efficient. With model size and compute power both being relevant constraints, a training objective must be able to overcome such bottlenecks.

Data

[edit]

Foundation models are trained on a large quantity of data, working under the maxim "the more data, the better."[70] Performance evaluation does show that more data generally leads to better performance, but other issues arise as data quantity grows. Tasks like managing the dataset, integrating data across new applications, ensuring adherence to data licenses, and maintaining data quality all become more difficult as data size grows. The specific demands of foundation models have only exacerbated such issues, as it remains the norm for large foundation models to use public web-scraped data. Foundation models include also search engines data and SEO meta tags data. Public web data remains a plentiful resource, but it also demands stringent moderation and data processing from foundation model developers before it can be successfully integrated into the training pipeline.[71]

Training foundation models often runs the risk of violating user privacy, as private data can be disclosed, collected, or used in ways beyond the stated scope. Even if no private data is leaked, models can still inadvertently compromise security through learned behavior in the resulting foundation model.[72] Data quality is another key point, as web-scraped data frequently contains biased, duplicate, and toxic material. Once foundation models are deployed, ensuring high-quality data is still an issue, as undesirable behavior can still emerge from small subsets of data.

Systems

[edit]

The size of foundation models also brings about issues with the computer systems they run on. The average foundation model is too large to be run within a single accelerator's memory and the initial training process requires an expensive amount of resources.[73] Such issues are predicted to further exacerbate in future as foundation models grow to new heights. Due to this constraint, researchers have begun looking into compressing model size through tight model inference.

GPUs are the most common choice of compute hardware for machine learning, due to high memory storage and strong power. Typical foundation model training requires many GPUs, all connected in parallel with fast interconnects. Acquiring a sufficient amount of GPUs of requisite compute efficiency is a challenge for many foundation model developers, one that has led to an increasing dilemma in the field. Larger models require greater compute power, but often at the cost of improved compute efficiency. Since training remains time-consuming and expensive, the tradeoff between compute power and compute efficiency has led only a few select companies to afford the production costs for large, state of the art foundation models. Some techniques like compression and distillation can make inference more affordable, but they fail to completely shore up this weakness.

Scaling

[edit]

The accuracy and capabilities of foundation models often scale predictably with the size of the model and the amount of the training data. Specifically, scaling laws have been discovered, which are data-based empirical trends that relate resources (data, model size, compute usage) to model capabilities. Particularly, a model's scale is defined by compute, dataset size, and the number of parameters, all of which exhibit a power-law relationship with end performance.

However, broken scaling laws[74] have been discovered in which this relationship smoothly transitions (at points referred to as break(s)) from a power law with one exponent to a power law with another (different) exponent. When one does not collect any points near (or after) the break(s), it can be difficult to obtain an accurate extrapolation.

Adaptation

[edit]

Foundation models are inherently multi-purpose: to use these model for a specific use case requires some form of adaptation. At a minimum, models need to be adapted to perform the task of interest (task specification), but often better performance can be achieved by more extensive adaptation to the domain of interest (domain specialization).

A variety of methods (e.g. prompting, in-context learning, fine-tuning, LoRA) provide different tradeoffs between the costs of adaptation and the extent to which models are specialized. Some major facets to consider when adapting a foundation model are compute budget and data availability. Foundation models can be very large, up to trillions of parameters in size, so adapting the entirety of a foundation model can be computationally expensive. Therefore, developers sometimes adapt only the last neural layer or only the bias vectors to save time and space.[75] For particularly niche applications, specific data may also not be available to adapt the foundation model sufficiently. In such circumstances, data must be manually labeled, which is costly and can demand expert knowledge.

Evaluation

[edit]

Evaluation is a key part of developing foundation models. Not only does evaluation allow for tracking progress of high-performance models, it also creates benchmarks for future model development. Stakeholders rely on evaluations to understand model behaviors and gain insight into their various attributes. Traditionally, foundation models are evaluated relative to each other through standardized task benchmarks like MMLU,[76] MMMU,[77] HumanEval,[78] and GSM8K.[79] Given that foundation models are multi-purpose, increasingly meta-benchmarks are developed that aggregate different underlying benchmarks. Examples include LM-Harness,[80] BIG-Bench,[81] HELM,[82] OpenLLM Leaderboard,[83] DecodingTrust,[84] and HEIM.[85]

Since foundation models' utility depends on their own general capabilities and the performance of fine-tuned applications, evaluation must cover both metrics. Proper evaluation examines both a foundation model's downstream applications in aggregate and the direct properties the foundation model holds. To ensure further equity in evaluation, certain existing evaluation frameworks account for all adaptation resources, which leads to more informed analyses for the benefit of all stakeholders.[86]

Supply chain

[edit]

Foundation models' general capabilities allow them to fulfill a unique role in the AI ecosystem,[87] fueled by many upstream and downstream technologies.[1] Training a foundation model requires several resources (e.g. data, compute, labor, hardware, code), with foundation models often involving immense amounts of data and compute (also referred to as computational power). Due to foundation models' large development costs and inexpensive adaptation requirements, the AI landscape has shifted to a small subset of AI companies making foundation models for downstream adaptation.[88] Thus, most foundation model companies outsource this step to specialized data providers (e.g. Scale AI,[89] Surge[90]) and compute providers (e.g. Amazon Bedrock, Google Cloud, Microsoft Azure).

Investment in computing capabilities to train larger AI models has rapidly increased.[91]

The foundation model developer itself will then take the data and use the supplied compute to actually train the foundation model. After the foundation model is completely built, much of the data and labor requirements abate. In this development process, hardware and compute are the most necessary, and also the most exclusive resources. To train larger and more complex AI, a sufficient amount of compute is key. However, compute is consolidated in the hands of a few, select entities, which most foundation model developers depend on. As such, the foundation model pipeline is concentrated heavily around these providers. Compute is also costly; in 2023, AI companies spent more than 80% of total capital on compute resources.[92]

Foundation models require a large amount of general data to power their capabilities. Early foundation models scraped from subsets of the internet to provide this data information. As the size and scope of foundation models grows, larger quantities of internet scraping becomes necessary, resulting in higher likelihoods of biased or toxic data. This toxic or biased data can disproportionately harm marginalized groups and exacerbate existing prejudices.[93]

To address this issue of low-quality data that arose with unsupervised training, some foundation model developers have turned to manual filtering. This practice, known as data labor, comes with its own host of issues.[94] Such manual data detoxification is often outsourced to reduce labor costs, with some workers making less than $2 per hour.[95]

The foundation model will then be hosted online either via the developer or via an external organization. Once released, other parties can create applications based on the foundation model, whether through fine-tuning or wholly new purposes. People can then access these applications to serve their various means, allowing one foundation model to power and reach a wide audience.

Release strategies

[edit]

After a foundation model is built, it can be released in one of many ways. There are many facets to a release: the asset itself, who has access, how access changes over time, and the conditions on use.[96] All these factors contribute to how a foundation model will affect downstream applications.[97] In particular, the two most common forms of foundation model release are through APIs and direct model downloads.

When a model is released via an API, users can query the model and receive responses, but cannot directly access the model itself. Comparatively, the model could be directly downloadable for users to access and modify. Both release strategies are often classified as an open release. The exact definition of an open release is disputed, but widely accepted requirements are provided by the Open Source Initiative.

Some open foundation models are: PaLM 2, Llama 2, Granite, and Mistral. While open foundation models can further research and development more easily, they are also more susceptible to misuse. Open foundation models can be downloaded by anyone, and particularly powerful models can be fine-tuned to intentionally or unintentionally cause harm.[citation needed]

During a closed release, the foundation model cannot be accessed by the public, but is used internally by an organization. Such releases are considered safer, but offer no additional value to the research community or the public at large.

Some foundation models like Google DeepMind's Flamingo[98] are fully closed, meaning they are available only to the model developer; others, such as OpenAI's GPT-4, are limited access, available to the public but only as a black box; and still others, such as Meta's Llama 2 are open, with broadly available model weights enabling downstream modification and scrutiny.

Practices and applications

[edit]

In practice, foundation models are often embedded into everyday software workflows as general-purpose services for drafting and summarizing text, answering questions, and generating structured output. Microsoft's documentation for Microsoft 365 Copilot describes the system as coordinating large language models to "understand, summarize, predict, and generate content" across applications such as Word, Excel, Outlook, and Teams.[99] In customer support operations, IBM has described using foundation models for automatic call summarization and topic extraction to update customer-relationship-management (CRM) systems and reduce pre- and post-call workload.[100]

Organizations also adapt foundation models to specialized domains in response to constraints such as cost, latency, and governance of sensitive data. In 2024, NeuralFabric co-founder John deVadoss argued that "Foundation models are the new applications," describing domain-specific foundation models as a new metaphor for software and emphasizing issues such as data sovereignty and the cost of training and inference in enterprise deployments.[101][102]

Software development is another prominent application area, where foundation models are used for code generation, refactoring, and multi-step "agentic" assistance. IEEE Spectrum described a competitive market of AI coding tools in which AI-first integrated development environments (IDEs) such as Cursor (a fork of Visual Studio Code) and model-provider tools such as Anthropic's Claude Code both seek to become central to developer workflows.[103] Reporting on Anthropic's Claude Code noted that Anthropic's models already powered third-party coding tools such as Cursor, while describing Claude Code as an "agentic" tool able to search and read code, edit files, write and run tests, and interact with version-control and command-line tooling.[104] Anthropic later released a native Visual Studio Code extension for Claude Code, further integrating a first-party model-provider tool into the IDE environment and overlapping with capabilities offered by standalone AI-first editors.[105]

Claude Code is an agentic coding assistant developed by Anthropic that runs in a command-line interface and is designed to help users carry out software-development tasks via natural-language instructions, including reading and editing project files and executing commands as part of a workflow.[106][107] Comparable tools include GitHub Copilot, which provides code completion and chat-based assistance integrated into development environments and related tooling, as well as Google's Gemini Code Assist and AWS's Amazon Q Developer, which are positioned as generative-AI assistants that support multiple parts of the software development lifecycle.[108][109][110] Such assistants have also been deployed through collaboration-platform integrations; for example, Anthropic introduced a Slack integration that routes tagged messages and thread context to Claude Code as a "research preview".[111]

Although products like Claude Code are often described as programming tools, Anthropic has reported internal use cases that extend into adjacent knowledge-work domains, including automating routine data engineering and operational troubleshooting, and enabling finance staff to execute data workflows described in plain text; the same report describes usage across departments such as marketing and legal.[112] Related model-based assistants have been integrated into mainstream productivity software, including Microsoft 365 Copilot and a Microsoft Excel worksheet function (COPILOT) that invokes an AI language model from within a cell formula for tasks such as summarization, while noting that outputs can be incorrect and are not intended for tasks requiring reproducible accuracy.[113][114] Surveys and systematic reviews have discussed additional applications of foundation-model and large-language-model systems outside software engineering, including healthcare (e.g., diagnostics, personalized treatment, and operational efficiency, alongside privacy and bias concerns) and education (e.g., intelligent tutoring systems, alongside issues such as over-reliance, fairness, and privacy).[115][116]

References

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
A foundation model is any model trained on broad —generally using at scale—that can be adapted to a wide range of downstream tasks. The term was introduced in a 2021 report by researchers at to describe a class of large-scale AI systems exhibiting emergent generalization capabilities across domains like , , and . These models leverage massive datasets and computational resources, often involving billions or trillions of parameters, enabling where pre- on general supports fine-tuning for specialized applications with minimal additional supervision. Key examples include transformer-based models such as OpenAI's GPT series and Google's , which have demonstrated scaling laws where performance predictably improves with increased model size, volume, and compute. While foundation models have accelerated AI advancements—facilitating breakthroughs in tasks from text generation to synthesis—they incur exorbitant costs, frequently exceeding hundreds of millions of dollars, and raise concerns over risks including amplification from uncurated corpora, to adversarial attacks, and potential societal harms from misuse. Their development underscores a concentration of capabilities among resource-rich entities, prompting debates on accessibility, safety, and the empirical limits of scaling without fundamental architectural innovations.

Definition and Characteristics

Core Definition

A foundation model refers to any model trained on broad data, typically using , that can be adapted—through fine-tuning, prompting, or other methods—to a wide range of downstream tasks. This definition emphasizes the model's foundational role in deriving generalizable capabilities from vast, unlabeled datasets rather than bespoke task engineering. Unlike supervised approaches reliant on labeled examples for specific objectives, foundation models leverage emergent properties from data scale, where representations encode patterns enabling versatile application across domains. Key characteristics include massive scale, often encompassing billions to trillions of parameters, which facilitates the compression of diverse data into reusable latent structures. These models exhibit generality across modalities such as text, images, audio, and video, allowing unified processing of heterogeneous inputs through shared pre-training objectives. Adaptation efficiency stems from , where minimal additional data or instructions suffice to specialize the model, contrasting with resource-intensive retraining from scratch. In distinction from narrow AI systems, which are engineered for singular, predefined functions without broad reusability, foundation models achieve capabilities via probabilistic pattern extraction from expansive corpora, yielding causal inferences grounded in data distributions rather than explicit programming. Narrow AI, by contrast, optimizes for isolated performance metrics through targeted , limiting extrapolation to untrained scenarios. This prioritizes empirical scaling laws, where model performance correlates predictably with data volume and compute, over domain-specific heuristics.

Distinguishing Attributes

Foundation models differ from prior AI paradigms, such as task-specific models, through their reliance on massive scale in parameters, data, and compute, enabling emergent abilities that arise discontinuously rather than through gradual performance gains. For instance, in-context learning—where models adapt to new tasks via prompts without parameter updates—emerged sharply in , a 175-billion-parameter model trained on approximately 570 gigabytes of filtered text data and released in June 2020, marking a departure from smaller models' predictable scaling curves. Subsequent analyses confirm these abilities, including few-shot adaptation, correlate empirically with model sizes exceeding 10^9 parameters and training datasets surpassing trillions of tokens, as smaller systems fail to exhibit such behaviors despite similar architectures. This scale-driven emergence underscores a foundational shift: capabilities previously requiring specialized now surface as byproducts of broad pre-training on internet-scale corpora. A core distinguishing attribute is versatility across tasks and modalities without exhaustive retraining, contrasting with traditional machine learning's dependence on curated, labeled datasets for each application. Foundation models undergo initial self-supervised pre-training on diverse, unlabeled data—often billions of examples spanning text, code, and images—allowing subsequent deployment via lightweight prompting or fine-tuning for downstream uses like , summarization, or code generation. Multimodal extensions exemplify this: , introduced by in January 2021, leverages pre-trained text-image alignments to generate images from textual descriptions, adapting foundational representations to vision tasks without starting from scratch, unlike conventional vision models requiring modality-specific training from raw pixels. This adaptability stems from learned latent representations that generalize across domains, though it remains bounded by the distributional coverage of pre-training data. Critically, foundation models' proficiency traces to statistical in observational rather than causal comprehension, highlighting limitations in causal realism absent from many prior paradigms' narrower scopes. They excel at predictive interpolation within training distributions but falter on novel , such as counterfactual reasoning or interventions in unseen graphs, where outputs revert to memorized correlations rather than mechanistic understanding. Empirical probes reveal this gap: even advanced models like struggle with tasks demanding distinction between spurious associations and true causes outside benchmark templates, underscoring that scale amplifies data-driven heuristics without bridging to first-principles . This attribute necessitates caution in applications presuming deeper reasoning, as capabilities reflect probabilistic approximations, not veridical world modeling.

Historical Development

Pre-Foundation Precursors

The Transformer architecture, proposed by Vaswani et al. in June 2017, marked a pivotal shift in by eschewing recurrent and convolutional layers in favor of self-attention mechanisms, which facilitated parallel and captured long-range dependencies more effectively than prior models. This design empirically demonstrated superior performance on tasks, with the model achieving a 28% reduction in score error compared to previous state-of-the-art systems on the WMT 2014 English-to-German dataset, laying the groundwork for scaling to larger datasets and model sizes without the sequential bottlenecks of recurrent neural networks. Building on this, early large-scale pre-training emerged with models like in 2018, which used bidirectional LSTMs trained on unsupervised objectives such as predicting internal word representations from context, enabling contextualized embeddings that improved transfer to six NLP tasks by averaging 4-7 percentage point gains over non-contextual baselines. Similarly, BERT, released by Devlin et al. in October 2018, introduced masked language modeling and next-sentence prediction for pre-training on 3.3 billion words from and , attaining state-of-the-art results on 11 NLP benchmarks like GLUE (80.5% average score) through fine-tuning, thus highlighting self-supervised learning's capacity for broad task adaptation without task-specific supervision from scratch. GPT-2, developed by and detailed in February 2019, further exemplified this trajectory by scaling unsupervised next-token prediction to a 1.5 billion parameter model trained on 40 gigabytes of WebText—a curated of 8 million web pages linked from —yielding coherent text generation and zero-shot performance on tasks like summarization (ROUGE scores competitive with supervised models) and translation, underscoring the viability of purely generative pre-training for emergent capabilities across domains. These pre-2020 efforts collectively demonstrated that large-scale, data-driven pre-training on unlabeled corpora could yield models with transferable representations, departing from the era's dominant paradigm of narrow, supervised architectures and empirically validating scaling as a path to .

Emergence of the Term (2021)

The term "foundation model" was formally introduced in the report On the Opportunities and Risks of Foundation Models, published on August 16, 2021, by researchers at Stanford University's Center for Research on Foundation Models (CRFM). The report, authored by Rishi Bommasani and colleagues including Percy Liang, defined foundation models as "any model trained on broad data (typically by self-supervision at scale) that can be adapted (e.g., fine-tuned) to a wide range of downstream tasks," emphasizing their role as reusable infrastructural bases rather than task-specific systems. This framing positioned models like OpenAI's (released June 2020) and Google's T5 (paper published October 2019, with implementations scaling in 2020) as exemplars, highlighting their capacity for across domains due to massive pre-training on diverse datasets. The motivation for coining the term stemmed from the escalating computational costs of training large-scale models—often exceeding hundreds of millions of dollars—and the recognition that such investments could be amortized through broad reusability, shifting AI development from siloed, narrow applications toward general-purpose foundations adaptable via fine-tuning or prompting. The CRFM report argued this paradigm enabled efficiency gains, as a single foundation model could underpin multiple specialized applications, but also introduced systemic risks like amplified biases from broad data ingestion and challenges in governance due to their infrastructural scale. Initial examples focused on language models, but the concept extended to multimodal systems, underscoring the need for interdisciplinary analysis of their societal implications. Following the report's release, the terminology saw rapid adoption in industry and academia, with organizations like and integrating it to describe their scalable AI architectures. , for instance, began referencing GPT-series models as foundation models in public communications and technical updates by late 2021, aligning with the report's emphasis on pre-trained bases for downstream adaptation. similarly embraced the term for systems like , framing them as foundational layers in cloud AI services to highlight interoperability and cost-sharing potential. This uptake reflected a consensus on the term's utility in capturing the shift toward models prioritizing scale and generality over bespoke training.

Key Milestones and Models (2020-2025)

In June 2020, released , a transformer-based with 175 billion parameters that demonstrated emergent few-shot learning capabilities, enabling task performance with minimal examples provided in prompts without fine-tuning. This marked a pivotal advancement in scaling laws, where larger models showed improved generalization across tasks like and question-answering, though limited by a 2048-token context window. Google's , announced on April 4, 2022, scaled to 540 billion parameters using the Pathways system for efficient distributed training, achieving breakthroughs in reasoning tasks such as arithmetic and commonsense inference through chain-of-thought prompting. In February 2023, Meta released LLaMA, a family of efficient models up to 65 billion parameters with open weights under a research license, which spurred widespread community fine-tuning and democratized access, intensifying competition beyond proprietary systems. The year 2023 saw an explosion in releases, with 149 foundation models documented globally—more than double the 2022 figure—including xAI's Grok-1 base model, whose pre-training concluded in October, emphasizing truth-seeking objectives in a 314 billion parameter mixture-of-experts architecture released openly in March 2024. Of these, 65.7% featured open weights, accelerating innovation through derivative models and efficiency optimizations. In May 2024, launched GPT-4o, a multimodal model integrating text, vision, and audio processing in a unified with a 128,000-token , enabling real-time applications like voice interaction while maintaining performance parity to prior versions at reduced inference costs. By 2025, releases continued apace, exemplified by Meta's LLaMA 4 in April, introducing natively multimodal variants like Scout (17 billion active parameters) with extended lengths, reflecting shifts toward efficiency gains amid sustained scaling in compute and data.

Frontier Models

Frontier models represent the most advanced subset of foundation models, characterized by their in empirical benchmarks and ability to demonstrate emergent capabilities that approach or exceed human performance in targeted domains. These systems are typically defined by high training compute scales—often exceeding 10^25 FLOPs—and broad generality, enabling superior results in reasoning, coding, and multimodal tasks, while introducing heightened risks from potential misuse or unintended behaviors. Unlike standard foundation models, frontier models are distinguished not merely by size but by verifiable outperformance on standardized evaluations, such as achieving scores that rival expert humans, though they remain limited in holistic real-world agency. Frontier models often outperform fine-tuned smaller models overall due to general intelligence advantages, with closed frontier models consistently ranking highest in crowdsourced arenas like the LMArena (formerly LMSYS Chatbot Arena) for reasoning, coding, math, and multifaceted tasks; even large open models like Llama 405B compete closely when fine-tuned, but smaller fine-tuned ones rarely surpass them on broad metrics. Their strength in zero/few-shot prompting, enhanced by techniques like chain-of-thought, frequently matches or exceeds fine-tuning results on smaller base models without custom training. Scaling advantages from larger parameter counts and superior pretraining provide edges in complex, open-ended reasoning. OpenAI's , released on March 14, 2023, exemplifies this category by attaining the 90th percentile on the Uniform Bar Examination, outperforming 90% of human examinees, and scoring in the 90th percentile on SAT reading and math sections. Similarly, Anthropic's Claude 3 family, introduced in March 2024, established new benchmarks in graduate-level reasoning (GPQA), undergraduate knowledge (MMLU), and vision tasks, with the Opus variant leading competitors in coding and multilingual proficiency. Google's Gemini 1.0, announced December 6, 2023, advanced multimodal integration, processing text, images, audio, and video to achieve state-of-the-art results on benchmarks like MMMU for visual reasoning. These models' capabilities stem from massive pre-training on diverse datasets, yielding emergent skills like few-shot learning that were not explicitly optimized. Due to their scale and potency, models carry elevated risk profiles, including amplified potential for adversarial exploitation or systemic impacts, as outlined in guidelines from the U.S. AI Safety Institute established in 2023 under the National Institute of Standards and Technology. The Institute's framework prioritizes rigorous pre-deployment testing and safeguards for models with compute thresholds indicative of advanced risks, emphasizing empirical validation over self-reported claims to address gaps in transparency and safety. This focus underscores causal links between model scale and emergent hazards, such as deceptive alignment or unintended amplification of biases in training data.

General-Purpose AI Systems

Foundation models share substantial conceptual overlap with general-purpose AI systems, frequently treated as synonymous in discourse, as exemplified by the EU AI Act's classification of general-purpose AI (GPAI) models—which encompass foundation models—as adaptable systems trained on extensive datasets to execute diverse tasks across applications without task-specific redesign. This equivalence arises from their broad applicability, yet foundation models distinctly prioritize statistical generality emergent from massive pre-training corpora over explicitly engineered modularity or hybrid architectures that might characterize some general-purpose designs. Empirical assessments reveal foundational constraints on these systems' purported generality, with no demonstrated and pronounced beyond training distributions; for instance, leading models score below 10% on the ARC-AGI benchmark's novel tasks, where humans routinely exceed 80%, indicating reliance on rather than causal understanding or flexible reasoning. Even recent advancements, such as OpenAI's o3 model achieving partial gains on public ARC subsets through enhanced chain-of-thought prompting, fail to close the gap on core challenges, affirming that capabilities remain distributionally bounded without evidence of scalable transfer. Regulatory approaches like the EU AI Act, which impose transparency, documentation, and systemic risk evaluations on GPAI models effective from August 2025, have drawn criticism for presuming unverified existential hazards—such as uncontrolled proliferation—absent causal mechanisms observed in deployed systems, thereby prioritizing speculative threats over documented limitations. Analyses contend that such frameworks, often shaped by precautionary biases in academic and circles, overlook empirical risk profiles favoring iterative competition and open benchmarking to foster verifiable , rather than decelerationist stances that conflate scaling artifacts with apocalyptic inevitability.

World Models and Multimodal Extensions

Foundation models incorporate world models as latent representations that predict environmental dynamics through causal forecasting, enabling internal for and rather than mere . These extensions draw from paradigms, where the model generates hypothetical future states based on actions, facilitating emergent behaviors like in simulated environments. For instance, Google DeepMind's Genie 3, introduced in August 2025, advances real-time interactive world modeling by generating consistent video predictions from latent states, supporting applications in game-like without explicit physics engines. However, empirical evaluations reveal that such models often prioritize statistical correlations over invariant causal structures, leading to brittle generalizations outside training distributions. In robotics, world models integrate with action primitives for grounded planning, as demonstrated by Google DeepMind's , a vision-language-action model released in July 2023. co-fine-tunes on internet-scale vision-language data and robotic trajectories, achieving up to 2x success rates on novel tasks like using objects as improvised tools through chain-of-thought reasoning over predicted outcomes. This causal prediction mechanism allows transfer of web-derived knowledge to physical control, improving manipulation in unseen scenarios by simulating action effects. Yet, critiques highlight deficiencies in encoding fundamental physical laws; a 2025 Harvard-MIT study found that foundation models, including world model variants, accurately predict outcomes in tested cases but fail to internalize principles like Newton's laws, relying instead on memorized heuristics that break under counterfactual perturbations. Multimodal extensions enhance world models by fusing modalities like vision and , promoting grounded reasoning through aligned representations. OpenAI's CLIP, pretrained in 2021 on 400 million image-text pairs via contrastive learning, establishes zero-shot cross-modal correspondences that anchor textual predictions to visual evidence, reducing hallucinations in simulation tasks. Subsequent integrations, such as in FOUNDER frameworks, map foundation model outputs to world model latents for open-ended task solving, yielding improved planning in embodied settings. Achievements include enhanced robotic control, with exhibiting emergent skills like semantic reasoning for object affordances. Nonetheless, these systems inherit data biases from curated corpora, amplifying representational skews—e.g., underrepresentation of diverse physical interactions—that propagate to causal predictions, as biases in training data lead to skewed outcome distributions rather than veridical simulations. True adherence to physical invariance remains elusive, with models critiqued for simulating superficial dynamics without underlying causal realism.

Technical Architecture

Core Architectures and Parameters

The transformer architecture, introduced in the 2017 paper "Attention Is All You Need," underpins the majority of foundation models through its self-attention mechanisms, which compute dependencies between sequence elements in parallel, eliminating the sequential processing constraints of recurrent neural networks like LSTMs. This design consists of stacked encoder and decoder layers, though many modern foundation models, such as those in the GPT series, employ decoder-only variants optimized for autoregressive generation. Self-attention enables efficient handling of long-range dependencies via scaled dot-product attention, formulated as Attention(Q,K,V)=softmax(QKTdk)V\text{Attention}(Q, K, V) = \text{softmax}\left(\frac{QK^T}{\sqrt{d_k}}\right)V
Add your contribution
Related Hubs
User Avatar
No comments yet.