Skip to content

Introduction to Foundation Models

What Is a Foundation Model

In 2021, Stanford HAI published the landmark report "On the Opportunities and Risks of Foundation Models", formally introducing the concept of the Foundation Model. It is defined as follows:

A foundation model is any model that is trained on broad data (generally using self-supervision at scale) that can be adapted to a wide range of downstream tasks.

The core idea: a large-scale model pretrained on massive data that can be transferred to a variety of downstream tasks through fine-tuning, prompting, and other adaptation methods.


Key Characteristics

Foundation Models exhibit the following key characteristics:

Characteristic Description
Scale Parameter counts ranging from hundreds of millions to trillions; training data from terabytes to petabytes
Generality Not designed for a single task, but capable of classification, generation, reasoning, retrieval, and more
Emergence Beyond a certain scale, the model exhibits abilities that were not explicitly optimized during training
Transfer Pretrained representations can be efficiently transferred to downstream tasks, reducing the need for labeled data
Homogenization Different tasks share the same foundation model, and technical approaches converge

Compared to traditional task-specific models, Foundation Models represent a fundamental paradigm shift:

Traditional:  One task → One model → One dataset
Foundation:   One base model → Many tasks → Unified representations

Historical Evolution

The development of Foundation Models can be divided into the following stages:

Stage 1: The Dawn of Distributed Representations (2013-2017)

  • Word2Vec (Mikolov et al., 2013): The first use of neural networks to learn distributed word representations
  • GloVe (2014): A word vector method based on co-occurrence matrices
  • Core idea: Mapping discrete symbols into a continuous vector space
\[ \text{Word2Vec}: \quad P(w_t | w_{t-k}, \ldots, w_{t-1}) = \text{softmax}(W \cdot h) \]

Stage 2: The Breakthrough of Contextual Representations (2018)

  • ELMo (Peters et al., 2018): Used bidirectional LSTMs to generate context-dependent word representations
  • Key innovation: The same word receives different vector representations in different contexts
  • Limitation: Based on LSTMs, making it difficult to capture long-range dependencies

Stage 3: The Pretrain + Fine-tune Paradigm (2018-2019)

  • GPT-1 (Radford et al., 2018): First application of the Transformer decoder for language model pretraining
  • BERT (Devlin et al., 2018): Masked Language Model + Next Sentence Prediction
  • Established the standard "pretrain + finetune" pipeline

Stage 4: Large-Scale Language Models (2020-2022)

  • GPT-3 (Brown et al., 2020): 175B parameters, demonstrating in-context learning capabilities
  • PaLM (Google, 2022): 540B parameters, the first systematic study of emergent abilities
  • Chinchilla (Hoffmann et al., 2022): Proposed compute-optimal Scaling Laws

Stage 5: Alignment and Instruction Following (2022-2023)

  • InstructGPT (Ouyang et al., 2022): RLHF to better align models with human instructions
  • ChatGPT (OpenAI, 2022): A milestone in conversational AI, bringing large models into mainstream applications

Stage 6: Multimodality and General Intelligence (2023-Present)

  • GPT-4 (OpenAI, 2023): Multimodal input with significantly improved reasoning capabilities
  • GPT-4o (OpenAI, 2024): Natively multimodal (text, image, audio)
  • Gemini (Google, 2024): A natively multimodal Foundation Model
Evolution roadmap:

Word2Vec → ELMo → GPT-1/BERT → GPT-3 → InstructGPT → ChatGPT → GPT-4 → GPT-4o/Gemini
  词向量     上下文    预训练       规模化     对齐          对话       多模态     全模态

Paradigm Shifts

The development of deep learning has undergone five major paradigm shifts:

1. Feature Engineering Era

Manually designed features (e.g., SIFT, HOG) fed into shallow classifiers such as SVMs.

2. Representation Learning Era

Deep networks learn features automatically — exemplified by the breakthrough of CNNs on ImageNet.

3. Pretrain + Fine-tune

Pretrain on large-scale unlabeled data, then fine-tune on a small amount of labeled data.

\[ \theta^* = \arg\min_\theta \mathcal{L}_{\text{downstream}}(f_\theta(x), y) \quad \text{其中 } \theta_0 \text{ 来自预训练} \]

4. Pretrain + Prompt

Instead of fine-tuning model parameters, design prompts to elicit the model's existing capabilities.

5. In-context Learning

The model completes tasks directly from a few examples provided in context, with no parameter updates required.

Paradigm evolution:

特征工程 → 表示学习 → Pretrain+Finetune → Pretrain+Prompt → In-context Learning
 (人工)      (自动)      (迁移学习)          (不动参数)         (零参数更新)

Core trend: Progressively less human intervention, progressively greater generality.


Emergent Abilities

Emergent abilities refer to capabilities that suddenly appear once a model exceeds a certain critical scale — capabilities that are absent in smaller models.

Definition

The formal characterization given by Wei et al. (2022):

An ability is emergent if it is not present in smaller models but is present in larger models.

Key characteristic: Emergence is not a gradual improvement but a sharp, step-function-like transition.

Typical Examples of Emergent Abilities

Ability Description Approximate Scale of Emergence
Few-shot Learning Completing new tasks from just a few examples ~10B parameters
Chain-of-Thought (CoT) Step-by-step reasoning for multi-step math/logic problems ~100B parameters
Instruction Following Understanding and executing natural language instructions ~10B+ parameters
Code Generation Generating code from natural language descriptions ~100B parameters

Chain-of-Thought Example

Standard prompting:

Q: Roger has 5 tennis balls. He buys 2 more. How many does he have?
A: 7

CoT prompting:

Q: Roger has 5 tennis balls. He buys 2 more. How many does he have?
A: Roger started with 5 balls. He bought 2 more. 5 + 2 = 7. The answer is 7.

By guiding the model to show intermediate reasoning steps, accuracy on complex problems improves significantly.

The Debate Around Emergence

Schaeffer et al. (2023) argued that emergence may be an artifact of evaluation metric choice (e.g., using nonlinear metrics) rather than a genuine qualitative change in the model. This debate remains ongoing.


Limitations and Challenges

Despite their impressive capabilities, Foundation Models still face numerous challenges:

1. Hallucination

Models generate content that appears plausible but is factually incorrect. This is one of the most serious issues with current large models.

2. Bias

Social biases present in training data are learned and amplified by the model, leading to unfair outputs along dimensions such as gender and race.

3. Computational Cost

  • The training cost of GPT-4 is estimated to exceed $100 million
  • Inference-time compute demands are also enormous
  • High costs limit the democratization of research

4. Interpretability

Foundation Models are essentially black boxes, making it difficult to understand their internal decision-making processes.

5. Safety and Alignment

Ensuring that model behavior conforms to human values and intentions remains a core open problem. See the Safety and Alignment section for details.

Large-scale training data may contain copyrighted content or personal information, raising legal and ethical concerns.


Summary

Foundation Models represent a major paradigm shift in AI development. The core idea is:

Pretrain a general-purpose model on massive data, then adapt it to specific tasks through various methods.

The success of this paradigm rests on three pillars:

  1. Data: Web-scale, high-quality training data
  2. Compute: Large-scale distributed training infrastructure
  3. Algorithms: The Transformer architecture combined with self-supervised learning objectives

Foundation Model research continues to advance rapidly, expanding from language to vision, multimodal systems, embodied intelligence, and beyond — with the ultimate goal of building truly general artificial intelligence systems.


评论 #