Skip to content
AI Workshop Switzerland
Contact

Foundation Models

One model, infinite applications.

AI Workshop Team

A Foundation Model is a large-scale AI model trained on a vast amount of data (often at internet scale) that can be adapted to a wide range of downstream tasks. They represent a paradigm shift from task-specific models to general-purpose engines.

01

What makes a model a "Foundation Model"?

It must be broadly capable. Unlike previous models designed for one task (e.g., sentiment analysis), a foundation model can write poetry, debug code, translate languages, and summarize text, all without specific retraining.

02

What is "Emergence"?

Foundation models exhibit emergence—capabilities that were not explicitly trained for. For example, a model trained simply to predict the next word in a sentence might emerge with the ability to translate languages, write code, or solve logic puzzles.

03

How are they built?

The lifecycle involves two stages:

  1. Pre-training: The expensive, compute-intensive phase where the model learns general patterns from massive datasets (e.g., "learning to read and write").
  2. Fine-tuning: The adaptation phase where the model is specialized for a specific task or behavior (e.g., "learning to be a helpful assistant").
04

Leading Foundation Models

Prominent foundation models include the GPT series (OpenAI), BERT (Google), Claude (Anthropic), and Stable Diffusion (Stability AI).

Frequently Asked Questions

Are Foundation Models the same as LLMs?

LLMs (Large Language Models) are a *type* of foundation model focused on text. But foundation models can also be multimodal, handling images, audio, and video.

What are the risks?

Because they are trained on internet data, they can inherit biases and toxic content. They can also 'hallucinate' (make things up) confidently.

Up Next

Ready to Deepen Your Understanding?

Join our hands-on workshops to master these concepts and apply them to real-world problems.