Foundation Model
A large-scale AI model trained on broad data that can be adapted to a wide range of downstream tasks.
Foundation models are large, general-purpose models trained on diverse data at scale. The term, coined by Stanford's Center for Research on Foundation Models, captures the idea that these models serve as a "foundation" that can be fine-tuned, prompted, or otherwise adapted for countless specific applications.
Examples include GPT-4, Claude, LLaMA, and Gemini for language; CLIP and DALL-E for vision-language; and Whisper for speech. These models demonstrate emergent capabilities — abilities that appear at scale but weren't explicitly trained for.
The foundation model paradigm has shifted AI from task-specific model development to adaptation and application building. Companies increasingly build on top of existing foundation models rather than training from scratch, creating demand for engineers who understand how to effectively leverage, customize, and deploy these models.
Related AI Job Categories
Related Terms
Large Language Model (LLM)
A neural network trained on massive text datasets that can understand and generate human language.
Fine-Tuning
The process of further training a pre-trained AI model on domain-specific data to improve its performance on particular tasks.
Transfer Learning
A technique where a model trained on one task is reused as the starting point for a model on a different but related task.
Generative AI
AI systems that create new content — text, images, code, audio, or video — based on patterns learned from training data.