Back to glossary

Zero-Shot Learning

The ability of a model to perform a task it was not explicitly trained on, using only a natural language description of the task without any task-specific examples.

Zero-shot learning is one of the most remarkable capabilities of large language models. You can ask an LLM to classify sentiment, extract entities, translate languages, or summarize documents without providing a single example of the desired output. The model generalizes from its broad pre-training to handle novel tasks based solely on the instruction.

This capability emerges from scale. Models trained on trillions of tokens have encountered enough diverse text to develop a general understanding of tasks described in natural language. When you prompt "Classify this review as positive or negative," the model draws on patterns from millions of similar classification contexts in its training data.

For product teams, zero-shot learning is transformative because it enables rapid prototyping. You can test an AI feature in hours rather than weeks, since there is no training data to collect or model to fine-tune. The trade-off is that zero-shot performance is typically 10-30% lower than few-shot or fine-tuned approaches. The practical strategy is to launch with zero-shot, measure quality, and invest in examples or fine-tuning only when zero-shot quality is insufficient for your use case.

Related Terms