Claude Haiku
FamilyAnthropic · Claude
Anthropic's fastest Claude line for latency-sensitive, high-volume, and cost-constrained workloads.
fast efficient coding vision tool-use cost-effective model-family
Updated March 6, 2026
Overview
This is a model family overview. For version-specific details, see the individual model entries linked below.
Claude Haiku is the speed-and-cost tier in Anthropic’s lineup. It sacrifices some capability depth compared with Sonnet and Opus but delivers significantly faster responses at a fraction of the price. Haiku is the right choice when throughput, latency, or cost matter more than peak reasoning quality.
Current Latest
Claude Haiku 4.5 remains the efficiency-oriented current reference in this repo snapshot.
Strengths
- Very fast response times — well suited for real-time applications
- Lowest per-token cost in the Claude family
- Surprisingly capable for its tier — handles many tasks that previously required larger models
- Multimodal support (images and text)
- Tool use and structured output support
When to Choose Haiku
Choose Haiku when speed or cost is the primary concern:
- High-volume classification, extraction, and routing tasks
- Real-time chat applications where latency matters most
- Bulk content processing pipelines
- Development prototyping before upgrading to Sonnet/Opus for production
- Cost-sensitive applications that can tolerate occasional quality drops on hard tasks
Access
- Anthropic API (direct)
- AWS Bedrock
- Google Vertex AI
- Claude consumer products (often the default for quick interactions)