Free
Explore AI using serverless pre-trained models in you cloud or your local dev environments.
- Up to 1K API Calls/month
- Pre-trained models
- 1 Request per Second
- Community support
Essential
Core AI development capabilities across our cloud or your self-hosted compute.
- Up to 30K API Calls/month
- Local development model containers
- Upload any model
- Fine-tune specialist models
- Dedicated GPU clusters with up to 2 nodes
- NVIDIA A10 and L4 GPUs
- 30 Requests per Second
- Email support
Professional
Development and production AI workloads, more advanced dedicated GPUs and Control Center.
Everything from Essential, and:
- Up to 100K API Calls/month
- Train & fine-tune models
- Dedicated GPU clusters with up to 5 nodes
- NVIDIA A10, L4 and L40S GPUs
- 100 Requests per Second
- Control center for model observability
- Email support
Hybrid-Cloud AI Enterprise
Unlimited SaaS AI development and production workloads.
- Unlimited API calls
- Clarifai’s SaaS control plane
- Multi-cloud and multi-region compute planes
- Self-managed cloud and bare-metal compute planes
- NVIDIA A10, L4, L40S, A100 and H100 GPUs
- Private data planes
- Unlimited GPU and CPU nodes
- GPU and CPU spot instances
- Full model exports
- Model leaderboards
- Custom rate limits
- Enterprise 99.99% SLAs
- 24/7 enterprise support
- Dedicated account manager
Private Hybrid-Cloud AI Enterprise
Hybrid AI enterprise deployments for regulated industries.
Everything from Hybrid-Cloud AI Enterprise, and:
- Private hybrid cloud deployments
- Virtual Private Cloud, Bare-metal or Air-Gapped control plane
- Multi-cloud and multi-region compute planes
- Private data planes
- Unlimited GPU and CPU nodes
- Private APM and observability integrations
- Enterprise 99.99% SLAs
- 24/7 enterprise support
Compare Plans
Compare features and benefits across every plan.
Free | Essential | Professional | Hybrid AI Enterprise | Private AI Enterprise | |
---|---|---|---|---|---|
Usage & limits | |||||
Monthly requests | Limited | 30,000 | 100,000 | Unlimited | Unlimited |
Requests per second | 1 | 15 | 100 | 1000+ | 1000+ |
SDK & API access | |||||
Compute | |||||
Deployment types | SaaS, Local Dev |
+ Hybrid Cloud (Self-Hosted) |
+ Hybrid Cloud (Self-Hosted) |
+ Hybrid Cloud (Self-Hosted) |
+ VPC, On-Prem, Air Gapped |
NVIDIA GPUs | Serverless | A10G, L4 | + L40S | + A100, H100 | + A100, H100 |
Intel & AMD CPUs | |||||
Inference | |||||
Pre-trained model access | |||||
Batch requests | |||||
Realtime bi-directional streaming | |||||
GPU fractioning | |||||
Scale to zero | |||||
Spot instances | |||||
Development & training | |||||
Custom model training | Fine-tune | Train & deploy | Full training | Enterprise AI | |
Model evaluation | |||||
Model upload | |||||
Model export | |||||
Dataset management | |||||
Vector search | |||||
Automated data labeling |
Inference Pricing
Over 500 leading open-source and closed-source language, multimodal, image, code, and embedding models are available for serverless and dedicated inference.
Dedicated Node Pricing
Only pay for the compute you use, down to the minute
Best-in-class model performance, effortless autoscaling, and blazing fast cold starts mean you get the most out of each GPU, saving money along the way.
Amazon Web Services
Google Cloud
Vultr

us-east-1
us-west-2
us-east-4
new-york
Node name | Cloud instance name | Price per min |
---|---|---|
NVIDIA T4 16GB XL | g4dn.xlarge | $0.011 |
NVIDIA A10G 24GB XL | g5.xlarge | $0.021 |
NVIDIA A10G 24GB 2XL | g5.2xlarge | $0.0253 |
NVIDIA L4 24GB XL | g6.xlarge | $0.0168 |
NVIDIA L4 24GB 2XL | g6.2xlarge | $0.0204 |
NVIDIA L40S 48GB XL | g6e.xlarge | $0.0388 |
NVIDIA L40S 192GB 12XL | g6e.12xlarge | $0.2186 |
AMD EPYC 7000 M | t3a.medium | $0.0012 |
AMD EPYC 7000 L | t3a.large | $0.0016 |
AMD EPYC 7000 XL | t3a.xlarge | $0.0031 |
AMD EPYC 7000 2XL | t3a.2xlarge | $0.0063 |
NVIDIA L4 24GB XL | g2-standard-4 | $0.0147 |
NVIDIA L4 24GB 2XL | g2-standard-8 | $0.0178 |
NVIDIA L4 24GB 3XL | g2-standard-12 | $0.0208 |
NVIDIA L4 24GB 4XL | g2-standard-16 | $0.0239 |
NVIDIA L4 24GB 5XL | g2-standard-32 | $0.0361 |
NVIDIA A100 80GB XL | a2-ultragpu-1g | $0.1189 |
NVIDIA H100 80GB XL | a3-highgpu-1g | $0.2304 |
Intel ICL/CSL CPU 8GB | n2-standard-2 | $0.0023 |
Intel ICL/CSL CPU 16GB | n2-standard-4 | $0.0046 |
Intel ICL/CSL CPU 32GB | n2-standard-8 | $0.0091 |
Intel ICL/CSL CPU 64GB | n2-standard-16 | $0.0182 |
NVIDIA T4 16GB XL | g4dn.xlarge | $0.011 |
NVIDIA A10G 24GB XL | g5.xlarge | $0.021 |
NVIDIA A10G 24GB 2XL | g5.2xlarge | $0.0253 |
NVIDIA L4 24GB XL | g6.xlarge | $0.0168 |
NVIDIA L4 24GB 2XL | g6.2xlarge | $0.0204 |
NVIDIA L40S 48GB XL | g6e.xlarge | $0.0388 |
NVIDIA L40S 192GB 12XL | g6e.12xlarge | $0.2186 |
AMD EPYC 7000 M | t3a.medium | $0.0012 |
AMD EPYC 7000 L | t3a.large | $0.0016 |
AMD EPYC 7000 XL | t3a.xlarge | $0.0031 |
AMD EPYC 7000 2XL | t3a.2xlarge | $0.0063 |
NVIDIA A16 M 16GB | vcg-a16-6c-64g-16vram | $0.0098 |
NVIDIA L40S XL 48GB | vcg-l40s-16c-180g-48vram | $0.0348 |
NVIDIA A100 XL 80GB | vcg-a100-12c-120g-80vram | $0.0499 |
Frequently Asked Questions
Each plan provides a monthly credit that can be applied toward various operations, such as model predictions, training, and data storage. If your usage exceeds the included credit, additional charges will apply based on the specific operations performed.
Each plan has a maximum number of operations and inputs that can be stored monthly. The Essential Plan allows up to 30,000 operations, while the Professional Plan permits up to 100,000. To exceed these operations, contact sales@clarifai.com for a custom package that includes committed volume discounts.
You can cancel your subscription any time by downgrading to the Community plan.
Clarifai has a team of AI experts that will help you implement AI projects. We can also offer custom development, depending on the project's need.
Please contact us to share more about your project.
Need services? Contact us for professional services ranging from AI Sprints to Custom AI development.

