AI Wisdom
☁️

Cloud AI Platforms

Managed AI inference services, GPU clouds, and model hosting platforms.

Graduated · 3Incubating · 6Sandbox · 110 total
← All categories

Azure OpenAI Service

Graduated
5/5

Enterprise GPT models with Azure compliance, RBAC, and private networking

The enterprise standard for deploying OpenAI models. VNet integration, managed identity, content filtering, and regional deployment. Best for regulated industries needing SOC 2, HIPAA, and GDPR.

Managed

AWS Bedrock

Graduated
4/5

Multi-model serverless AI service with Claude, Llama, and more

Best multi-model platform — access Claude, Llama, Mistral, and Titan from one API. Serverless with no infrastructure to manage. Knowledge bases and agents built in.

Managed

Google Vertex AI

Graduated
4/5

GCP's unified AI platform with Gemini, tuning, and evaluation

Full-stack AI platform with native Gemini access, Model Garden for 100+ models, and built-in evaluation. Best for GCP shops and teams using BigQuery for data. Strong MLOps tooling.

Managed

Together AI

Incubating
5/5

Fastest open-source model inference with fine-tuning support

Best platform for running open-source models at scale. Competitive pricing, fast inference, and one-click fine-tuning. Strong Llama, Mistral, and FLUX support. Developer-friendly API.

Managed

Groq

Incubating
5/5

Ultra-fast LPU inference — 10× faster than GPU-based alternatives

Fastest AI inference available — custom LPU hardware delivers tokens at unprecedented speed. Great for latency-sensitive applications. Limited model selection but growing.

Managed

Fireworks AI

Incubating
4/5

Fast model serving with function calling and compound AI systems

Excellent for compound AI systems — fast function calling, grammar-constrained generation, and JSON mode. FireAttention delivers top-tier speed. Good for agentic applications.

Managed

Replicate

Incubating
4/5

Run any ML model via API with one-click deployment

Easiest way to run open-source models. Huge model library, pay-per-second billing, and Cog packaging for custom models. Great for prototyping and small-to-medium production workloads.

Managed

HF Inference Endpoints

Incubating
4/5

Deploy any Hugging Face model to dedicated infrastructure

Seamless deployment of any HF Hub model to dedicated GPU instances. Auto-scaling, custom containers, and VPC support. Best for teams already in the Hugging Face ecosystem.

Managed

Anyscale

Incubating
3/5

Ray-based platform for scalable AI workloads and serving

Best for teams already using Ray. Managed Ray clusters for training, fine-tuning, and serving. Supports any model. Good for complex pipelines needing distributed computing.

Managed

Lambda Cloud

Sandbox
3/5

GPU cloud built for AI — H100 and A100 clusters on demand

Competitive GPU pricing with AI-focused networking and storage. Good for training runs and self-hosted inference. Less managed services than hyperscalers but lower cost per GPU-hour.

Managed