Running real inference at scale? Apply for our limited $10K credit program —
Find out more
Friendli Docs home page
Search...
⌘K
Dashboard
Dashboard
Search...
Navigation
Page Not Found
Documentation
API Reference
Models
Friendli Suite Guide
Tutorial
Changelog
Website
Blog
Get Started
Overview
Supported Models
OpenAI Compatibility
Capabilities
Tool Calling
Structured Outputs
Reasoning
Multi‑modality
Friendli Dedicated Endpoints
Introduction
QuickStart
Plans & Pricing
Models
Endpoints
Autoscaling
Online Quantization
Speculative Decoding
Multi-LoRA Serving
Versioning
Dataset
Beta
FAQs
Friendli Serverless Endpoints
Introduction
QuickStart
Plans & Pricing
Integrations
Tool Assisted API
Beta
Friendli Container
Introduction
QuickStart
Running Friendli Container
CUDA Compatibility
Quantized Models Serving
Multi-LoRA Serving
MoE Models Serving
Optimizing Inference with Policy Search
SageMaker Integration
Inference with gRPC
Monitoring
404
Page Not Found
We couldn't find the page you were looking for. Maybe you were looking for?
Running Friendli Container
CUDA Compatibility
QuickStart: Friendli Container Trial
Assistant
Responses are generated using AI and may contain mistakes.