- Dec 15, 2025
- 5 min read
Enabling the Next Level of Efficient Agentic AI: FriendliAI Supports NVIDIA Nemotron 3 Nano Launch
Read full article

- March 15, 2026
- 7 min read
Integrating FriendliAI with OpenClaw
LLM Inference
GLM-5
Qwen3-30B-A3B

- March 13, 2026
- 7 min read
Your Coding Agent is Only as Fast as Your Model API
LLM Inference
GLM-5
MiniMax-M2.5

- March 12, 2026
- 4 min read
FriendliAI Launches InferenceSense™ to Monetize Idle GPU Capacity
InferenceSense

- March 11, 2026
- 3 min read
Nemotron 3 Super is Live on FriendliAI: multi-agent applications and for specialized agentic AI systems
Nemotron
NVIDIA

- March 4, 2026
- 3 min read
Serving GLM-5 at Scale: Why Inference Infrastructure Now Defines Model Capability
GLM-5
Serving
Inference

- February 11, 2026
- 3 min read
GLM-5: The Open-Source Model for Production-Grade Coding Agents
GLM-5
Z.ai
Partnership

- January 4, 2026
- 7 min read
Rethinking AI Inference Kubernetes Cluster Consistency with Atomic State Reconciliation
Reliability
Kubernetes
Atomic State Reconciliation

- December 31, 2025
- 2 min read
K-EXAONE Is Now Available on Friendli Serverless Endpoints
LG AI Research
K-EXAONE
Partnership

- December 30, 2025
- 3 min read
Serverless vs. Dedicated AI Inference: Choosing the Right Friendli Endpoint for Your Workload
Serverless Endpoints
Dedicated Endpoints
Inference