Skip to main content
SGLang Documentation home page
Search...
⌘K
Ask AI
Search...
Navigation
Multi-Node Deployment
Multi-Node Deployment
Get Started
User Guide
Hardware
Cookbook
SGLang Diffusion
Basic Usage
Basic Usage
OpenAI-Compatible APIs
Ollama-Compatible API
Offline Engine API
SGLang Native APIs
Sampling Parameters
Popular Model Usage
Advanced Features
Advanced Features
Server Arguments
Hyperparameter Tuning
Attention Backend
Speculative Decoding
Structured Outputs
Structured Outputs For Reasoning Models
Tool Parser
Reasoning Parser
Quantization
Quantized KV Cache
Expert Parallelism
LoRA Serving
PD Disaggregation
EPD Disaggregation
Pipeline Parallelism for Long Context
Hierarchical KV Caching (HiCache)
Query VLM with Offline Engine
DP for Multi-Modal Encoder in SGLang
Cuda Graph for Multi-Modal Encoder in SGLang
SGLang Model Gateway
Deterministic Inference
Observability
Checkpoint Engine Integration
SGLang for RL Systems
Supported Models
Supported models
Text Generation
Retrieval and Ranking
Specialized Models
Extending SGLang
Developer Guide
Developer Guide
Contribution Guide
Development
Benchmarking
Evaluating New Models with SGLang
References
References
Troubleshooting and Frequently Asked Questions
Environment Variables
Production Metrics
Production Request Tracing
Multi-Node Deployment
Multi-Node Deployment
Multi-Node Deployment
Deploy On Kubernetes
LWS Based PD Deploy
DeepSeekV32-Exp RBG Based PD Deploy
Custom Chat Template
Frontend Language
Cookbook
Post-Training Integration
Multi-Node Deployment
Multi-Node Deployment
Copy page
Copy page
Multi Node
Deploy On K8S
Lws Pd Deploy
Deepseekv32 Pd
Deploying DeepSeek with PD Disaggregation on 96 H100 GPUs
Deploying Kimi K2 with PD Disaggregation on 128 H200 GPUs
Production Request Tracing
Previous
Multi-Node Deployment
Next
⌘I
Assistant
Responses are generated using AI and may contain mistakes.