Skip to main content
SGLang Documentation home page
Search...
⌘K
Ask AI
Search...
Navigation
Advanced Features
Advanced Features
Get Started
User Guide
Hardware
Cookbook
SGLang Diffusion
Basic Usage
Basic Usage
OpenAI-Compatible APIs
Ollama-Compatible API
Offline Engine API
SGLang Native APIs
Sampling Parameters
Popular Model Usage
Advanced Features
Advanced Features
Server Arguments
Hyperparameter Tuning
Attention Backend
Speculative Decoding
Structured Outputs
Structured Outputs For Reasoning Models
Tool Parser
Reasoning Parser
Quantization
Quantized KV Cache
Expert Parallelism
LoRA Serving
PD Disaggregation
EPD Disaggregation
Pipeline Parallelism for Long Context
Hierarchical KV Caching (HiCache)
Query VLM with Offline Engine
DP for Multi-Modal Encoder in SGLang
Cuda Graph for Multi-Modal Encoder in SGLang
SGLang Model Gateway
Deterministic Inference
Observability
Checkpoint Engine Integration
SGLang for RL Systems
Supported Models
Supported models
Text Generation
Retrieval and Ranking
Specialized Models
Extending SGLang
Developer Guide
Developer Guide
Contribution Guide
Development
Benchmarking
Evaluating New Models with SGLang
References
References
Troubleshooting and Frequently Asked Questions
Environment Variables
Production Metrics
Production Request Tracing
Multi-Node Deployment
Custom Chat Template
Frontend Language
Cookbook
Post-Training Integration
Advanced Features
Advanced Features
Copy page
Advanced configuration, optimization, and deployment features for SGLang.
Copy page
Server Arguments
Hyperparameter Tuning
Attention Backend
Speculative Decoding
Structured Outputs
Quantization
Expert Parallelism
LoRA
PD Disaggregation
Pipeline Parallelism
HiCache
Observability
And more…
Llama4 Usage
Previous
Server Arguments
Next
⌘I
Assistant
Responses are generated using AI and may contain mistakes.