Getting Started#
- Installation
- Using Xinference
- Logging in Xinference
- Xinference Docker Image
- Xinference on Kubernetes
- Troubleshooting
- No huggingface repo access
- Incompatibility Between NVIDIA Driver and PyTorch Version
- Xinference service cannot be accessed from external systems through
<IP>:9997
- Launching a built-in model takes a long time, and sometimes the model fails to download
- When using the official Docker image, RayWorkerVllm died due to OOM, causing the model to fail to load
- Missing
model_engine
parameter when launching LLM models
- Environments Variables