
Xorbits Inference: Model Serving Made Easy 
English | 中文介绍 | 日本語
Xorbits Inference(Xinference) is a powerful and versatile library designed to serve language, speech recognition, and multimodal models. With Xorbits Inference, you can effortlessly deploy and serve your or state-of-the-art built-in models using just a single command. Whether you are a researcher, developer, or data scientist, Xorbits Inference empowers you to unleash the full potential of cutting-edge AI models.
Hot Topics
Framework Enhancements
- Support Continuous batching for Transformers engine: #1724
- Support MLX backend for Apple Silicon chips: #1765
- Support specifying worker and GPU indexes for launching models: #1195
- Support SGLang backend: #1161
- Support LoRA for LLM and image models: #1080
- Support speech recognition model: #929
- Metrics support: #906
New Models
- Built-in support for Qwen 2.5 Series: #2325
- Built-in support for Fish Speech V1.4: #2295
- Built-in support for DeepSeek-V2.5: #2292
- Built-in support for Qwen2-Audio: #2271
- Built-in support for Qwen2-vl-instruct: #2205
- Built-in support for MiniCPM3-4B: #2263
- Built-in support for CogVideoX: #2049
- Built-in support for flux.1-schnell & flux.1-dev: #2007
Integrations
- Dify: an LLMOps platform that enables developers (and even non-developers) to quickly build useful applications based on large language models, ensuring they are visual, operable, and improvable.
- FastGPT: a knowledge-based platform built on the LLM, offers out-of-the-box data processing and model invocation capabilities, allows for workflow orchestration through Flow visualization.
- Chatbox: a desktop client for multiple cutting-edge LLM models, available on Windows, Mac and Linux.
- RAGFlow: is an open-source RAG engine based on deep document understanding.
Key Features
Model Serving Made Easy: Simplify the process of serving large language, speech
recognition, and multimodal models. You can set up and deploy your models
for experimentation and production with a single command.
State-of-the-Art Models: Experiment with cutting-edge built-in models using a single
command. Inference provides access to state-of-the-art open-source models!
Heterogeneous Hardware Utilization: Make the most of your hardware resources with
ggml. Xorbits Inference intelligently utilizes heterogeneous
hardware, including GPUs and CPUs, to accelerate your model inference tasks.
Flexible API and Interfaces: Offer multiple interfaces for interacting
with your models, supporting OpenAI compatible RESTful API (including Function Calling API), RPC, CLI
and WebUI for seamless model management and interaction.
Distributed Deployment: Excel in distributed deployment scenarios,
allowing the seamless distribution of model inference across multiple devices or machines.
Built-in Integration with Third-Party Libraries: Xorbits Inference seamlessly integrates
with popular third-party libraries including LangChain, LlamaIndex, Dify, and Chatbox.
Why Xinference
Feature | Xinference | FastChat | OpenLLM | RayLLM |
---|---|---|---|---|
OpenAI-Compatible RESTful API | ||||
vLLM Integrations | ||||
More Inference Engines (GGML, TensorRT) | ||||
More Platforms (CPU, Metal) | ||||
Multi-node Cluster Deployment | ||||
Image Models (Text-to-Image) | ||||
Text Embedding Models | ||||
Multimodal Models | ||||
Audio Models | ||||
More OpenAI Functionalities (Function Calling) |
Getting Started
Please give us a star before you begin, and you’ll receive instant notifications for every new release on GitHub!
- Docs
- Built-in Models
- Custom Models
- Deployment Docs
- Examples and Tutorials
Jupyter Notebook
The lightest way to experience Xinference is to try our Jupyter Notebook on Google Colab.
Docker
Nvidia GPU users can start Xinference server using Xinference Docker Image. Prior to executing the installation command, ensure that both Docker and CUDA are set up on your system.
docker run --name xinference -d -p 9997:9997 -e XINFERENCE_HOME=/data -v </on/your/host>:/data --gpus all xprobe/xinference:latest xinference-local -H 0.0.0.0
K8s via helm
Ensure that you have GPU support in your Kubernetes cluster, then install as follows.
# add repo
helm repo add xinference https://xorbitsai.github.io/xinference-helm-charts
# update indexes and query xinference versions
helm repo update xinference
helm search repo xinference/xinference --devel --versions
# install xinference
helm install xinference xinference/xinference -n xinference --version 0.0.1-v<xinference_release_version>
For more customized installation methods on K8s, please refer to the documentation.
Quick Start
Install Xinference by using pip as follows. (For more options, see Installation page.)
pip install "xinference[all]"
To start a local instance of Xinference, run the following command:
$ xinference-local
Once Xinference is running, there are multiple ways you can try it: via the web UI, via cURL, via the command line, or via the Xinference’s python client. Check out our docs for the guide.
Getting involved
Platform | Purpose |
---|---|
Github Issues | Reporting bugs and filing feature requests. |
Slack | Collaborating with other Xorbits users. |
Staying up-to-date on new features. |
Contributors
Xorbits Inference
Project Details
- zhanghaiqiangshigezhu/inference
- Apache License 2.0
- Last Updated: 9/21/2024
Recomended MCP Servers
CrateDB MCP Server for Text-to-SQL and documentation retrieval.
MCP server for interacting with RabbitMQ
TS based companion MCP server for the Drupal MCP module that works with the STDIO transport.
WhatsApp Web MCP Server
用 Vue3 和 Go 搭建的微软 New Bing 演示站点,拥有一致的 UI 体验,支持 ChatGPT 提示词,国内可用。
Apache AGE MCP Server
A Model Context Protocol (MCP) server for interacting with Shortcut (formerly Clubhouse)
MCP server for anki