Installation
InstallXinference through PyPI:
Deploy Xinference Locally or in a Distributed Cluster
For local deployment, runxinference.
To deploy Xinference in a cluster, first start an Xinference supervisor using the xinference-supervisor. You can also use the option -p to specify the port and -H to specify the host. The default port is 9997.
Then, start the Xinference workers using xinference-worker on each server you want to run them on.
You can consult the README file from Xinference for more information.
Wrapper
To use Xinference with LangChain, you need to first launch a model. You can use command line interface (CLI) to do so:Integrate with a LLMChain
Connect these docs programmatically to Claude, VSCode, and more via MCP for real-time answers.