We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
V100 * 8
13.3
xinference-local --host 0.0.0.0 --port 8080
目前在机器上用了两张V100卡部署GLM4,现在发现显存不太够用,如何无缝衔接的再添加两张显卡
The text was updated successfully, but these errors were encountered:
n-gpu 指定卡的个数。
Sorry, something went wrong.
那是要重新启动后台服务? @qinxuye
不需要,停掉模型重新 launch。
也就是模型短暂是不可用的对吧
如何不停模型服务,直接扩容呢?
开源不支持这个特性。
No branches or pull requests
System Info / 系統信息
V100 * 8
Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
Version info / 版本信息
13.3
The command used to start Xinference / 用以启动 xinference 的命令
xinference-local --host 0.0.0.0 --port 8080
Reproduction / 复现过程
目前在机器上用了两张V100卡部署GLM4,现在发现显存不太够用,如何无缝衔接的再添加两张显卡
Expected behavior / 期待表现
目前在机器上用了两张V100卡部署GLM4,现在发现显存不太够用,如何无缝衔接的再添加两张显卡
The text was updated successfully, but these errors were encountered: