ENH: optimize availability of vLLM #4277
Workflow file for this run
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
name: Python CI | |
on: | |
push: | |
branches: | |
- '*' | |
pull_request: | |
types: ['opened', 'reopened', 'synchronize'] | |
concurrency: | |
group: ${{ github.workflow }}-${{ github.ref }} | |
cancel-in-progress: true | |
jobs: | |
lint: | |
runs-on: ${{ matrix.os }} | |
strategy: | |
fail-fast: false | |
matrix: | |
os: [ "ubuntu-latest" ] | |
python-version: [ "3.10" ] | |
steps: | |
- name: Check out code | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 0 | |
submodules: recursive | |
- name: Set up Python environment | |
uses: actions/setup-python@v4 | |
with: | |
python-version: "3.10" | |
- name: flake8 Lint | |
uses: py-actions/flake8@v2 | |
with: | |
path: "xinference" | |
args: "--config setup.cfg" | |
- name: black | |
uses: psf/black@stable | |
with: | |
src: "xinference" | |
options: "--check" | |
version: "23.12.0" | |
- uses: isort/isort-action@master | |
with: | |
sortPaths: "xinference" | |
configuration: "--check-only --diff --sp setup.cfg" | |
- name: mypy | |
run: pip install mypy && mypy --install-types --non-interactive xinference | |
- name: codespell | |
run: pip install codespell && codespell --ignore-words-list thirdparty xinference | |
- name: Set up Node.js | |
uses: actions/setup-node@v1 | |
with: | |
node-version: 16 | |
# ESLint and Prettier must be in `package.json` | |
- name: Install Node.js dependencies | |
run: cd xinference/web/ui && npm ci | |
- name: ESLint Check | |
run: cd xinference/web/ui && npx eslint . | |
- name: Prettier Check | |
run: cd xinference/web/ui && ./node_modules/.bin/prettier --check . | |
build_test_job: | |
runs-on: ${{ matrix.os }} | |
needs: lint | |
env: | |
CONDA_ENV: test | |
SELF_HOST_PYTHON: /root/miniconda3/envs/inference_test/bin/python | |
SELF_HOST_CONDA: /root/miniconda3/condabin/conda | |
defaults: | |
run: | |
shell: bash -l {0} | |
strategy: | |
fail-fast: false | |
matrix: | |
os: [ "ubuntu-latest", "macos-12", "windows-latest" ] | |
python-version: [ "3.8", "3.9", "3.10", "3.11" ] | |
module: [ "xinference" ] | |
exclude: | |
- { os: macos-12, python-version: 3.9 } | |
- { os: macos-12, python-version: 3.10 } | |
- { os: windows-latest, python-version: 3.9 } | |
- { os: windows-latest, python-version: 3.10 } | |
include: | |
- { os: self-hosted, module: gpu, python-version: 3.9} | |
- { os: macos-latest, module: metal, python-version: "3.10" } | |
steps: | |
- name: Check out code | |
uses: actions/checkout@v3 | |
with: | |
fetch-depth: 0 | |
submodules: recursive | |
- name: Set up conda ${{ matrix.python-version }} | |
uses: conda-incubator/setup-miniconda@v3 | |
if: ${{ matrix.module != 'gpu' }} | |
with: | |
python-version: ${{ matrix.python-version }} | |
activate-environment: ${{ env.CONDA_ENV }} | |
- name: Install dependencies | |
env: | |
MODULE: ${{ matrix.module }} | |
OS: ${{ matrix.os }} | |
if: ${{ matrix.module != 'gpu' }} | |
run: | | |
if [ "$OS" == "ubuntu-latest" ]; then | |
sudo rm -rf /usr/share/dotnet | |
sudo rm -rf /opt/ghc | |
sudo rm -rf "/usr/local/share/boost" | |
sudo rm -rf "$AGENT_TOOLSDIRECTORY" | |
fi | |
if [ "$MODULE" == "metal" ]; then | |
pip install mlx-lm | |
fi | |
pip install "llama-cpp-python==0.2.77" --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cpu | |
pip install transformers | |
pip install attrdict | |
pip install "timm>=0.9.16" | |
pip install torch | |
pip install torchvision | |
pip install accelerate | |
pip install sentencepiece | |
pip install transformers_stream_generator | |
pip install bitsandbytes | |
pip install "sentence-transformers>=2.3.1" | |
pip install s3fs | |
pip install modelscope | |
pip install diffusers | |
pip install protobuf | |
pip install FlagEmbedding | |
pip install "tenacity>=8.2.0,<8.4.0" | |
pip install -e ".[dev]" | |
pip install "jinja2==3.1.2" | |
pip install tensorizer | |
working-directory: . | |
- name: Test with pytest | |
env: | |
MODULE: ${{ matrix.module }} | |
run: | | |
if [ "$MODULE" == "gpu" ]; then | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U "openai>1,<1.40" | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U modelscope | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U sse_starlette | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U xoscar | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U "python-jose[cryptography]" | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U "passlib[bcrypt]" | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U "aioprometheus[starlette]" | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U "pynvml" | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U "transformers" | |
${{ env.SELF_HOST_CONDA }} install -c conda-forge pynini=2.1.5 | |
${{ env.SELF_HOST_CONDA }} install -c conda-forge "ffmpeg<7" | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U funasr | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U nemo_text_processing | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U omegaconf~=2.3.0 | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U WeTextProcessing | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U librosa | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U xxhash | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U "ChatTTS>0.1" | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U HyperPyYAML | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U matcha-tts | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U onnxruntime-gpu==1.16.0; sys_platform == 'linux' | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U openai-whisper | |
${{ env.SELF_HOST_PYTHON }} -m pip install -U "torch==2.1.0" "torchaudio==2.1.0" | |
${{ env.SELF_HOST_PYTHON }} -m pytest --timeout=1500 \ | |
-W ignore::PendingDeprecationWarning \ | |
--cov-config=setup.cfg --cov-report=xml --cov=xinference xinference/model/image/tests/test_stable_diffusion.py | |
${{ env.SELF_HOST_PYTHON }} -m pytest --timeout=1500 \ | |
-W ignore::PendingDeprecationWarning \ | |
--cov-config=setup.cfg --cov-report=xml --cov=xinference xinference/model/audio/tests/test_whisper.py | |
${{ env.SELF_HOST_PYTHON }} -m pytest --timeout=1500 \ | |
-W ignore::PendingDeprecationWarning \ | |
--cov-config=setup.cfg --cov-report=xml --cov=xinference xinference/model/audio/tests/test_funasr.py | |
${{ env.SELF_HOST_PYTHON }} -m pytest --timeout=1500 \ | |
-W ignore::PendingDeprecationWarning \ | |
--cov-config=setup.cfg --cov-report=xml --cov=xinference xinference/model/audio/tests/test_chattts.py | |
${{ env.SELF_HOST_PYTHON }} -m pytest --timeout=1500 \ | |
-W ignore::PendingDeprecationWarning \ | |
--cov-config=setup.cfg --cov-report=xml --cov=xinference xinference/model/audio/tests/test_cosyvoice.py | |
elif [ "$MODULE" == "metal" ]; then | |
pytest --timeout=1500 \ | |
-W ignore::PendingDeprecationWarning \ | |
--cov-config=setup.cfg --cov-report=xml --cov=xinference xinference/model/llm/mlx/tests/test_mlx.py | |
else | |
pytest --timeout=1500 \ | |
-W ignore::PendingDeprecationWarning \ | |
--cov-config=setup.cfg --cov-report=xml --cov=xinference xinference/client/tests/test_client.py | |
pytest --timeout=1500 \ | |
-W ignore::PendingDeprecationWarning \ | |
--cov-config=setup.cfg --cov-report=xml --cov=xinference --ignore xinference/client/tests/test_client.py --ignore xinference/model/image/tests/test_stable_diffusion.py --ignore xinference/model/audio/tests xinference | |
fi | |
working-directory: . |