Skip to content

Releases: aniketmaurya/fastserve-ai

v0.0.3

25 Feb 13:32
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.0.2...v0.0.3

v0.0.2

21 Dec 17:12
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.0.1...v0.0.2

v0.0.1

06 Dec 19:40
Compare
Choose a tag to compare

FastServe

Machine Learning Serving focused on GenAI & LLMs with simplicity as the top priority.

Installation

git clone https://github.com/aniketmaurya/fastserve.git
cd fastserve
pip install .

Run locally

python -m fastserve

Usage/Examples

Serve Mistral-7B with Llama-cpp

from fastserve.models import ServeLlamaCpp

model_path = "openhermes-2-mistral-7b.Q5_K_M.gguf"
serve = ServeLlamaCpp(model_path=model_path, )
serve.run_server()

or, run python -m fastserve.models --model llama-cpp --model_path openhermes-2-mistral-7b.Q5_K_M.gguf from terminal.

Serve SDXL Turbo

from fastserve.models import ServeSDXLTurbo

serve = ServeSDXLTurbo(device="cuda", batch_size=2, timeout=1)
serve.run_server()

or, run python -m fastserve.models --model sdxl-turbo --batch_size 2 --timeout 1 from terminal.