Fedir Zadniprovskyi
feat: dependency injection
bf48682
raw
history blame
2.04 kB
from __future__ import annotations
import gc
from fastapi import (
APIRouter,
Response,
)
import huggingface_hub
from huggingface_hub.hf_api import RepositoryNotFoundError
from faster_whisper_server import hf_utils
from faster_whisper_server.dependencies import ModelManagerDependency # noqa: TCH001
router = APIRouter()
@router.get("/health")
def health() -> Response:
return Response(status_code=200, content="OK")
@router.post("/api/pull/{model_name:path}", tags=["experimental"], summary="Download a model from Hugging Face.")
def pull_model(model_name: str) -> Response:
if hf_utils.does_local_model_exist(model_name):
return Response(status_code=200, content="Model already exists")
try:
huggingface_hub.snapshot_download(model_name, repo_type="model")
except RepositoryNotFoundError as e:
return Response(status_code=404, content=str(e))
return Response(status_code=201, content="Model downloaded")
@router.get("/api/ps", tags=["experimental"], summary="Get a list of loaded models.")
def get_running_models(
model_manager: ModelManagerDependency,
) -> dict[str, list[str]]:
return {"models": list(model_manager.loaded_models.keys())}
@router.post("/api/ps/{model_name:path}", tags=["experimental"], summary="Load a model into memory.")
def load_model_route(model_manager: ModelManagerDependency, model_name: str) -> Response:
if model_name in model_manager.loaded_models:
return Response(status_code=409, content="Model already loaded")
model_manager.load_model(model_name)
return Response(status_code=201)
@router.delete("/api/ps/{model_name:path}", tags=["experimental"], summary="Unload a model from memory.")
def stop_running_model(model_manager: ModelManagerDependency, model_name: str) -> Response:
model = model_manager.loaded_models.get(model_name)
if model is not None:
del model_manager.loaded_models[model_name]
gc.collect()
return Response(status_code=204)
return Response(status_code=404)