Fix tgi doc

This commit is contained in:
Ashwin Bharambe 2024-11-19 21:05:59 -08:00
parent 1086b500f9
commit 89f5093dfc
6 changed files with 132 additions and 135 deletions

View file

@ -1,177 +1,171 @@
{ {
"together": [ "together": [
"pypdf", "aiosqlite",
"sentencepiece",
"pandas",
"redis",
"nltk",
"psycopg2-binary",
"scikit-learn",
"chardet",
"matplotlib",
"pillow",
"tqdm",
"chromadb-client",
"transformers",
"blobfile", "blobfile",
"aiosqlite", "chardet",
"together", "chromadb-client",
"faiss-cpu", "faiss-cpu",
"scipy",
"numpy",
"torch --index-url https://download.pytorch.org/whl/cpu",
"sentence-transformers --no-deps",
"aiosqlite",
"fastapi", "fastapi",
"fire", "fire",
"httpx", "httpx",
"uvicorn" "matplotlib",
"nltk",
"numpy",
"pandas",
"pillow",
"psycopg2-binary",
"pypdf",
"redis",
"scikit-learn",
"scipy",
"sentencepiece",
"together",
"tqdm",
"transformers",
"uvicorn",
"sentence-transformers --no-deps",
"torch --index-url https://download.pytorch.org/whl/cpu"
], ],
"remote-vllm": [ "remote-vllm": [
"pypdf", "aiosqlite",
"sentencepiece",
"pandas",
"redis",
"nltk",
"psycopg2-binary",
"scikit-learn",
"chardet",
"matplotlib",
"pillow",
"tqdm",
"chromadb-client",
"transformers",
"openai",
"blobfile", "blobfile",
"aiosqlite", "chardet",
"chromadb-client",
"faiss-cpu", "faiss-cpu",
"scipy",
"numpy",
"torch --index-url https://download.pytorch.org/whl/cpu",
"sentence-transformers --no-deps",
"aiosqlite",
"fastapi", "fastapi",
"fire", "fire",
"httpx", "httpx",
"uvicorn" "matplotlib",
"nltk",
"numpy",
"openai",
"pandas",
"pillow",
"psycopg2-binary",
"pypdf",
"redis",
"scikit-learn",
"scipy",
"sentencepiece",
"tqdm",
"transformers",
"uvicorn",
"sentence-transformers --no-deps",
"torch --index-url https://download.pytorch.org/whl/cpu"
], ],
"fireworks": [ "fireworks": [
"pypdf", "aiosqlite",
"sentencepiece",
"pandas",
"redis",
"nltk",
"psycopg2-binary",
"scikit-learn",
"chardet",
"fireworks-ai",
"matplotlib",
"pillow",
"tqdm",
"chromadb-client",
"transformers",
"blobfile", "blobfile",
"aiosqlite", "chardet",
"chromadb-client",
"faiss-cpu", "faiss-cpu",
"scipy",
"numpy",
"torch --index-url https://download.pytorch.org/whl/cpu",
"sentence-transformers --no-deps",
"aiosqlite",
"fastapi", "fastapi",
"fire", "fire",
"fireworks-ai",
"httpx", "httpx",
"uvicorn" "matplotlib",
"nltk",
"numpy",
"pandas",
"pillow",
"psycopg2-binary",
"pypdf",
"redis",
"scikit-learn",
"scipy",
"sentencepiece",
"tqdm",
"transformers",
"uvicorn",
"sentence-transformers --no-deps",
"torch --index-url https://download.pytorch.org/whl/cpu"
], ],
"tgi": [ "tgi": [
"pypdf",
"sentencepiece",
"pandas",
"redis",
"nltk",
"psycopg2-binary",
"scikit-learn",
"chardet",
"matplotlib",
"pillow",
"tqdm",
"huggingface_hub",
"chromadb-client",
"aiohttp", "aiohttp",
"transformers", "aiosqlite",
"blobfile", "blobfile",
"aiosqlite", "chardet",
"chromadb-client",
"faiss-cpu", "faiss-cpu",
"scipy",
"numpy",
"torch --index-url https://download.pytorch.org/whl/cpu",
"sentence-transformers --no-deps",
"aiosqlite",
"fastapi", "fastapi",
"fire", "fire",
"httpx", "httpx",
"uvicorn" "huggingface_hub",
"matplotlib",
"nltk",
"numpy",
"pandas",
"pillow",
"psycopg2-binary",
"pypdf",
"redis",
"scikit-learn",
"scipy",
"sentencepiece",
"tqdm",
"transformers",
"uvicorn",
"sentence-transformers --no-deps",
"torch --index-url https://download.pytorch.org/whl/cpu"
], ],
"meta-reference-gpu": [ "meta-reference-gpu": [
"accelerate",
"aiosqlite",
"blobfile",
"chardet",
"chromadb-client",
"fairscale",
"faiss-cpu",
"fastapi",
"fire",
"httpx",
"lm-format-enforcer",
"matplotlib",
"nltk",
"numpy",
"pandas",
"pillow",
"psycopg2-binary",
"pypdf", "pypdf",
"redis",
"scikit-learn",
"scipy",
"sentencepiece", "sentencepiece",
"torch", "torch",
"pandas",
"redis",
"nltk",
"psycopg2-binary",
"scikit-learn",
"chardet",
"accelerate",
"matplotlib",
"pillow",
"fairscale",
"tqdm",
"lm-format-enforcer",
"chromadb-client",
"transformers",
"blobfile",
"aiosqlite",
"torchvision", "torchvision",
"faiss-cpu", "tqdm",
"transformers",
"uvicorn",
"zmq", "zmq",
"scipy",
"numpy",
"torch --index-url https://download.pytorch.org/whl/cpu",
"sentence-transformers --no-deps", "sentence-transformers --no-deps",
"aiosqlite", "torch --index-url https://download.pytorch.org/whl/cpu"
"fastapi",
"fire",
"httpx",
"uvicorn"
], ],
"ollama": [ "ollama": [
"ollama",
"pypdf",
"sentencepiece",
"pandas",
"redis",
"nltk",
"psycopg2-binary",
"scikit-learn",
"chardet",
"matplotlib",
"pillow",
"tqdm",
"chromadb-client",
"aiohttp", "aiohttp",
"transformers", "aiosqlite",
"blobfile", "blobfile",
"aiosqlite", "chardet",
"chromadb-client",
"faiss-cpu", "faiss-cpu",
"scipy",
"numpy",
"torch --index-url https://download.pytorch.org/whl/cpu",
"sentence-transformers --no-deps",
"aiosqlite",
"fastapi", "fastapi",
"fire", "fire",
"httpx", "httpx",
"uvicorn" "matplotlib",
"nltk",
"numpy",
"ollama",
"pandas",
"pillow",
"psycopg2-binary",
"pypdf",
"redis",
"scikit-learn",
"scipy",
"sentencepiece",
"tqdm",
"transformers",
"uvicorn",
"sentence-transformers --no-deps",
"torch --index-url https://download.pytorch.org/whl/cpu"
] ]
} }

View file

@ -82,7 +82,10 @@ def collect_template_dependencies(template_dir: Path) -> tuple[str, list[str]]:
template = template_func() template = template_func()
normal_deps, special_deps = get_provider_dependencies(template.providers) normal_deps, special_deps = get_provider_dependencies(template.providers)
# Combine all dependencies in order: normal deps, special deps, server deps # Combine all dependencies in order: normal deps, special deps, server deps
all_deps = normal_deps + special_deps + SERVER_DEPENDENCIES all_deps = sorted(list(set(normal_deps + SERVER_DEPENDENCIES))) + sorted(
list(set(special_deps))
)
return template.name, all_deps return template.name, all_deps
except Exception: except Exception:
return None, [] return None, []

View file

@ -2,7 +2,7 @@ version: '2'
name: tgi name: tgi
distribution_spec: distribution_spec:
description: Use (an external) TGI server for running LLM inference description: Use (an external) TGI server for running LLM inference
docker_image: llamastack/distribution-tgi:test-0.0.52rc3 docker_image: null
providers: providers:
inference: inference:
- remote::tgi - remote::tgi

View file

@ -1,6 +1,6 @@
version: '2' version: '2'
image_name: tgi image_name: tgi
docker_image: llamastack/distribution-tgi:test-0.0.52rc3 docker_image: null
conda_env: tgi conda_env: tgi
apis: apis:
- agents - agents

View file

@ -1,6 +1,6 @@
version: '2' version: '2'
image_name: tgi image_name: tgi
docker_image: llamastack/distribution-tgi:test-0.0.52rc3 docker_image: null
conda_env: tgi conda_env: tgi
apis: apis:
- agents - agents

View file

@ -41,7 +41,7 @@ def get_distribution_template() -> DistributionTemplate:
name="tgi", name="tgi",
distro_type="self_hosted", distro_type="self_hosted",
description="Use (an external) TGI server for running LLM inference", description="Use (an external) TGI server for running LLM inference",
docker_image="llamastack/distribution-tgi:test-0.0.52rc3", docker_image=None,
template_path=Path(__file__).parent / "doc_template.md", template_path=Path(__file__).parent / "doc_template.md",
providers=providers, providers=providers,
default_models=[inference_model, safety_model], default_models=[inference_model, safety_model],