From 2da93c883533d49dd070f58b8f3ab5bc019c136c Mon Sep 17 00:00:00 2001 From: Xi Yan Date: Tue, 19 Nov 2024 14:20:07 -0800 Subject: [PATCH] fix 3.2-1b fireworks --- distributions/dependencies.json | 204 +++++++++--------- .../self_hosted_distro/fireworks.md | 2 +- .../remote/inference/fireworks/fireworks.py | 2 +- llama_stack/templates/fireworks/run.yaml | 2 +- 4 files changed, 105 insertions(+), 105 deletions(-) diff --git a/distributions/dependencies.json b/distributions/dependencies.json index 469b6f14e..0f85b70c6 100644 --- a/distributions/dependencies.json +++ b/distributions/dependencies.json @@ -1,26 +1,26 @@ { "together": [ - "scipy", - "blobfile", - "together", - "tqdm", - "sentencepiece", - "matplotlib", - "pandas", "pypdf", - "scikit-learn", - "nltk", - "faiss-cpu", - "chardet", - "numpy", - "psycopg2-binary", - "aiosqlite", - "pillow", + "sentencepiece", + "pandas", "redis", - "transformers", + "nltk", + "psycopg2-binary", + "scikit-learn", + "chardet", + "matplotlib", + "pillow", + "tqdm", "chromadb-client", - "sentence-transformers --no-deps", + "transformers", + "blobfile", + "aiosqlite", + "together", + "faiss-cpu", + "scipy", + "numpy", "torch --index-url https://download.pytorch.org/whl/cpu", + "sentence-transformers --no-deps", "aiosqlite", "fastapi", "fire", @@ -28,27 +28,27 @@ "uvicorn" ], "remote-vllm": [ - "scipy", - "blobfile", - "tqdm", - "sentencepiece", - "matplotlib", - "pandas", "pypdf", - "scikit-learn", - "nltk", - "faiss-cpu", - "chardet", - "openai", - "numpy", - "psycopg2-binary", - "aiosqlite", - "pillow", + "sentencepiece", + "pandas", "redis", - "transformers", + "nltk", + "psycopg2-binary", + "scikit-learn", + "chardet", + "matplotlib", + "pillow", + "tqdm", "chromadb-client", - "sentence-transformers --no-deps", + "transformers", + "openai", + "blobfile", + "aiosqlite", + "faiss-cpu", + "scipy", + "numpy", "torch --index-url https://download.pytorch.org/whl/cpu", + "sentence-transformers --no-deps", "aiosqlite", "fastapi", "fire", @@ -56,27 +56,27 @@ "uvicorn" ], "fireworks": [ - "scipy", - "blobfile", - "tqdm", + "pypdf", "sentencepiece", + "pandas", + "redis", + "nltk", + "psycopg2-binary", + "scikit-learn", + "chardet", "fireworks-ai", "matplotlib", - "pandas", - "pypdf", - "scikit-learn", - "nltk", - "faiss-cpu", - "chardet", - "numpy", - "psycopg2-binary", - "aiosqlite", "pillow", - "redis", - "transformers", + "tqdm", "chromadb-client", - "sentence-transformers --no-deps", + "transformers", + "blobfile", + "aiosqlite", + "faiss-cpu", + "scipy", + "numpy", "torch --index-url https://download.pytorch.org/whl/cpu", + "sentence-transformers --no-deps", "aiosqlite", "fastapi", "fire", @@ -84,28 +84,28 @@ "uvicorn" ], "tgi": [ - "scipy", - "blobfile", + "pypdf", + "sentencepiece", + "pandas", + "redis", + "nltk", + "psycopg2-binary", + "scikit-learn", + "chardet", + "matplotlib", + "pillow", "tqdm", "huggingface_hub", - "sentencepiece", - "matplotlib", - "pandas", - "pypdf", - "scikit-learn", - "nltk", - "faiss-cpu", - "chardet", - "numpy", - "psycopg2-binary", - "aiosqlite", - "pillow", - "redis", - "transformers", "chromadb-client", "aiohttp", - "sentence-transformers --no-deps", + "transformers", + "blobfile", + "aiosqlite", + "faiss-cpu", + "scipy", + "numpy", "torch --index-url https://download.pytorch.org/whl/cpu", + "sentence-transformers --no-deps", "aiosqlite", "fastapi", "fire", @@ -113,32 +113,32 @@ "uvicorn" ], "meta-reference-gpu": [ - "scipy", - "blobfile", - "tqdm", - "torchvision", - "sentencepiece", - "zmq", - "matplotlib", - "pandas", "pypdf", - "scikit-learn", - "accelerate", - "nltk", - "faiss-cpu", + "sentencepiece", "torch", - "chardet", - "numpy", - "psycopg2-binary", - "aiosqlite", - "pillow", + "pandas", "redis", + "nltk", + "psycopg2-binary", + "scikit-learn", + "chardet", + "accelerate", + "matplotlib", + "pillow", "fairscale", + "tqdm", "lm-format-enforcer", - "transformers", "chromadb-client", - "sentence-transformers --no-deps", + "transformers", + "blobfile", + "aiosqlite", + "torchvision", + "faiss-cpu", + "zmq", + "scipy", + "numpy", "torch --index-url https://download.pytorch.org/whl/cpu", + "sentence-transformers --no-deps", "aiosqlite", "fastapi", "fire", @@ -146,28 +146,28 @@ "uvicorn" ], "ollama": [ - "scipy", - "blobfile", - "tqdm", - "sentencepiece", - "matplotlib", - "pandas", - "pypdf", - "scikit-learn", - "nltk", "ollama", - "faiss-cpu", - "chardet", - "numpy", - "psycopg2-binary", - "aiosqlite", - "pillow", + "pypdf", + "sentencepiece", + "pandas", "redis", - "transformers", + "nltk", + "psycopg2-binary", + "scikit-learn", + "chardet", + "matplotlib", + "pillow", + "tqdm", "chromadb-client", "aiohttp", - "sentence-transformers --no-deps", + "transformers", + "blobfile", + "aiosqlite", + "faiss-cpu", + "scipy", + "numpy", "torch --index-url https://download.pytorch.org/whl/cpu", + "sentence-transformers --no-deps", "aiosqlite", "fastapi", "fire", diff --git a/docs/source/getting_started/distributions/self_hosted_distro/fireworks.md b/docs/source/getting_started/distributions/self_hosted_distro/fireworks.md index 66a150f50..cca1155e1 100644 --- a/docs/source/getting_started/distributions/self_hosted_distro/fireworks.md +++ b/docs/source/getting_started/distributions/self_hosted_distro/fireworks.md @@ -25,7 +25,7 @@ The following models are available by default: - `meta-llama/Llama-3.1-8B-Instruct (fireworks/llama-v3p1-8b-instruct)` - `meta-llama/Llama-3.1-70B-Instruct (fireworks/llama-v3p1-70b-instruct)` - `meta-llama/Llama-3.1-405B-Instruct-FP8 (fireworks/llama-v3p1-405b-instruct)` -- `meta-llama/Llama-3.2-3B-Instruct (fireworks/llama-v3p2-1b-instruct)` +- `meta-llama/Llama-3.2-1B-Instruct (fireworks/llama-v3p2-1b-instruct)` - `meta-llama/Llama-3.2-3B-Instruct (fireworks/llama-v3p2-3b-instruct)` - `meta-llama/Llama-3.2-11B-Vision-Instruct (fireworks/llama-v3p2-11b-vision-instruct)` - `meta-llama/Llama-3.2-90B-Vision-Instruct (fireworks/llama-v3p2-90b-vision-instruct)` diff --git a/llama_stack/providers/remote/inference/fireworks/fireworks.py b/llama_stack/providers/remote/inference/fireworks/fireworks.py index d8cbca5f9..c3e634155 100644 --- a/llama_stack/providers/remote/inference/fireworks/fireworks.py +++ b/llama_stack/providers/remote/inference/fireworks/fireworks.py @@ -50,7 +50,7 @@ MODEL_ALIASES = [ ), build_model_alias( "fireworks/llama-v3p2-1b-instruct", - CoreModelId.llama3_2_3b_instruct.value, + CoreModelId.llama3_2_1b_instruct.value, ), build_model_alias( "fireworks/llama-v3p2-3b-instruct", diff --git a/llama_stack/templates/fireworks/run.yaml b/llama_stack/templates/fireworks/run.yaml index aa44f0f84..6add39c3a 100644 --- a/llama_stack/templates/fireworks/run.yaml +++ b/llama_stack/templates/fireworks/run.yaml @@ -57,7 +57,7 @@ models: provider_id: null provider_model_id: fireworks/llama-v3p1-405b-instruct - metadata: {} - model_id: meta-llama/Llama-3.2-3B-Instruct + model_id: meta-llama/Llama-3.2-1B-Instruct provider_id: null provider_model_id: fireworks/llama-v3p2-1b-instruct - metadata: {}