diff --git a/COMMANDS.md b/COMMANDS.md new file mode 100644 index 000000000..55289267c --- /dev/null +++ b/COMMANDS.md @@ -0,0 +1,22 @@ +```bash +source ~/miniconda3/bin/activate +conda create --prefix ./envs python=3.10 + +source ~/miniconda3/bin/activate +conda activate ./envs + +pip install -e . \ +&& llama stack build --config ./build.yaml --image-type conda \ +&& llama stack run ./run.yaml \ + --port 5001 + +pytest llama_stack/providers/tests/inference/test_text_inference.py -v -k groq --lf -s + +trash .git/hooks/pre-commit + +llama stack build --template ollama --image-type conda \ +&& llama stack run ./distributions/ollama/run.yaml \ + --port $LLAMA_STACK_PORT \ + --env INFERENCE_MODEL=$INFERENCE_MODEL \ + --env OLLAMA_URL=http://localhost:11434 +``` diff --git a/llama_stack/providers/remote/inference/ollama/ollama.py b/llama_stack/providers/remote/inference/ollama/ollama.py index acd5b62bc..163333668 100644 --- a/llama_stack/providers/remote/inference/ollama/ollama.py +++ b/llama_stack/providers/remote/inference/ollama/ollama.py @@ -48,10 +48,10 @@ model_aliases = [ "llama3.1:8b-instruct-fp16", CoreModelId.llama3_1_8b_instruct.value, ), - build_model_alias_with_just_provider_model_id( - "llama3.1:8b", - CoreModelId.llama3_1_8b_instruct.value, - ), + # build_model_alias_with_just_provider_model_id( + # "llama3.1:8b", + # CoreModelId.llama3_1_8b_instruct.value, + # ), build_model_alias( "llama3.1:70b-instruct-fp16", CoreModelId.llama3_1_70b_instruct.value,