dell tgi fix

This commit is contained in:
Xi Yan 2024-10-16 15:37:05 -07:00
parent c4d5d6bb91
commit cfc97df6d5
3 changed files with 55 additions and 1 deletions

View file

@ -10,6 +10,8 @@ distribution_spec:
- remote::fireworks
safety: meta-reference
agents: meta-reference
memory: meta-reference
memory:
- remote::chromadb
- meta-reference
telemetry: meta-reference
image_type: docker

View file

@ -52,6 +52,10 @@ class _HfAdapter(Inference, ModelsProtocolPrivate):
async def list_models(self) -> List[ModelDef]:
repo = self.model_id
# tmp hack to support Dell
if repo not in self.huggingface_repo_to_llama_model_id:
repo = "meta-llama/Llama-3.1-8B-Instruct"
identifier = self.huggingface_repo_to_llama_model_id[repo]
return [
ModelDef(

View file

@ -0,0 +1,48 @@
version: '2'
built_at: '2024-10-08T17:40:45.325529'
image_name: local
docker_image: null
conda_env: local
apis:
- shields
- agents
- models
- memory
- memory_banks
- inference
- safety
providers:
inference:
- provider_id: remote::tgi
provider_type: remote::tgi
config:
url: http://127.0.0.1:5009
safety:
- provider_id: meta-reference
provider_type: meta-reference
config:
llama_guard_shield:
model: Llama-Guard-3-1B
excluded_categories: []
disable_input_check: false
disable_output_check: false
prompt_guard_shield:
model: Prompt-Guard-86M
memory:
- provider_id: remote::chromadb
provider_type: remote::chromadb
config:
host: localhost
port: 6000
agents:
- provider_id: meta-reference
provider_type: meta-reference
config:
persistence_store:
namespace: null
type: sqlite
db_path: /home/xiyan/.llama/runtime/kvstore.db
telemetry:
- provider_id: meta-reference
provider_type: meta-reference
config: {}