mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-11 20:40:40 +00:00
init commit
This commit is contained in:
parent
27d5892dfa
commit
34d7fb649e
2 changed files with 26 additions and 4 deletions
|
@ -62,6 +62,16 @@ def get_inference_providers() -> Tuple[List[Provider], List[ModelInput]]:
|
||||||
],
|
],
|
||||||
GeminiConfig.sample_run_config(api_key="${env.GEMINI_API_KEY:}"),
|
GeminiConfig.sample_run_config(api_key="${env.GEMINI_API_KEY:}"),
|
||||||
),
|
),
|
||||||
|
(
|
||||||
|
"groq",
|
||||||
|
[],
|
||||||
|
GroqConfig.sample_run_config(api_key="${env.GROQ_API_KEY:}"),
|
||||||
|
),
|
||||||
|
(
|
||||||
|
"together",
|
||||||
|
[],
|
||||||
|
TogetherImplConfig.sample_run_config(api_key="${env.TOGETHER_API_KEY:}"),
|
||||||
|
),
|
||||||
]
|
]
|
||||||
inference_providers = []
|
inference_providers = []
|
||||||
available_models = {}
|
available_models = {}
|
||||||
|
@ -135,7 +145,20 @@ def get_distribution_template() -> DistributionTemplate:
|
||||||
),
|
),
|
||||||
]
|
]
|
||||||
|
|
||||||
default_models = get_model_registry(available_models)
|
default_models = get_model_registry(available_models) + [
|
||||||
|
ModelInput(
|
||||||
|
model_id="meta-llama/Llama-3.3-70B-Instruct",
|
||||||
|
provider_id="groq",
|
||||||
|
provider_model_id="groq/llama-3.3-70b-versatile",
|
||||||
|
model_type=ModelType.llm,
|
||||||
|
),
|
||||||
|
ModelInput(
|
||||||
|
model_id="meta-llama/Llama-3.1-405B-Instruct",
|
||||||
|
provider_id="together",
|
||||||
|
provider_model_id="meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo",
|
||||||
|
model_type=ModelType.llm,
|
||||||
|
),
|
||||||
|
]
|
||||||
|
|
||||||
default_datasets = [
|
default_datasets = [
|
||||||
DatasetInput(
|
DatasetInput(
|
||||||
|
|
|
@ -132,12 +132,12 @@ models:
|
||||||
provider_model_id: gemini/gemini-1.5-flash
|
provider_model_id: gemini/gemini-1.5-flash
|
||||||
model_type: llm
|
model_type: llm
|
||||||
- metadata: {}
|
- metadata: {}
|
||||||
model_id: groq/llama-3.3-70b-versatile
|
model_id: meta-llama/Llama-3.3-70B-Instruct
|
||||||
provider_id: groq
|
provider_id: groq
|
||||||
provider_model_id: groq/llama-3.3-70b-versatile
|
provider_model_id: groq/llama-3.3-70b-versatile
|
||||||
model_type: llm
|
model_type: llm
|
||||||
- metadata: {}
|
- metadata: {}
|
||||||
model_id: meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo
|
model_id: meta-llama/Llama-3.1-405B-Instruct
|
||||||
provider_id: together
|
provider_id: together
|
||||||
provider_model_id: meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo
|
provider_model_id: meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo
|
||||||
model_type: llm
|
model_type: llm
|
||||||
|
@ -156,7 +156,6 @@ datasets:
|
||||||
uri: https://huggingface.co/datasets/llamastack/simpleqa
|
uri: https://huggingface.co/datasets/llamastack/simpleqa
|
||||||
metadata:
|
metadata:
|
||||||
path: llamastack/simpleqa
|
path: llamastack/simpleqa
|
||||||
name: null
|
|
||||||
split: train
|
split: train
|
||||||
dataset_id: simpleqa
|
dataset_id: simpleqa
|
||||||
provider_id: huggingface
|
provider_id: huggingface
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue