mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-14 06:23:45 +00:00
Add toolchain from agentic system here
This commit is contained in:
parent
f6b2b2fb39
commit
95781ec85d
71 changed files with 11899 additions and 0 deletions
9
toolchain/configs/ashwin.yaml
Normal file
9
toolchain/configs/ashwin.yaml
Normal file
|
@ -0,0 +1,9 @@
|
|||
model_inference_config:
|
||||
impl_type: "inline"
|
||||
inline_config:
|
||||
checkpoint_type: "pytorch"
|
||||
checkpoint_dir: /home/ashwin/local/checkpoints/Meta-Llama-3.1-8B-Instruct-20240710150000
|
||||
tokenizer_path: /home/ashwin/local/checkpoints/Meta-Llama-3.1-8B-Instruct-20240710150000/tokenizer.model
|
||||
model_parallel_size: 1
|
||||
max_seq_len: 2048
|
||||
max_batch_size: 1
|
9
toolchain/configs/chrisluc.yaml
Normal file
9
toolchain/configs/chrisluc.yaml
Normal file
|
@ -0,0 +1,9 @@
|
|||
model_inference_config:
|
||||
impl_type: "inline"
|
||||
inline_config:
|
||||
checkpoint_type: "pytorch"
|
||||
checkpoint_dir: /home/chrisluc/models/Meta-Llama-3.1-8B-Instruct-20240710150000
|
||||
tokenizer_path: /home/chrisluc/models/Meta-Llama-3.1-8B-Instruct-20240710150000/tokenizer.model
|
||||
model_parallel_size: 1
|
||||
max_seq_len: 2048
|
||||
max_batch_size: 1
|
9
toolchain/configs/default.yaml
Normal file
9
toolchain/configs/default.yaml
Normal file
|
@ -0,0 +1,9 @@
|
|||
model_inference_config:
|
||||
impl_type: "inline"
|
||||
inline_config:
|
||||
checkpoint_type: "pytorch"
|
||||
checkpoint_dir: /home/dalton/models/Meta-Llama-3.1-8B-Instruct-20240710150000
|
||||
tokenizer_path: /home/dalton/models/Meta-Llama-3.1-8B-Instruct-20240710150000/tokenizer.model
|
||||
model_parallel_size: 1
|
||||
max_seq_len: 2048
|
||||
max_batch_size: 1
|
9
toolchain/configs/hjshah.yaml
Normal file
9
toolchain/configs/hjshah.yaml
Normal file
|
@ -0,0 +1,9 @@
|
|||
model_inference_config:
|
||||
impl_type: "inline"
|
||||
inline_config:
|
||||
checkpoint_type: "pytorch"
|
||||
checkpoint_dir: /home/hjshah/local/checkpoints/Meta-Llama-3.1-8B-Instruct-20240710150000
|
||||
tokenizer_path: /home/hjshah/local/checkpoints/Meta-Llama-3.1-8B-Instruct-20240710150000/tokenizer.model
|
||||
model_parallel_size: 1
|
||||
max_seq_len: 2048
|
||||
max_batch_size: 1
|
9
toolchain/configs/long_seqlen.yaml
Normal file
9
toolchain/configs/long_seqlen.yaml
Normal file
|
@ -0,0 +1,9 @@
|
|||
model_inference_config:
|
||||
impl_type: "inline"
|
||||
inline_config:
|
||||
checkpoint_type: "pytorch"
|
||||
checkpoint_dir: /home/hjshah/local/checkpoints/Meta-Llama-3.1-8B-Instruct-20240710150000
|
||||
tokenizer_path: /home/hjshah/local/checkpoints/Meta-Llama-3.1-8B-Instruct-20240710150000/tokenizer.model
|
||||
model_parallel_size: 1
|
||||
max_seq_len: 8192
|
||||
max_batch_size: 1
|
Loading…
Add table
Add a link
Reference in a new issue