llama-stack-mirror/llama_stack/models/llama
skamenan7 f5c1935c18 fix: Resolve Llama4 tool calling 500 errors
This commit addresses issue #2584 by:
- Implementing lazy torch imports in llama4/chat_format.py and datatypes.py to prevent ModuleNotFoundError in torch-free environments.
- Adding comprehensive unit tests to verify that text-only functionality works without torch and that vision features fail gracefully.
- Ensuring the module remains importable and functional for text-based operations, thus resolving the 500 internal server errors.
2025-07-23 15:20:17 -04:00
..
llama3 chore(api): add mypy coverage to chat_format (#2654) 2025-07-18 11:56:53 +02:00
llama3_1 chore: enable pyupgrade fixes (#1806) 2025-05-01 14:23:50 -07:00
llama3_2 refactor: move all llama code to models/llama out of meta reference (#1887) 2025-04-07 15:03:58 -07:00
llama3_3 chore(api): add mypy coverage to prompts (#2657) 2025-07-09 10:07:00 +02:00
llama4 fix: Resolve Llama4 tool calling 500 errors 2025-07-23 15:20:17 -04:00
resources feat: introduce llama4 support (#1877) 2025-04-05 11:53:35 -07:00
__init__.py feat: introduce llama4 support (#1877) 2025-04-05 11:53:35 -07:00
checkpoint.py chore: enable pyupgrade fixes (#1806) 2025-05-01 14:23:50 -07:00
datatypes.py fix: finish conversion to StrEnum (#2514) 2025-06-26 08:01:26 +05:30
hadamard_utils.py refactor: move all llama code to models/llama out of meta reference (#1887) 2025-04-07 15:03:58 -07:00
prompt_format.py chore: enable pyupgrade fixes (#1806) 2025-05-01 14:23:50 -07:00
quantize_impls.py chore: enable pyupgrade fixes (#1806) 2025-05-01 14:23:50 -07:00
sku_list.py refactor: make sku_list resolve provider aliases generically 2025-07-23 15:20:17 -04:00
sku_types.py chore: enable pyupgrade fixes (#1806) 2025-05-01 14:23:50 -07:00
tokenizer_utils.py chore: remove usage of load_tiktoken_bpe (#2276) 2025-06-02 07:33:37 -07:00