llama-stack-mirror/llama_stack/models/llama/llama4/quantization
jiawenliu64 c8a0b110c0 fix: on-the-fly int4 quantize parameter
Before this PR:
```
[rank1]: TypeError: quantize_int4() got multiple values for argument 'output_device'
```
2025-04-09 13:35:11 -07:00
..
__init__.py refactor: move all llama code to models/llama out of meta reference (#1887) 2025-04-07 15:03:58 -07:00
loader.py fix: on-the-fly int4 quantize parameter 2025-04-09 13:35:11 -07:00