llama-stack-mirror/llama_stack/providers/tests/inference
Sixian Yi bf610adf9b Temporary Commit at 1/6/2025, 8:13:09 PM
Summary:
### THIS DIFF

### PLAN

### CONTEXT

### DESIGN

Test Plan:
# Test Setup

**Type checker and check that the build compiles**

**Unit Tests**

**E2E Tests**

// Screenshots and videos
| Before | After |
|--|
| … | … |

# Monitoring Plan.
2025-01-06 20:16:00 -08:00
..
__init__.py Remove "routing_table" and "routing_key" concepts for the user (#201) 2024-10-10 10:24:13 -07:00
ci_test.py Amend 2025-01-05 23:47:05 -08:00
conftest.py Make embedding generation go through inference (#606) 2024-12-12 11:47:50 -08:00
fixtures.py Update the "InterleavedTextMedia" type (#635) 2024-12-17 11:18:31 -08:00
pasta.jpeg Enable vision models for (Together, Fireworks, Meta-Reference, Ollama) (#376) 2024-11-05 16:22:33 -08:00
test_embeddings.py add embedding model by default to distribution templates (#617) 2024-12-13 12:48:00 -08:00
test_model_registration.py [4/n][torchtune integration] support lazy load model during inference (#620) 2024-12-18 16:30:53 -08:00
test_prompt_adapter.py [remove import *] clean up import *'s (#689) 2024-12-27 15:45:44 -08:00
test_text_inference.py Temporary Commit at 1/6/2025, 8:13:09 PM 2025-01-06 20:16:00 -08:00
test_vision_inference.py [remove import *] clean up import *'s (#689) 2024-12-27 15:45:44 -08:00
utils.py Enable vision models for (Together, Fireworks, Meta-Reference, Ollama) (#376) 2024-11-05 16:22:33 -08:00