llama-stack/llama_stack/providers/inline
Sarthak Deshpande 5ece262976
chore: Make code interpreter async (#1654)
# What does this PR do?
 Made code interpreter tool call to be async such that its non blocking

## Test Plan
pytest -s -v tests/integration/agents/test_agents.py
--stack-config=together --text-model=meta-llama/Llama-3.3-70B-Instruct
<img width="1693" alt="image"
src="https://github.com/user-attachments/assets/42520bb6-7acf-42d5-b71f-b35ca149d722"
/>


[//]: # (## Documentation)

Co-authored-by: sarthakdeshpande <sarthak.deshpande@engati.com>
2025-03-18 14:13:46 -07:00
..
agents feat(agent): support multiple tool groups (#1556) 2025-03-17 22:13:09 -07:00
datasetio feat(api): (1/n) datasets api clean up (#1573) 2025-03-17 16:55:45 -07:00
eval feat(api): (1/n) datasets api clean up (#1573) 2025-03-17 16:55:45 -07:00
inference test: add unit test to ensure all config types are instantiable (#1601) 2025-03-12 22:29:58 -07:00
ios/inference chore: removed executorch submodule (#1265) 2025-02-25 21:57:21 -08:00
post_training feat(api): (1/n) datasets api clean up (#1573) 2025-03-17 16:55:45 -07:00
safety feat(agent): support multiple tool groups (#1556) 2025-03-17 22:13:09 -07:00
scoring feat(api): (1/n) datasets api clean up (#1573) 2025-03-17 16:55:45 -07:00
telemetry refactor: move all datetime.now() calls to UTC (#1589) 2025-03-13 15:34:53 -07:00
tool_runtime chore: Make code interpreter async (#1654) 2025-03-18 14:13:46 -07:00
vector_io feat: Qdrant inline provider (#1273) 2025-03-18 14:04:21 -07:00
__init__.py impls -> inline, adapters -> remote (#381) 2024-11-06 14:54:05 -08:00