Commit graph

  • 12cf140b3a precommit Yunlu Li 2024-11-20 21:59:36 -08:00
  • 09302347d3
    Merge branch 'meta-llama:main' into main Chacksu 2024-11-21 10:21:49 -05:00
  • 3ed2e816fa use pydantic v2's model_dump() instead of dict() Matthew Farrellee 2024-11-21 10:04:26 -05:00
  • 0027c5b614 update nb-1 Sanyam Bhutani 2024-11-21 06:53:47 -08:00
  • 06b699f5ec Update quickstart.md Sanyam Bhutani 2024-11-21 06:40:24 -08:00
  • 4deb95eaae Update quickstart.md Sanyam Bhutani 2024-11-21 06:37:21 -08:00
  • 5fbfb9d854 Merge branch 'main' into add-nvidia-inference-adapter Matthew Farrellee 2024-11-21 06:49:13 -05:00
  • 914cea8939 allow users to provide a tool Matthew Farrellee 2024-11-21 06:46:26 -05:00
  • cf079a22a0 Plurals Ashwin Bharambe 2024-11-20 23:24:59 -08:00
  • cd6ccb664c Integrate distro docs into the restructured docs Ashwin Bharambe 2024-11-20 23:20:05 -08:00
  • 2411a44833 Update more distribution docs to be simpler and partially codegen'ed Ashwin Bharambe 2024-11-20 14:44:04 -08:00
  • 511b054b77 Fix fp8 quantization script. Yunlu Li 2024-11-20 21:32:53 -08:00
  • 490c5fb730 Undo None check and temporarily move if model check before builder Connor Hack 2024-11-20 19:17:44 -08:00
  • 16ffe19a20 Account for if a permitted model is None Connor Hack 2024-11-20 18:03:34 -08:00
  • 05f1041bfa
    Merge branch 'meta-llama:main' into main Chacksu 2024-11-20 19:21:20 -05:00
  • e84d4436b5
    Since we are pushing for HF repos, we should accept them in inference configs (#497) Ashwin Bharambe 2024-11-20 16:14:37 -08:00
  • 0d4565349b Since we are pushing for HF repos, we should accept them in inference configs Ashwin Bharambe 2024-11-20 16:07:29 -08:00
  • b3f9e8b2f2
    Restructure docs (#494) Dinesh Yeduguru 2024-11-20 15:54:47 -08:00
  • efdc67e5b7 address feedback Dinesh Yeduguru 2024-11-20 15:51:55 -08:00
  • 0ec4ddd179
    Merge branch 'meta-llama:main' into main Chacksu 2024-11-20 18:46:45 -05:00
  • 068ac00a3b
    Don't depend on templates.py when print llama stack build messages (#496) Ashwin Bharambe 2024-11-20 15:44:49 -08:00
  • b95533ce08 Don't depend on templates.py when print llama stack build messages Ashwin Bharambe 2024-11-20 15:40:14 -08:00
  • a5acb59407
    Merge branch 'meta-llama:main' into main Chacksu 2024-11-20 18:30:01 -05:00
  • 3b5ea74267 Merge remote-tracking branch 'mattf/add-nvidia-inference-adapter' into cdgamarose/add_nvidia_distro Chantal D Gama Rose 2024-11-20 23:06:31 +00:00
  • 6759744235 Added distributions for inline and remote Chantal D Gama Rose 2024-11-20 23:04:48 +00:00
  • 2795731434 Update model name for mete-reference template Connor Hack 2024-11-20 14:37:27 -08:00
  • 00816cc8ef make sure codegen doesn't cause spurious diffs for no reason v0.0.53 Ashwin Bharambe 2024-11-20 13:55:43 -08:00
  • edfd92d81f
    Merge branch 'meta-llama:main' into main Chacksu 2024-11-20 16:12:38 -05:00
  • 681322731b
    Make run yaml optional so dockers can start with just --env (#492) Ashwin Bharambe 2024-11-20 13:11:40 -08:00
  • 990a997405 Make sure llama stack run ollama also works correctly Ashwin Bharambe 2024-11-20 12:52:43 -08:00
  • 15200e23ba address feedback Dinesh Yeduguru 2024-11-20 12:19:12 -08:00
  • 21a11b2bf7 restructure docs: Dinesh Yeduguru 2024-11-20 12:06:22 -08:00
  • adbb605a50 Make run yaml optional so dockers can start with just --env Ashwin Bharambe 2024-11-20 11:56:08 -08:00
  • 1d8d0593af
    register with provider even if present in stack (#491) Dinesh Yeduguru 2024-11-20 11:05:50 -08:00
  • 7197d73b12 add a new getting started Dinesh Yeduguru 2024-11-19 22:34:08 -08:00
  • 57fd974872 register with provider even if present in stack Dinesh Yeduguru 2024-11-20 10:42:43 -08:00
  • 91e7efbc91
    fall to back to read from chroma/pgvector when not in cache (#489) Dinesh Yeduguru 2024-11-20 10:30:23 -08:00
  • ae49a4cb97
    Reorganizing Zero to Hero Folder structure (#447) Justin Lee 2024-11-20 10:27:29 -08:00
  • 8d5fdeedda add _get_and_cache_bank_index Dinesh Yeduguru 2024-11-20 10:21:55 -08:00
  • a68e58f873 update notebooks, remove colab links Justin Lee 2024-11-20 09:08:56 -08:00
  • f4648b0b26 fix relative path Justin Lee 2024-11-13 13:14:43 -08:00
  • b18077b7a9 rebase to main Justin Lee 2024-11-20 09:05:42 -08:00
  • ac1791f8b1 Adding memory provider mocks Vladimir Ivic 2024-11-20 08:56:13 -08:00
  • 8a35dc8b0e Merge branch 'main' into add-nvidia-inference-adapter Matthew Farrellee 2024-11-20 09:37:48 -05:00
  • ac244d18c1 for pgvector Dinesh Yeduguru 2024-11-20 00:03:15 -08:00
  • 49a440acfb fall to back to read from chroma when not in cache Dinesh Yeduguru 2024-11-19 23:59:15 -08:00
  • 89f5093dfc Fix tgi doc Ashwin Bharambe 2024-11-19 21:05:59 -08:00
  • 1086b500f9
    Support Tavily as built-in search tool. (#485) Mengtao Yuan 2024-11-19 20:59:02 -08:00
  • 9b76224c28 Support Tavily as built-in search tool. Martin Yuan 2024-11-18 17:59:50 -08:00
  • 08be023290
    Added optional md5 validate command once download is completed (#486) varunfb 2024-11-19 17:42:43 -08:00
  • e670f99ef7
    add changelog (#487) Dinesh Yeduguru 2024-11-19 17:36:08 -08:00
  • 1bd43fd951 fixes Dinesh Yeduguru 2024-11-19 17:13:40 -08:00
  • ee7e5c0379 fixes Dinesh Yeduguru 2024-11-19 17:12:41 -08:00
  • c50a3f70a5 remove development and bug fixes sections Dinesh Yeduguru 2024-11-19 17:11:03 -08:00
  • 21cd86913f address feedback from review Dinesh Yeduguru 2024-11-19 17:08:30 -08:00
  • f3c6cbfd53 add evals work Dinesh Yeduguru 2024-11-19 17:06:54 -08:00
  • 7df97445a2 add changelog Dinesh Yeduguru 2024-11-19 17:02:00 -08:00
  • dd5466e17d Bump version to 0.0.53 Ashwin Bharambe 2024-11-19 16:44:15 -08:00
  • b0fdf7552a docs Xi Yan 2024-11-19 16:41:45 -08:00
  • c49acc5226 docs Xi Yan 2024-11-19 16:39:40 -08:00
  • f78200b189 docs Xi Yan 2024-11-19 16:37:30 -08:00
  • e605d57fb7 use API version in "remote" stack client Ashwin Bharambe 2024-11-19 15:59:47 -08:00
  • 7bfcfe80b5 Add logs (prints :/) to dump out what URL vllm / tgi is connecting to Ashwin Bharambe 2024-11-19 15:50:26 -08:00
  • 887ccc2143 Ensure llama-stack-client is installed in the container with TEST_PYPI Ashwin Bharambe 2024-11-19 15:20:51 -08:00
  • 4ccf4ef641 align with other remote adapters, rename config base_url -> url Matthew Farrellee 2024-11-19 17:36:08 -05:00
  • 2da93c8835 fix 3.2-1b fireworks Xi Yan 2024-11-19 14:20:07 -08:00
  • 189df6358a codegen docs Xi Yan 2024-11-19 14:16:00 -08:00
  • 185df4b568 fix fireworks registration Xi Yan 2024-11-19 14:09:00 -08:00
  • ae99ceae69 Fixed precommit lint issues varunfb 2024-11-19 13:45:29 -08:00
  • 38ba3b9f0c Fix fireworks stream completion Ashwin Bharambe 2024-11-19 13:36:14 -08:00
  • 4277e24661 Fixed text for path to md5 file location varunfb 2024-11-19 13:30:36 -08:00
  • 05d1ead02f Update condition in tests to handle llama-3.1 vs llama3.1 (HF names) Ashwin Bharambe 2024-11-19 13:25:36 -08:00
  • a5d413045c Add nvidia remote distro Chantal D Gama Rose 2024-11-19 21:02:20 +00:00
  • 42acff502c Added optional md5 validate command once download is completed varunfb 2024-11-19 12:05:34 -08:00
  • 394519d68a Add llama-stack-client as a legitimate dependency for llama-stack Ashwin Bharambe 2024-11-19 11:44:35 -08:00
  • c46b462c22 Updates to docker build script Ashwin Bharambe 2024-11-19 11:36:53 -08:00
  • 39e99b39fe
    update quick start to have the working instruction (#467) Henry Tai 2024-11-20 02:32:19 +08:00
  • 1b0f5fff5a fix curl endpoint Xi Yan 2024-11-19 10:26:05 -08:00
  • 6bf6c79bd6 update quick start to have the working instruction Henry Tai 2024-11-17 12:38:01 +08:00
  • 18e8f18749 Merge remote-tracking branch 'mattf/add-nvidia-inference-adapter' into cdgamarose/add_nvidia_distro merging matt's changes Chantal D Gama Rose 2024-11-19 17:59:50 +00:00
  • 1619d37cc6 codegen per-distro dependencies; not hooked into setup.py yet Ashwin Bharambe 2024-11-19 09:54:30 -08:00
  • 52e470cb23 map llama model -> provider model id in ModelRegistryHelper Matthew Farrellee 2024-11-19 12:49:14 -05:00
  • 2980a18920 map llama model -> provider model id in ModelRegistryHelper Matthew Farrellee 2024-11-19 12:49:14 -05:00
  • 5e4ac1b7c1 Make sure server code uses version prefixed routes Ashwin Bharambe 2024-11-19 09:15:05 -08:00
  • 2a25ace2fa Merge branch 'main' into add-nvidia-inference-adapter Matthew Farrellee 2024-11-19 10:25:50 -05:00
  • 88979c1815 Update Ollama supported llama model list Martin Hickey 2024-11-19 16:18:48 +00:00
  • 1f3f0f9f4f Add cache for PGVector memory adapter Martin Hickey 2024-11-19 15:32:54 +00:00
  • c24f882f31 Merge branch 'main' into add-nvidia-inference-adapter Matthew Farrellee 2024-11-17 15:47:13 -05:00
  • 84d5f35a48 Update the model alias for llama guard models in ollama Ashwin Bharambe 2024-11-19 00:22:24 -08:00
  • e8d3eee095 Fix docs yet again Ashwin Bharambe 2024-11-18 23:51:25 -08:00
  • 02f1c47416
    support adding alias for models without hf repo/sku entry (#481) Dinesh Yeduguru 2024-11-18 23:50:18 -08:00
  • ddb0ee23a5 remove x/ Dinesh Yeduguru 2024-11-18 23:49:27 -08:00
  • d04c9d9e4e change function name Dinesh Yeduguru 2024-11-18 23:43:47 -08:00
  • 8ed79ad0f3 Fix the pyopenapi generator avoid potential circular imports Ashwin Bharambe 2024-11-18 23:37:52 -08:00
  • 8bd0a33206 support adding alias for models without hf repo/sku entry Dinesh Yeduguru 2024-11-18 23:31:04 -08:00
  • d463d68e1e Update docs Ashwin Bharambe 2024-11-18 23:21:25 -08:00
  • 93abb8e208 Include all yamls Ashwin Bharambe 2024-11-18 22:46:07 -08:00
  • 0dc7f5fa89
    Add version to REST API url (#478) Ashwin Bharambe 2024-11-18 22:44:14 -08:00
  • 62a21a3072 add comment for the noqa Ashwin Bharambe 2024-11-18 22:43:23 -08:00
  • c73fe390ad Add version to REST API url Ashwin Bharambe 2024-11-18 18:53:06 -08:00