Commit graph

  • edfd92d81f
    Merge branch 'meta-llama:main' into main Chacksu 2024-11-20 16:12:38 -05:00
  • 681322731b
    Make run yaml optional so dockers can start with just --env (#492) Ashwin Bharambe 2024-11-20 13:11:40 -08:00
  • 990a997405 Make sure llama stack run ollama also works correctly Ashwin Bharambe 2024-11-20 12:52:43 -08:00
  • 15200e23ba address feedback Dinesh Yeduguru 2024-11-20 12:19:12 -08:00
  • 21a11b2bf7 restructure docs: Dinesh Yeduguru 2024-11-20 12:06:22 -08:00
  • adbb605a50 Make run yaml optional so dockers can start with just --env Ashwin Bharambe 2024-11-20 11:56:08 -08:00
  • 1d8d0593af
    register with provider even if present in stack (#491) Dinesh Yeduguru 2024-11-20 11:05:50 -08:00
  • 7197d73b12 add a new getting started Dinesh Yeduguru 2024-11-19 22:34:08 -08:00
  • 57fd974872 register with provider even if present in stack Dinesh Yeduguru 2024-11-20 10:42:43 -08:00
  • 91e7efbc91
    fall to back to read from chroma/pgvector when not in cache (#489) Dinesh Yeduguru 2024-11-20 10:30:23 -08:00
  • ae49a4cb97
    Reorganizing Zero to Hero Folder structure (#447) Justin Lee 2024-11-20 10:27:29 -08:00
  • 8d5fdeedda add _get_and_cache_bank_index Dinesh Yeduguru 2024-11-20 10:21:55 -08:00
  • a68e58f873 update notebooks, remove colab links Justin Lee 2024-11-20 09:08:56 -08:00
  • f4648b0b26 fix relative path Justin Lee 2024-11-13 13:14:43 -08:00
  • b18077b7a9 rebase to main Justin Lee 2024-11-20 09:05:42 -08:00
  • ac1791f8b1 Adding memory provider mocks Vladimir Ivic 2024-11-20 08:56:13 -08:00
  • 8a35dc8b0e Merge branch 'main' into add-nvidia-inference-adapter Matthew Farrellee 2024-11-20 09:37:48 -05:00
  • ac244d18c1 for pgvector Dinesh Yeduguru 2024-11-20 00:03:15 -08:00
  • 49a440acfb fall to back to read from chroma when not in cache Dinesh Yeduguru 2024-11-19 23:59:15 -08:00
  • 89f5093dfc Fix tgi doc Ashwin Bharambe 2024-11-19 21:05:59 -08:00
  • 1086b500f9
    Support Tavily as built-in search tool. (#485) Mengtao Yuan 2024-11-19 20:59:02 -08:00
  • 9b76224c28 Support Tavily as built-in search tool. Martin Yuan 2024-11-18 17:59:50 -08:00
  • 08be023290
    Added optional md5 validate command once download is completed (#486) varunfb 2024-11-19 17:42:43 -08:00
  • e670f99ef7
    add changelog (#487) Dinesh Yeduguru 2024-11-19 17:36:08 -08:00
  • 1bd43fd951 fixes Dinesh Yeduguru 2024-11-19 17:13:40 -08:00
  • ee7e5c0379 fixes Dinesh Yeduguru 2024-11-19 17:12:41 -08:00
  • c50a3f70a5 remove development and bug fixes sections Dinesh Yeduguru 2024-11-19 17:11:03 -08:00
  • 21cd86913f address feedback from review Dinesh Yeduguru 2024-11-19 17:08:30 -08:00
  • f3c6cbfd53 add evals work Dinesh Yeduguru 2024-11-19 17:06:54 -08:00
  • 7df97445a2 add changelog Dinesh Yeduguru 2024-11-19 17:02:00 -08:00
  • dd5466e17d Bump version to 0.0.53 Ashwin Bharambe 2024-11-19 16:44:15 -08:00
  • b0fdf7552a docs Xi Yan 2024-11-19 16:41:45 -08:00
  • c49acc5226 docs Xi Yan 2024-11-19 16:39:40 -08:00
  • f78200b189 docs Xi Yan 2024-11-19 16:37:30 -08:00
  • e605d57fb7 use API version in "remote" stack client Ashwin Bharambe 2024-11-19 15:59:47 -08:00
  • 7bfcfe80b5 Add logs (prints :/) to dump out what URL vllm / tgi is connecting to Ashwin Bharambe 2024-11-19 15:50:26 -08:00
  • 887ccc2143 Ensure llama-stack-client is installed in the container with TEST_PYPI Ashwin Bharambe 2024-11-19 15:20:51 -08:00
  • 4ccf4ef641 align with other remote adapters, rename config base_url -> url Matthew Farrellee 2024-11-19 17:36:08 -05:00
  • 2da93c8835 fix 3.2-1b fireworks Xi Yan 2024-11-19 14:20:07 -08:00
  • 189df6358a codegen docs Xi Yan 2024-11-19 14:16:00 -08:00
  • 185df4b568 fix fireworks registration Xi Yan 2024-11-19 14:09:00 -08:00
  • ae99ceae69 Fixed precommit lint issues varunfb 2024-11-19 13:45:29 -08:00
  • 38ba3b9f0c Fix fireworks stream completion Ashwin Bharambe 2024-11-19 13:36:14 -08:00
  • 4277e24661 Fixed text for path to md5 file location varunfb 2024-11-19 13:30:36 -08:00
  • 05d1ead02f Update condition in tests to handle llama-3.1 vs llama3.1 (HF names) Ashwin Bharambe 2024-11-19 13:25:36 -08:00
  • a5d413045c Add nvidia remote distro Chantal D Gama Rose 2024-11-19 21:02:20 +00:00
  • 42acff502c Added optional md5 validate command once download is completed varunfb 2024-11-19 12:05:34 -08:00
  • 394519d68a Add llama-stack-client as a legitimate dependency for llama-stack Ashwin Bharambe 2024-11-19 11:44:35 -08:00
  • c46b462c22 Updates to docker build script Ashwin Bharambe 2024-11-19 11:36:53 -08:00
  • 39e99b39fe
    update quick start to have the working instruction (#467) Henry Tai 2024-11-20 02:32:19 +08:00
  • 1b0f5fff5a fix curl endpoint Xi Yan 2024-11-19 10:26:05 -08:00
  • 6bf6c79bd6 update quick start to have the working instruction Henry Tai 2024-11-17 12:38:01 +08:00
  • 18e8f18749 Merge remote-tracking branch 'mattf/add-nvidia-inference-adapter' into cdgamarose/add_nvidia_distro merging matt's changes Chantal D Gama Rose 2024-11-19 17:59:50 +00:00
  • 1619d37cc6 codegen per-distro dependencies; not hooked into setup.py yet Ashwin Bharambe 2024-11-19 09:54:30 -08:00
  • 52e470cb23 map llama model -> provider model id in ModelRegistryHelper Matthew Farrellee 2024-11-19 12:49:14 -05:00
  • 2980a18920 map llama model -> provider model id in ModelRegistryHelper Matthew Farrellee 2024-11-19 12:49:14 -05:00
  • 5e4ac1b7c1 Make sure server code uses version prefixed routes Ashwin Bharambe 2024-11-19 09:15:05 -08:00
  • 2a25ace2fa Merge branch 'main' into add-nvidia-inference-adapter Matthew Farrellee 2024-11-19 10:25:50 -05:00
  • 88979c1815 Update Ollama supported llama model list Martin Hickey 2024-11-19 16:18:48 +00:00
  • 1f3f0f9f4f Add cache for PGVector memory adapter Martin Hickey 2024-11-19 15:32:54 +00:00
  • c24f882f31 Merge branch 'main' into add-nvidia-inference-adapter Matthew Farrellee 2024-11-17 15:47:13 -05:00
  • 84d5f35a48 Update the model alias for llama guard models in ollama Ashwin Bharambe 2024-11-19 00:22:24 -08:00
  • e8d3eee095 Fix docs yet again Ashwin Bharambe 2024-11-18 23:51:25 -08:00
  • 02f1c47416
    support adding alias for models without hf repo/sku entry (#481) Dinesh Yeduguru 2024-11-18 23:50:18 -08:00
  • ddb0ee23a5 remove x/ Dinesh Yeduguru 2024-11-18 23:49:27 -08:00
  • d04c9d9e4e change function name Dinesh Yeduguru 2024-11-18 23:43:47 -08:00
  • 8ed79ad0f3 Fix the pyopenapi generator avoid potential circular imports Ashwin Bharambe 2024-11-18 23:37:52 -08:00
  • 8bd0a33206 support adding alias for models without hf repo/sku entry Dinesh Yeduguru 2024-11-18 23:31:04 -08:00
  • d463d68e1e Update docs Ashwin Bharambe 2024-11-18 23:21:25 -08:00
  • 93abb8e208 Include all yamls Ashwin Bharambe 2024-11-18 22:46:07 -08:00
  • 0dc7f5fa89
    Add version to REST API url (#478) Ashwin Bharambe 2024-11-18 22:44:14 -08:00
  • 62a21a3072 add comment for the noqa Ashwin Bharambe 2024-11-18 22:43:23 -08:00
  • c73fe390ad Add version to REST API url Ashwin Bharambe 2024-11-18 18:53:06 -08:00
  • 05e93bd2f7 together default Xi Yan 2024-11-18 22:39:45 -08:00
  • 7693786322 Use HF names for registering fireworks and together models Ashwin Bharambe 2024-11-18 22:34:26 -08:00
  • 6765fd76ff
    fix llama stack build for together & llama stack build from templates (#479) Xi Yan 2024-11-18 22:29:16 -08:00
  • f967a740be new build msg from template Xi Yan 2024-11-18 22:22:58 -08:00
  • ea52a3ee1c minor enhancement for test fixtures Ashwin Bharambe 2024-11-18 22:20:59 -08:00
  • d25cae3995 fix together build Xi Yan 2024-11-18 22:06:10 -08:00
  • e6ae344eb4 fix together build Xi Yan 2024-11-18 21:42:18 -08:00
  • a042f41d03 fix core model ids for ollama Dinesh Yeduguru 2024-11-18 20:55:54 -08:00
  • fcc2132e6f
    remove pydantic namespace warnings using model_config (#470) Matthew Farrellee 2024-11-18 22:24:14 -05:00
  • 2108a779f2
    Update kotlin client docs (#476) Riandy 2024-11-18 19:13:20 -08:00
  • d2b7c5aeae
    add quantized model ollama support (#471) Kai Wu 2024-11-18 18:55:23 -08:00
  • 14c75c3f21 Update CONTRIBUTING to include info about pre-commit Ashwin Bharambe 2024-11-18 18:17:41 -08:00
  • fe19076838
    get stack run config based on template name (#477) Dinesh Yeduguru 2024-11-18 18:05:05 -08:00
  • edd946a207 remove async Dinesh Yeduguru 2024-11-18 18:02:32 -08:00
  • 50d539e6d7 update tests --inference-model to hf id Xi Yan 2024-11-18 17:36:58 -08:00
  • 939056e265 More documentation fixes Ashwin Bharambe 2024-11-18 17:06:13 -08:00
  • e40404625b Update to docs Ashwin Bharambe 2024-11-18 16:52:48 -08:00
  • 41069b4b05 fix on rebase Dinesh Yeduguru 2024-11-18 16:22:33 -08:00
  • 511346cd41 get stack run config based on template name Dinesh Yeduguru 2024-11-18 16:07:11 -08:00
  • 91f3009c67 No more built_at Ashwin Bharambe 2024-11-18 16:38:51 -08:00
  • afa4f0b19f Update remote vllm docs Ashwin Bharambe 2024-11-18 16:34:33 -08:00
  • fb15ff4a97 Move to use argparse, fix issues with multiple --env cmdline options Ashwin Bharambe 2024-11-18 16:31:59 -08:00
  • b87f3ac499 Allow server to accept --env key pairs Ashwin Bharambe 2024-11-18 16:17:59 -08:00
  • 1fb61137ad Add conda_env Ashwin Bharambe 2024-11-18 16:08:03 -08:00
  • b822149098 Update start conda Ashwin Bharambe 2024-11-18 16:07:27 -08:00
  • 47c37fd831 Fixes Ashwin Bharambe 2024-11-18 16:03:20 -08:00
  • 3aedde2ab4 Add a pre-commit for distro_codegen but it does not work yet Ashwin Bharambe 2024-11-18 15:20:49 -08:00