Commit graph

  • 4884c62190 Merge branch 'main' into watsonx-infer-fix Sajikumar JS 2025-04-26 18:31:59 +05:30
  • a9e4d1f00e pre-commit updates Sajikumar JS 2025-04-26 18:29:47 +05:30
  • 7f1e4bf075 Updated parameters Sajikumar JS 2025-04-26 18:29:06 +05:30
  • 3588c5bcd7 Updated SamplingParams Sajikumar JS 2025-04-26 18:25:58 +05:30
  • 0266b20535
    docs: update prompt_format.md for llama4 (#2035) ehhuang 2025-04-25 15:52:15 -07:00
  • 045836ebfc doc: update prompt_format.md for llama4 Eric Huang 2025-04-25 15:48:06 -07:00
  • 1e8fce126f build: Bump version to 0.2.3 v0.2.3 release-0.2.3 Ashwin Bharambe 2025-04-25 15:38:49 -07:00
  • bb1a85c9a0 fix: make sure test works equally well against llama stack as a server Ashwin Bharambe 2025-04-25 15:23:53 -07:00
  • 3ca284a52b Release candidate 0.2.3rc5 v0.2.3rc5 github-actions[bot] 2025-04-25 22:07:01 +00:00
  • 8713d67ce3
    fix: Correctly parse algorithm_config when launching NVIDIA customization job; fix internal request handler (#2025) Jash Gulabrai 2025-04-25 16:21:50 -04:00
  • b5d8e44e81 fix: only sleep for tests when they pass or fail Ashwin Bharambe 2025-04-25 13:15:52 -07:00
  • 1b2e116a2a
    fix: tool call encoded twice (#2034) ehhuang 2025-04-25 13:16:16 -07:00
  • dd349b2176 Remove check to parse either dict or pydantic model Jash Gulabrai 2025-04-25 16:13:39 -04:00
  • ff6081a353 fix: tool call encoded twice Eric Huang 2025-04-25 13:05:06 -07:00
  • 6659ed995a Merge branch 'main' into fix/nvidia-launch-customization Jash Gulabrai 2025-04-25 16:01:22 -04:00
  • bb142435db Use correct shapes in unit tests; remove use of unsupported params Jash Gulabrai 2025-04-25 15:52:12 -04:00
  • 03a25a7753 updated the additional params to pass any type of values Sajikumar JS 2025-04-26 01:19:37 +05:30
  • 1d6ef73dd7 added additional params and new functions required to watsonx Sajikumar JS 2025-04-26 01:09:46 +05:30
  • 4fb583b407
    fix: check that llama stack client plain can be used as a subst for OpenAI client (#2032) Ashwin Bharambe 2025-04-25 12:23:33 -07:00
  • 0e4307de0f
    docs: Fix missing --gpu all flag in Docker run commands (#2026) Derek Higgins 2025-04-25 20:17:31 +01:00
  • 1deab94ea0
    chore: exclude test, provider, and template directories from coverage (#2028) Sébastien Han 2025-04-25 21:16:57 +02:00
  • a233bdc76e add unit tests for content from doc Kevin 2025-04-25 15:09:33 -04:00
  • 7b34153fff fix: check that llama stack client plain can be used as a subst for OpenAI client Ashwin Bharambe 2025-04-25 12:13:20 -07:00
  • cfc6bdae68 llama-4-scout-17b-16e-instruct passing tests Matt Clayton 2025-04-25 13:38:40 -04:00
  • 6135bdec22 add tests/verification/conf/lmstudio.yaml Neil Mehta 2025-04-22 13:09:36 -04:00
  • 6377b1912b Revert "Use int for year in test case" Neil Mehta 2025-03-24 18:54:52 -04:00
  • 357d7ea9ea Use int for year in test case Neil Mehta 2025-03-24 17:57:48 -04:00
  • 00affd1f02 Fix async streaming Neil Mehta 2025-03-24 14:10:49 -04:00
  • 05777dfb52 implement error handling, improve completion, tool calling and streaming Justin Lee 2025-03-21 16:52:32 -07:00
  • fe575a0fdf Update report.md to reflect current version support Rugved Somwanshi 2025-03-19 18:35:32 -04:00
  • a0ff1f0464 Update README.md Rugved Somwanshi 2025-03-18 17:31:20 -04:00
  • 302d72cc47 Fix python3.10 async Neil Mehta 2025-03-18 15:53:41 -04:00
  • aa9562e104 Addressed comments Rugved Somwanshi 2025-03-14 16:33:53 -04:00
  • 1a5cfd1b6f Fix stream generate Neil Mehta 2025-03-14 15:51:12 -04:00
  • 9c83ca415d Fix lmstudio name Neil Mehta 2025-03-14 15:40:50 -04:00
  • 461eec425d LM Studio inference integration Neil Mehta 2025-03-14 15:21:15 -04:00
  • 1bb1d9b2ba
    feat: Add watsonx inference adapter (#1895) Sajikumar JS 2025-04-25 23:59:21 +05:30
  • 29072f40ab
    feat: new system prompt for llama4 (#2031) ehhuang 2025-04-25 11:29:08 -07:00
  • 17edf138e8 new prompt Eric Huang 2025-04-25 11:15:04 -07:00
  • 8e9217774a new prompt Eric Huang 2025-04-24 13:16:42 -07:00
  • 4bbd0c0693 fix: add endpoint route debugs Ashwin Bharambe 2025-04-25 10:39:30 -07:00
  • 8409109ca7 docs(readme): add one-line installer snippet reluctantfuturist 2025-04-25 10:01:55 -07:00
  • bed5a9f55a chore(installer): remove ollama-models bind-mount for a stateless install reluctantfuturist 2025-04-25 09:39:19 -07:00
  • a5a842fa76 feat(installer): dump container logs on health-check failure reluctantfuturist 2025-04-24 10:26:59 -07:00
  • 0ae46f9417 chore(installer): fully silence container output by redirecting stderr Alexey Rybak 2025-04-24 10:21:07 -07:00
  • 876fd6e80b chore(ci): refine shellcheck reluctantfuturist 2025-04-23 11:50:16 -07:00
  • 6a135e80c7 chore(ci): refine shellcheck reluctantfuturist 2025-04-23 11:39:01 -07:00
  • b67940e5cc ci(installer): pin actions to SHAs, add ShellCheck, drop redundant steps reluctantfuturist 2025-04-23 11:27:56 -07:00
  • 19ad7ba513 chore(ci): remove redundant steps and simplify network setup reluctantfuturist 2025-04-18 15:00:25 -07:00
  • d4e5d4c1fa chore(installer): make install.sh executable in repo reluctantfuturist 2025-04-18 13:51:13 -07:00
  • d843672482 ci(installer): end-to-end smoke test and scheduled health check reluctantfuturist 2025-04-17 15:45:14 -07:00
  • 9edda486ae feat(installer): add Podman support, bump default model to 3b, document internal DNS reluctantfuturist 2025-04-17 11:37:44 -07:00
  • e70c4e67de refactor(install): simplify demo to two‑container flow, drop host‑level installs reluctantfuturist 2025-04-16 12:05:58 -07:00
  • 6d5d1480c9 ollama distro install reluctantfuturist 2025-03-04 20:46:35 -08:00
  • 6d4a4438ac feat(install): add Llama Stack Meta Reference installation script reluctantfuturist 2025-03-03 22:47:16 -08:00
  • f5dae0517c
    feat: Support ReAct Agent on Tools Playground (#2012) Andy Xie 2025-04-25 11:01:51 -04:00
  • 121c73c2f5
    feat(cli): add interactive tab completion for image type selection (#2027) Roland Huß 2025-04-25 16:57:42 +02:00
  • 92142fe0cd
    Maded changes on streamlit with ReAct agent mode Andy Xie 2025-04-24 02:01:09 -04:00
  • 74bacbe77b
    fixed precommit check Roland Huß 2025-04-25 15:40:30 +02:00
  • c71006f9cc
    chore: exclude test, provider, and template directories from coverage Sébastien Han 2025-04-25 14:14:29 +02:00
  • 8f701a831d
    feat(cli): add interactive tab completion for image type selection Roland Huß 2025-04-25 13:55:59 +02:00
  • ec4c04fa2d docs: Fix missing --gpu all flag in Docker run commands Derek Higgins 2025-04-25 12:06:08 +01:00
  • 7397534497 docs: Remove docs for meta-reference-quantized-gpu Derek Higgins 2025-04-25 12:33:06 +01:00
  • 59b7593609
    feat: Enhance tool display in Tools sidebar by simplifying tool identifiers (#2024) Surya Prakash Pathak 2025-04-25 01:22:22 -07:00
  • d9e00fca66
    fix: specify nbformat version in nb (#2023) Kevin Postlethwait 2025-04-25 04:10:37 -04:00
  • 2f498ad776 updated the string to watsonx Sajikumar JS 2025-04-25 11:03:21 +05:30
  • 6fe8b292b1 Merge branch 'main' into add-watsonx-inference-adapter Sajikumar JS 2025-04-25 10:57:45 +05:30
  • ace82836c1
    feat: NVIDIA allow non-llama model registration (#1859) Rashmi Pawar 2025-04-25 05:43:33 +05:30
  • cc77f79f55
    feat: Add NVIDIA Eval integration (#1890) Jash Gulabrai 2025-04-24 20:12:42 -04:00
  • 0b6cd45950
    fix: Additional streaming error handling (#2007) Ben Browning 2025-04-24 20:01:45 -04:00
  • c8797f1125
    fix: Including tool call in chat (#1931) Derek Higgins 2025-04-25 00:59:10 +01:00
  • 26c10b5ab5 fix: Correctly parse algorithm_config when launching NVIDIA customization job Jash Gulabrai 2025-04-24 17:17:00 -04:00
  • f557e16d0c Enhance tool display in Tools sidebar by simplifying tool identifiers Surya Prakash Pathak 2025-04-24 13:26:40 -07:00
  • 9d05aff960 specify nbformat version in nb Kevin 2025-04-24 15:51:07 -04:00
  • 7ed137e963
    fix: meta ref inference (#2022) ehhuang 2025-04-24 13:03:35 -07:00
  • 9b58624479 fixes Eric Huang 2025-04-24 12:08:11 -07:00
  • 693c709c27 fixes Eric Huang 2025-04-24 12:03:55 -07:00
  • a452c64300 docs: Add TrustyAI LM-Eval to list of known external providers Christina Xu 2025-04-24 14:45:48 -04:00
  • a5d6ab16b2 fix: meta-reference parallel utils bug, use isinstance not equality Ashwin Bharambe 2025-04-24 11:27:49 -07:00
  • 0990f60dad Merge branch 'main' of https://github.com/meta-llama/llama-stack into register_custom_model raspawar 2025-04-24 21:44:32 +05:30
  • 70488abe9c
    chore: Remove distributions/** from integration, external provider, and unit tests (#2018) Francisco Arceo 2025-04-24 09:39:31 -06:00
  • 516f8dffb4 chore: Remove distributions/** from integration, external provider, and unit tests Francisco Javier Arceo 2025-04-24 11:31:42 -04:00
  • dc0d4763a0
    chore: Update External Providers CI to not run on changes to docs, rfcs, and scripts (#2009) Francisco Arceo 2025-04-24 09:24:07 -06:00
  • 8d7225507f removing integration tests Francisco Javier Arceo 2025-04-24 11:02:54 -04:00
  • e24959ea9e Fix variable name Jash Gulabrai 2025-04-24 10:41:38 -04:00
  • e664ba91d8
    fix: prevent the knowledge search tool from confusing the model with long content (#1908) Ilya Kolchinsky 2025-04-24 16:38:38 +02:00
  • edbf85366d adding workflow changes as well Francisco Javier Arceo 2025-04-24 09:35:26 -04:00
  • 2f91b73bcb
    feat(pre-commit): enhance pre-commit hooks with additional checks Sébastien Han 2025-04-24 14:54:04 +02:00
  • f7257e33dc removing integration-tests Francisco Javier Arceo 2025-04-24 08:57:53 -04:00
  • bcc77f829c swapping to paths instead of paths-ignore to be consistent with the integragtion and unit tests Francisco Javier Arceo 2025-04-24 08:56:55 -04:00
  • 4ac7ead646 remove extra space Francisco Javier Arceo 2025-04-24 08:01:32 -04:00
  • 294e91724a
    fix: do not override the entrypoint when running container Sébastien Han 2025-04-24 13:56:08 +02:00
  • 14e60e3c02
    feat: include run.yaml in the container image (#2005) Sébastien Han 2025-04-24 11:29:53 +02:00
  • a487b0b902
    feat: include run.yaml in the container image Sébastien Han 2025-04-23 11:04:29 +02:00
  • a673697858
    chore: rename ramalama provider (#2008) Charlie Doern 2025-04-24 03:34:15 -04:00
  • 803b836323 chore: Update External Providers CI to not run on changes to docs, rfcs, and scripts Francisco Javier Arceo 2025-04-23 22:24:57 -04:00
  • 050d4b8cc9 chore: rename ramalama provider Charlie Doern 2025-04-23 15:34:46 -04:00
  • cfa4b61a01 fix: Additional streaming error handling Ben Browning 2025-04-23 13:04:16 -04:00
  • fa5dfee07b
    fix: Return HTTP 400 for OpenAI API validation errors (#2002) Ben Browning 2025-04-23 11:48:32 -04:00
  • 6a44e7ba20
    docs: add API to external providers table (#2006) Nathan Weinberg 2025-04-23 09:58:10 -04:00