mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-15 14:08:00 +00:00
more fixes
This commit is contained in:
parent
1302aabc60
commit
0b6a7abb28
15 changed files with 18 additions and 15 deletions
|
@ -1,4 +1,4 @@
|
||||||
# Agents
|
# Agents
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
# Datasetio
|
# Datasetio
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
# Eval
|
# Eval
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
# Files
|
# Files
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
# Inference
|
# Inference
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
# Post_Training
|
# Post_Training
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
|
|
|
@ -27,7 +27,7 @@ HuggingFace-based post-training provider for fine-tuning models using the Huggin
|
||||||
| `dpo_beta` | `<class 'float'>` | No | 0.1 | |
|
| `dpo_beta` | `<class 'float'>` | No | 0.1 | |
|
||||||
| `use_reference_model` | `<class 'bool'>` | No | True | |
|
| `use_reference_model` | `<class 'bool'>` | No | True | |
|
||||||
| `dpo_loss_type` | `Literal['sigmoid', 'hinge', 'ipo', 'kto_pair'` | No | sigmoid | |
|
| `dpo_loss_type` | `Literal['sigmoid', 'hinge', 'ipo', 'kto_pair'` | No | sigmoid | |
|
||||||
| `dpo_output_dir` | `<class 'str'>` | No | ./checkpoints/dpo | |
|
| `dpo_output_dir` | `<class 'str'>` | No | /var/folders/mj/t_stv1ys7637vqf2_b4yf67m0000gn/T/dpo_output_q24jwy0r | |
|
||||||
|
|
||||||
## Sample Configuration
|
## Sample Configuration
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
# Safety
|
# Safety
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
# Scoring
|
# Scoring
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
# Telemetry
|
# Telemetry
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
# Tool_Runtime
|
# Tool_Runtime
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
# Vector_Io
|
# Vector_Io
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
|
|
||||||
|
|
|
@ -4,10 +4,10 @@
|
||||||
# This source code is licensed under the terms described in the LICENSE file in
|
# This source code is licensed under the terms described in the LICENSE file in
|
||||||
# the root directory of this source tree.
|
# the root directory of this source tree.
|
||||||
|
|
||||||
|
import tempfile
|
||||||
from typing import Any, Literal
|
from typing import Any, Literal
|
||||||
|
|
||||||
from pydantic import BaseModel, Field
|
from pydantic import BaseModel, Field
|
||||||
import tempfile
|
|
||||||
|
|
||||||
|
|
||||||
class HuggingFacePostTrainingConfig(BaseModel):
|
class HuggingFacePostTrainingConfig(BaseModel):
|
||||||
|
|
|
@ -255,7 +255,7 @@ def process_provider_registry(progress, change_tracker: ChangedPathTracker) -> N
|
||||||
change_tracker.add_paths(doc_output_dir)
|
change_tracker.add_paths(doc_output_dir)
|
||||||
|
|
||||||
index_content = []
|
index_content = []
|
||||||
index_content.append(f"# {api_name.title()} \n")
|
index_content.append(f"# {api_name.title()}\n")
|
||||||
index_content.append("## Overview\n")
|
index_content.append("## Overview\n")
|
||||||
|
|
||||||
index_content.append(
|
index_content.append(
|
||||||
|
|
|
@ -4,7 +4,7 @@
|
||||||
# This source code is licensed under the terms described in the LICENSE file in
|
# This source code is licensed under the terms described in the LICENSE file in
|
||||||
# the root directory of this source tree.
|
# the root directory of this source tree.
|
||||||
import pytest
|
import pytest
|
||||||
from openai import BadRequestError
|
from openai import BadRequestError, OpenAI
|
||||||
|
|
||||||
from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
||||||
|
|
||||||
|
@ -36,6 +36,9 @@ from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
def test_responses_store(compat_client, text_model_id, stream, tools):
|
def test_responses_store(compat_client, text_model_id, stream, tools):
|
||||||
|
if not isinstance(compat_client, OpenAI):
|
||||||
|
pytest.skip("OpenAI client is required until responses.delete() exists in llama-stack-client")
|
||||||
|
|
||||||
message = "What's the weather in Tokyo?" + (
|
message = "What's the weather in Tokyo?" + (
|
||||||
" YOU MUST USE THE get_weather function to get the weather." if tools else ""
|
" YOU MUST USE THE get_weather function to get the weather." if tools else ""
|
||||||
)
|
)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue