mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 15:23:51 +00:00
sdg improvements
This commit is contained in:
parent
c1f6816d76
commit
d9367054df
1 changed files with 5 additions and 10 deletions
|
@ -283,11 +283,9 @@ class FilteringFunction(Enum):
|
|||
class SyntheticDataGenerationRequest:
|
||||
"""Request to generate synthetic data. A small batch of prompts and a filtering function"""
|
||||
|
||||
prompts: List[str]
|
||||
prompts: List[Message]
|
||||
filtering_function: FilteringFunction = FilteringFunction.none
|
||||
|
||||
# TODO(ragho): fix this
|
||||
# reward_scoring: RewardScoring
|
||||
reward_scoring: Optional[RewardScoring] = None
|
||||
|
||||
|
||||
@json_schema_type
|
||||
|
@ -295,11 +293,8 @@ class SyntheticDataGenerationRequest:
|
|||
class SyntheticDataGenerationResponse:
|
||||
"""Response from the synthetic data generation. Batch of (prompt, response, score) tuples that pass the threshold."""
|
||||
|
||||
synthetic_data: List[Tuple[str, str, float]]
|
||||
|
||||
"""The actual synthetic data"""
|
||||
statistics: Dict[str, float]
|
||||
"""Statistics on how many prompts were generated and how many were filtered out"""
|
||||
synthetic_data: List[KScoredPromptGenerations]
|
||||
statistics: Optional[Dict[str, Any]] = None
|
||||
|
||||
|
||||
class SyntheticDataGeneration(Protocol):
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue