mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-28 23:04:21 +00:00
sdg improvements
This commit is contained in:
parent
c1f6816d76
commit
d9367054df
1 changed files with 5 additions and 10 deletions
|
@ -283,23 +283,18 @@ class FilteringFunction(Enum):
|
|||
class SyntheticDataGenerationRequest:
|
||||
"""Request to generate synthetic data. A small batch of prompts and a filtering function"""
|
||||
|
||||
prompts: List[str]
|
||||
prompts: List[Message]
|
||||
filtering_function: FilteringFunction = FilteringFunction.none
|
||||
|
||||
# TODO(ragho): fix this
|
||||
# reward_scoring: RewardScoring
|
||||
reward_scoring: Optional[RewardScoring] = None
|
||||
|
||||
|
||||
@json_schema_type
|
||||
@dataclass
|
||||
class SyntheticDataGenerationResponse:
|
||||
"""Response from the synthetic data generation. Batch of (prompt, response, score) tuples that pass the threshold."""
|
||||
"""Response from the synthetic data generation. Batch of (prompt, response, score) tuples that pass the threshold."""
|
||||
|
||||
synthetic_data: List[Tuple[str, str, float]]
|
||||
|
||||
"""The actual synthetic data"""
|
||||
statistics: Dict[str, float]
|
||||
"""Statistics on how many prompts were generated and how many were filtered out"""
|
||||
synthetic_data: List[KScoredPromptGenerations]
|
||||
statistics: Optional[Dict[str, Any]] = None
|
||||
|
||||
|
||||
class SyntheticDataGeneration(Protocol):
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue