mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-03 19:57:35 +00:00
35 lines
1.1 KiB
Python
35 lines
1.1 KiB
Python
from typing import Any, Dict, List, Optional, Protocol
|
|
|
|
from pydantic import BaseModel
|
|
|
|
from pyopenapi import webmethod
|
|
from strong_typing.schema import json_schema_type
|
|
|
|
from models.llama3_1.api.datatypes import * # noqa: F403
|
|
from toolchain.reward_scoring.api.datatypes import * # noqa: F403
|
|
from .datatypes import * # noqa: F403
|
|
|
|
|
|
@json_schema_type
|
|
class SyntheticDataGenerationRequest(BaseModel):
|
|
"""Request to generate synthetic data. A small batch of prompts and a filtering function"""
|
|
|
|
dialogs: List[Message]
|
|
filtering_function: FilteringFunction = FilteringFunction.none
|
|
model: Optional[RewardModel] = None
|
|
|
|
|
|
@json_schema_type
|
|
class SyntheticDataGenerationResponse(BaseModel):
|
|
"""Response from the synthetic data generation. Batch of (prompt, response, score) tuples that pass the threshold."""
|
|
|
|
synthetic_data: List[ScoredDialogGenerations]
|
|
statistics: Optional[Dict[str, Any]] = None
|
|
|
|
|
|
class SyntheticDataGeneration(Protocol):
|
|
@webmethod(route="/synthetic_data_generation/generate")
|
|
def post_generate(
|
|
self,
|
|
request: SyntheticDataGenerationRequest,
|
|
) -> Union[SyntheticDataGenerationResponse]: ...
|