eleuther generate until

This commit is contained in:
Xi Yan 2024-10-08 23:57:22 -07:00
parent 6abef716dd
commit 9c38d9ae13
3 changed files with 34 additions and 40 deletions

View file

@ -91,10 +91,9 @@ class MetaReferenceInferenceImpl(Inference):
else: else:
return self._nonstream_chat_completion(request) return self._nonstream_chat_completion(request)
async def _nonstream_chat_completion( def _nonstream_chat_completion(
self, request: ChatCompletionRequest self, request: ChatCompletionRequest
) -> ChatCompletionResponse: ) -> ChatCompletionResponse:
async with SEMAPHORE:
messages = chat_completion_request_to_messages(request) messages = chat_completion_request_to_messages(request)
tokens = [] tokens = []
@ -121,18 +120,14 @@ class MetaReferenceInferenceImpl(Inference):
logprobs.append( logprobs.append(
TokenLogProbs( TokenLogProbs(
logprobs_by_token={ logprobs_by_token={token_result.text: token_result.logprobs[0]}
token_result.text: token_result.logprobs[0]
}
) )
) )
if stop_reason is None: if stop_reason is None:
stop_reason = StopReason.out_of_tokens stop_reason = StopReason.out_of_tokens
message = self.generator.formatter.decode_assistant_message( message = self.generator.formatter.decode_assistant_message(tokens, stop_reason)
tokens, stop_reason
)
return ChatCompletionResponse( return ChatCompletionResponse(
completion_message=message, completion_message=message,
logprobs=logprobs if request.logprobs else None, logprobs=logprobs if request.logprobs else None,

View file

@ -102,9 +102,8 @@ class EleutherEvalsWrapper(LM):
stream=False, stream=False,
) )
print(response) print(response)
res.append(response.completion_message) res.append(response.completion_message.content)
print(response)
return res return res

View file

@ -14,8 +14,8 @@ apis:
- evals - evals
providers: providers:
evals: evals:
- provider_id: meta-reference - provider_id: eleuther
provider_type: meta-reference provider_type: eleuther
config: {} config: {}
inference: inference:
- provider_id: meta-reference - provider_id: meta-reference