mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-14 22:33:48 +00:00
IDs are now deterministic hashes based on request content, and timestamps are normalized to constants, eliminating spurious changes when re-recording tests. ## Changes - Updated `inference_recorder.py` to normalize IDs and timestamps during recording - Added `scripts/normalize_recordings.py` utility to re-normalize existing recordings - Created documentation in `tests/integration/recordings/README.md` - Normalized 350 existing recording files
56 lines
4.2 KiB
JSON
56 lines
4.2 KiB
JSON
{
|
|
"request": {
|
|
"method": "POST",
|
|
"url": "http://0.0.0.0:11434/v1/v1/chat/completions",
|
|
"headers": {},
|
|
"body": {
|
|
"model": "llama3.2:3b-instruct-fp16",
|
|
"messages": [
|
|
{
|
|
"role": "user",
|
|
"content": "Test trace openai 2"
|
|
}
|
|
],
|
|
"stream": false
|
|
},
|
|
"endpoint": "/v1/chat/completions",
|
|
"model": "llama3.2:3b-instruct-fp16"
|
|
},
|
|
"response": {
|
|
"body": {
|
|
"__type__": "openai.types.chat.chat_completion.ChatCompletion",
|
|
"__data__": {
|
|
"id": "rec-8295382a8e7c",
|
|
"choices": [
|
|
{
|
|
"finish_reason": "stop",
|
|
"index": 0,
|
|
"logprobs": null,
|
|
"message": {
|
|
"content": "I'd be happy to help you test the Transformer-XL (denoted as \"Test Trace OpenAI 2\") model, but first I need to clarify a few things:\n\n1. **Transformer-XL** is not an official name from OpenAI. It's actually a variant of the Transformer model proposed in the paper \"Long-Short Term Memory Are General: A Study on The Curvature of Time\" (2017) by Jinyu Chen, et al.\n2. **Trace OpenAI 2** sounds like a specific version or configuration of the Transformer-XL model, possibly developed by OpenAI.\n\nGiven these clarifications, I'll provide you with a general idea of how to test the Transformer-XL (or \"Test Trace OpenAI 2\") model using popular frameworks and libraries. Please note that this is not an exhaustive or definitive guide.\n\nTo test the Transformer-XL model, you can follow these steps:\n\n1. **Install the necessary dependencies**: You'll need a deep learning framework like TensorFlow or PyTorch, as well as a library for natural language processing (NLP) like Hugging Face's transformers.\n2. **Load the pre-trained weights**: You can use a pre-trained model checkpoint from Hugging Face's Transformers library or load your own weights trained on a specific task or dataset.\n3. **Prepare your data**: Load your text data into tokens, such as words or characters, and preprocess it according to the specific requirements of the Transformer-XL architecture (e.g., tokenization, padding, etc.).\n4. **Configure the model**: Adjust the hyperparameters to suit your specific test case, including the model's configuration, batch size, learning rate, etc.\n5. **Run the inference**: Use the loaded pre-trained weights to perform inference on your test data.\n\nHere's some sample Python code using PyTorch and Hugging Face's Transformers library to get you started:\n```python\nimport torch\nfrom transformers import LongformerForSequenceClassification, LongformerTokenizer\n\n# Load pre-trained weights\nmodel = LongformerForSequenceClassification.from_pretrained('test-trace-openai-2')\n\n# Prepare data\ntokenizer = model.tokenizer\ntext = \"This is a test sentence\"\ninputs = tokenizer(text, return_tensors='pt')\noutput = model(inputs['input_ids'], attention_mask=inputs['attention_mask'])\n\n# Print the results\nprint(output.logits)\n```\nPlease note that this code snippet is just an example and may not work as-is. You'll need to adapt it to your specific requirements and test data.\n\nKeep in mind that testing a model's performance on a specific task or dataset requires careful consideration of factors like:\n\n* **Test data quality**: Your test data should accurately represent the underlying distribution of your target dataset.\n* **Model evaluation metrics**: Choose relevant evaluation metrics that measure the model's performance on your specific task, such as accuracy, precision, recall, F1-score, etc.\n\nFeel free to ask if you have any further questions or need more guidance!",
|
|
"refusal": null,
|
|
"role": "assistant",
|
|
"annotations": null,
|
|
"audio": null,
|
|
"function_call": null,
|
|
"tool_calls": null
|
|
}
|
|
}
|
|
],
|
|
"created": 0,
|
|
"model": "llama3.2:3b-instruct-fp16",
|
|
"object": "chat.completion",
|
|
"service_tier": null,
|
|
"system_fingerprint": "fp_ollama",
|
|
"usage": {
|
|
"completion_tokens": 614,
|
|
"prompt_tokens": 31,
|
|
"total_tokens": 645,
|
|
"completion_tokens_details": null,
|
|
"prompt_tokens_details": null
|
|
}
|
|
}
|
|
},
|
|
"is_streaming": false
|
|
}
|
|
}
|