fix: bump snaps for mllama (#3202)

This commit is contained in:
drbh 2025-05-01 10:20:45 -04:00 committed by GitHub
parent 12ea8d74c7
commit d303c1e37e
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
3 changed files with 8 additions and 8 deletions

View File

@ -14,11 +14,11 @@
"usage": null
}
],
"created": 1739290197,
"created": 1746054921,
"id": "",
"model": "meta-llama/Llama-3.2-11B-Vision-Instruct",
"object": "chat.completion",
"system_fingerprint": "3.1.1-dev0-native",
"system_fingerprint": "3.2.3-dev0-native",
"usage": {
"completion_tokens": 10,
"prompt_tokens": 45,
@ -40,11 +40,11 @@
"usage": null
}
],
"created": 1739290197,
"created": 1746054921,
"id": "",
"model": "meta-llama/Llama-3.2-11B-Vision-Instruct",
"object": "chat.completion",
"system_fingerprint": "3.1.1-dev0-native",
"system_fingerprint": "3.2.3-dev0-native",
"usage": {
"completion_tokens": 10,
"prompt_tokens": 45,

View File

@ -5,7 +5,7 @@
"index": 0,
"logprobs": null,
"message": {
"content": "A chicken sits on a pile of money, looking",
"content": "A chicken stands on a pile of money, looking",
"name": null,
"role": "assistant",
"tool_calls": null
@ -13,11 +13,11 @@
"usage": null
}
],
"created": 1739290152,
"created": 1746054919,
"id": "",
"model": "meta-llama/Llama-3.2-11B-Vision-Instruct",
"object": "chat.completion",
"system_fingerprint": "3.1.1-dev0-native",
"system_fingerprint": "3.2.3-dev0-native",
"usage": {
"completion_tokens": 10,
"prompt_tokens": 45,

View File

@ -48,7 +48,7 @@ async def test_mllama_simpl(mllama, response_snapshot):
}
assert (
response.choices[0].message.content
== "A chicken sits on a pile of money, looking"
== "A chicken stands on a pile of money, looking"
)
assert response == response_snapshot