ArthurZ HF Staff commited on
Commit
d92fa0d
·
verified ·
1 Parent(s): 6a9aa4d

Update CircleCI artifacts for PR 41621 (ba4a4a05a349)

Browse files
pr-41621/sha-ba4a4a05a349/failure_summary.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "failures": [
3
+ {
4
+ "job_name": "tests_tokenization",
5
+ "test_name": "tests/test_tokenization_mistral_common.py::TestMistralCommonBackend::test_all_special_ids",
6
+ "short_error": "AttributeError: LlamaTokenizer(name_or_path='hf-internal-testing/namespace-mistralai-repo_name-Mistral-Small-3.1-24B-Instruct-2503', vocab_size=3, model_max_length=1000000000000000019884624838656, padding_side='left', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>'}, added_tokens_decoder={",
7
+ "error": "/usr/local/lib/python3.10/unittest/mock.py:1420: AttributeError: LlamaTokenizer(name_or_path='hf-internal-testing/namespace-mistralai-repo_name-Mistral-Small-3.1-24B-Instruct-2503', vocab_size=3, model_max_length=1000000000000000019884624838656, padding_side='left', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>'}, added_tokens_decoder={",
8
+ "model_name": null
9
+ },
10
+ {
11
+ "job_name": "tests_tokenization",
12
+ "test_name": "tests/test_tokenization_mistral_common.py::TestMistralCommonBackend::test_all_special_tokens",
13
+ "short_error": "AttributeError: LlamaTokenizer(name_or_path='hf-internal-testing/namespace-mistralai-repo_name-Mistral-Small-3.1-24B-Instruct-2503', vocab_size=3, model_max_length=1000000000000000019884624838656, padding_side='left', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>'}, added_tokens_decoder={",
14
+ "error": "/usr/local/lib/python3.10/unittest/mock.py:1420: AttributeError: LlamaTokenizer(name_or_path='hf-internal-testing/namespace-mistralai-repo_name-Mistral-Small-3.1-24B-Instruct-2503', vocab_size=3, model_max_length=1000000000000000019884624838656, padding_side='left', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>'}, added_tokens_decoder={",
15
+ "model_name": null
16
+ }
17
+ ],
18
+ "by_test": {
19
+ "tests/test_tokenization_mistral_common.py::TestMistralCommonBackend::test_all_special_ids": {
20
+ "count": 1,
21
+ "errors": {
22
+ "/usr/local/lib/python3.10/unittest/mock.py:1420: AttributeError: LlamaTokenizer(name_or_path='hf-internal-testing/namespace-mistralai-repo_name-Mistral-Small-3.1-24B-Instruct-2503', vocab_size=3, model_max_length=1000000000000000019884624838656, padding_side='left', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>'}, added_tokens_decoder={": 1
23
+ },
24
+ "jobs": [
25
+ "tests_tokenization"
26
+ ],
27
+ "variants": [
28
+ "tests/test_tokenization_mistral_common.py::TestMistralCommonBackend::test_all_special_ids"
29
+ ]
30
+ },
31
+ "tests/test_tokenization_mistral_common.py::TestMistralCommonBackend::test_all_special_tokens": {
32
+ "count": 1,
33
+ "errors": {
34
+ "/usr/local/lib/python3.10/unittest/mock.py:1420: AttributeError: LlamaTokenizer(name_or_path='hf-internal-testing/namespace-mistralai-repo_name-Mistral-Small-3.1-24B-Instruct-2503', vocab_size=3, model_max_length=1000000000000000019884624838656, padding_side='left', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>'}, added_tokens_decoder={": 1
35
+ },
36
+ "jobs": [
37
+ "tests_tokenization"
38
+ ],
39
+ "variants": [
40
+ "tests/test_tokenization_mistral_common.py::TestMistralCommonBackend::test_all_special_tokens"
41
+ ]
42
+ }
43
+ },
44
+ "by_model": {}
45
+ }