feihu.hf commited on
Commit
1d5ada7
·
1 Parent(s): 44823bd

update README

Browse files
Files changed (4) hide show
  1. LICENSE +0 -0
  2. README.md +5 -5
  3. generation_config.json +12 -5
  4. tokenizer_config.json +0 -1
LICENSE CHANGED
File without changes
README.md CHANGED
@@ -322,10 +322,10 @@ YaRN is currently supported by several inference frameworks, e.g., `transformers
322
 
323
  | Mode | QUANTIZATION TYPE | LiveBench 2024-11-25 | GPQA | MMLU-Redux | AIME24 |
324
  | --- | --- | --- | --- | --- | --- |
325
- | Thinking | bf16 | 74.3 | 65.8 | 89.5 | 80.4 |
326
- | Thinking | GPTQ-int4 | 71.5 | 60.1 | 88.8 | 79.5 |
327
- | Non-Thinking | bf16 | 59.4 | 54.8 | 84.1 | - |
328
- | Non-Thinking | GPTQ-int4 | 57.2 | 50.4 | 83.5 | - |
329
 
330
  ## Best Practices
331
 
@@ -356,4 +356,4 @@ If you find our work helpful, feel free to give us a cite.
356
  month = {April},
357
  year = {2025}
358
  }
359
- ```
 
322
 
323
  | Mode | QUANTIZATION TYPE | LiveBench 2024-11-25 | GPQA | MMLU-Redux | AIME24 |
324
  | --- | --- | --- | --- | --- | --- |
325
+ | Thinking | bf16 | 77.1 | 71.1 | 92.7 | - |
326
+ | Thinking | GPTQ-int4 | 75.1 | 71.9 | 92.0 | - |
327
+ | Non-Thinking | bf16 | 62.5 | 62.9 | 89.2 | - |
328
+ | Non-Thinking | GPTQ-int4 | 61.1 | 62.8 | 89.0 | - |
329
 
330
  ## Best Practices
331
 
 
356
  month = {April},
357
  year = {2025}
358
  }
359
+ ```
generation_config.json CHANGED
@@ -1,6 +1,13 @@
1
  {
2
- "_from_model_config": true,
3
- "eos_token_id": 151645,
4
- "pad_token_id": 151643,
5
- "transformers_version": "4.51.3"
6
- }
 
 
 
 
 
 
 
 
1
  {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "temperature": 0.6,
10
+ "top_k": 20,
11
+ "top_p": 0.95,
12
+ "transformers_version": "4.51.0"
13
+ }
tokenizer_config.json CHANGED
@@ -231,7 +231,6 @@
231
  "clean_up_tokenization_spaces": false,
232
  "eos_token": "<|im_end|>",
233
  "errors": "replace",
234
- "extra_special_tokens": {},
235
  "model_max_length": 131072,
236
  "pad_token": "<|endoftext|>",
237
  "split_special_tokens": false,
 
231
  "clean_up_tokenization_spaces": false,
232
  "eos_token": "<|im_end|>",
233
  "errors": "replace",
 
234
  "model_max_length": 131072,
235
  "pad_token": "<|endoftext|>",
236
  "split_special_tokens": false,