zerofata commited on
Commit
cfc3cf9
·
verified ·
1 Parent(s): d491170

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. args.json +467 -0
  3. chat_template.jinja +103 -0
  4. config.json +43 -0
  5. generation_config.json +10 -0
  6. model-00001-of-00043.safetensors +3 -0
  7. model-00002-of-00043.safetensors +3 -0
  8. model-00003-of-00043.safetensors +3 -0
  9. model-00004-of-00043.safetensors +3 -0
  10. model-00005-of-00043.safetensors +3 -0
  11. model-00006-of-00043.safetensors +3 -0
  12. model-00007-of-00043.safetensors +3 -0
  13. model-00008-of-00043.safetensors +3 -0
  14. model-00009-of-00043.safetensors +3 -0
  15. model-00010-of-00043.safetensors +3 -0
  16. model-00011-of-00043.safetensors +3 -0
  17. model-00012-of-00043.safetensors +3 -0
  18. model-00013-of-00043.safetensors +3 -0
  19. model-00014-of-00043.safetensors +3 -0
  20. model-00015-of-00043.safetensors +3 -0
  21. model-00016-of-00043.safetensors +3 -0
  22. model-00017-of-00043.safetensors +3 -0
  23. model-00018-of-00043.safetensors +3 -0
  24. model-00019-of-00043.safetensors +3 -0
  25. model-00020-of-00043.safetensors +3 -0
  26. model-00021-of-00043.safetensors +3 -0
  27. model-00022-of-00043.safetensors +3 -0
  28. model-00023-of-00043.safetensors +3 -0
  29. model-00024-of-00043.safetensors +3 -0
  30. model-00025-of-00043.safetensors +3 -0
  31. model-00026-of-00043.safetensors +3 -0
  32. model-00027-of-00043.safetensors +3 -0
  33. model-00028-of-00043.safetensors +3 -0
  34. model-00029-of-00043.safetensors +3 -0
  35. model-00030-of-00043.safetensors +3 -0
  36. model-00031-of-00043.safetensors +3 -0
  37. model-00032-of-00043.safetensors +3 -0
  38. model-00033-of-00043.safetensors +3 -0
  39. model-00034-of-00043.safetensors +3 -0
  40. model-00035-of-00043.safetensors +3 -0
  41. model-00036-of-00043.safetensors +3 -0
  42. model-00037-of-00043.safetensors +3 -0
  43. model-00038-of-00043.safetensors +3 -0
  44. model-00039-of-00043.safetensors +3 -0
  45. model-00040-of-00043.safetensors +3 -0
  46. model-00041-of-00043.safetensors +3 -0
  47. model-00042-of-00043.safetensors +3 -0
  48. model-00043-of-00043.safetensors +3 -0
  49. model.safetensors.index.json +0 -0
  50. special_tokens_map.json +40 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
args.json ADDED
@@ -0,0 +1,467 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "use_ray": false,
3
+ "ray_exp_name": null,
4
+ "device_groups": null,
5
+ "model": "zai-org/GLM-4.5-Air",
6
+ "model_type": "glm4_5",
7
+ "model_revision": null,
8
+ "task_type": "causal_lm",
9
+ "torch_dtype": "bfloat16",
10
+ "attn_impl": null,
11
+ "new_special_tokens": [],
12
+ "num_labels": null,
13
+ "problem_type": null,
14
+ "rope_scaling": null,
15
+ "device_map": null,
16
+ "max_memory": {},
17
+ "max_model_len": null,
18
+ "local_repo_path": null,
19
+ "init_strategy": null,
20
+ "template": "glm4_5",
21
+ "system": null,
22
+ "max_length": 10280,
23
+ "truncation_strategy": "delete",
24
+ "max_pixels": null,
25
+ "agent_template": null,
26
+ "norm_bbox": null,
27
+ "use_chat_template": true,
28
+ "padding_free": true,
29
+ "padding_side": "right",
30
+ "loss_scale": "default",
31
+ "sequence_parallel_size": 1,
32
+ "response_prefix": null,
33
+ "template_backend": "swift",
34
+ "dataset": [
35
+ "/workspace/joined_dataset_cleaned_modified.jsonl"
36
+ ],
37
+ "val_dataset": [],
38
+ "split_dataset_ratio": 0.01,
39
+ "data_seed": 42,
40
+ "dataset_num_proc": 8,
41
+ "load_from_cache_file": true,
42
+ "dataset_shuffle": true,
43
+ "val_dataset_shuffle": false,
44
+ "streaming": false,
45
+ "interleave_prob": null,
46
+ "stopping_strategy": "first_exhausted",
47
+ "shuffle_buffer_size": 1000,
48
+ "download_mode": "reuse_dataset_if_exists",
49
+ "columns": {},
50
+ "strict": false,
51
+ "remove_unused_columns": true,
52
+ "model_name": null,
53
+ "model_author": null,
54
+ "custom_dataset_info": [],
55
+ "quant_method": null,
56
+ "quant_bits": null,
57
+ "hqq_axis": null,
58
+ "bnb_4bit_compute_dtype": "bfloat16",
59
+ "bnb_4bit_quant_type": "nf4",
60
+ "bnb_4bit_use_double_quant": true,
61
+ "bnb_4bit_quant_storage": null,
62
+ "max_new_tokens": null,
63
+ "temperature": null,
64
+ "top_k": null,
65
+ "top_p": null,
66
+ "repetition_penalty": null,
67
+ "num_beams": 1,
68
+ "stream": false,
69
+ "stop_words": [],
70
+ "logprobs": false,
71
+ "top_logprobs": null,
72
+ "ckpt_dir": "/workspace/glm-4.5-air-mcore",
73
+ "lora_modules": [],
74
+ "tuner_backend": "peft",
75
+ "train_type": "lora",
76
+ "adapters": [],
77
+ "external_plugins": [],
78
+ "seed": 42,
79
+ "model_kwargs": {},
80
+ "load_args": false,
81
+ "load_data_args": false,
82
+ "packing": true,
83
+ "packing_length": 10280,
84
+ "lazy_tokenize": false,
85
+ "cached_dataset": [],
86
+ "custom_register_path": [],
87
+ "use_hf": false,
88
+ "hub_token": null,
89
+ "ddp_timeout": 18000000,
90
+ "ddp_backend": null,
91
+ "ignore_args_error": false,
92
+ "use_swift_lora": false,
93
+ "freeze_llm": false,
94
+ "freeze_vit": true,
95
+ "freeze_aligner": true,
96
+ "freeze_parameters": [],
97
+ "freeze_parameters_regex": null,
98
+ "freeze_parameters_ratio": 0.0,
99
+ "trainable_parameters": [],
100
+ "trainable_parameters_regex": null,
101
+ "adapter_load": null,
102
+ "target_modules": [
103
+ "all-linear"
104
+ ],
105
+ "target_regex": null,
106
+ "modules_to_save": [],
107
+ "lora_rank": 256,
108
+ "lora_alpha": 16,
109
+ "lora_dropout": 0.05,
110
+ "lora_bias": "none",
111
+ "lora_dtype": null,
112
+ "use_rslora": true,
113
+ "rlhf_type": null,
114
+ "ref_load": null,
115
+ "ref_adapter_load": null,
116
+ "beta": 0.1,
117
+ "rpo_alpha": null,
118
+ "reference_free": false,
119
+ "label_smoothing": 0.0,
120
+ "f_divergence_type": "reverse_kl",
121
+ "loss_type": null,
122
+ "desirable_weight": 1.0,
123
+ "undesirable_weight": 1.0,
124
+ "calculate_KL": null,
125
+ "center_rewards_coefficient": null,
126
+ "padded_vocab_size": 151552,
127
+ "initialize_embedding": false,
128
+ "mlp_padding_free": false,
129
+ "dataloader_persistent_workers": true,
130
+ "dataloader_prefetch_factor": 10,
131
+ "architectures": "Glm4MoeForCausalLM",
132
+ "llm_architectures": null,
133
+ "max_epochs": 2,
134
+ "enable_dft_loss": false,
135
+ "enable_channel_loss": false,
136
+ "original_max_position_embeddings": null,
137
+ "partial_rotary_factor": 0.5,
138
+ "use_shared_expert_gate": false,
139
+ "vit_gradient_checkpointing": true,
140
+ "gradient_checkpointing_kwargs": null,
141
+ "linear_num_value_heads": null,
142
+ "linear_num_key_heads": null,
143
+ "linear_key_head_dim": null,
144
+ "linear_value_head_dim": null,
145
+ "linear_conv_kernel_dim": null,
146
+ "layer_types": null,
147
+ "mrope_interleaved": false,
148
+ "micro_batch_size": 4,
149
+ "global_batch_size": 32,
150
+ "recompute_granularity": "full",
151
+ "recompute_method": "uniform",
152
+ "recompute_num_layers": 1,
153
+ "recompute_modules": [
154
+ "core_attn"
155
+ ],
156
+ "use_cpu_initialization": false,
157
+ "deterministic_mode": false,
158
+ "train_iters": null,
159
+ "log_interval": 5,
160
+ "tensorboard_dir": "/workspace/megatron_output/Iceblink-v3-SFT-3/v0-20251101-214719/runs",
161
+ "no_masked_softmax_fusion": false,
162
+ "no_bias_dropout_fusion": false,
163
+ "no_bias_swiglu_fusion": false,
164
+ "no_rope_fusion": false,
165
+ "no_gradient_accumulation_fusion": false,
166
+ "cross_entropy_loss_fusion": true,
167
+ "cross_entropy_fusion_impl": "native",
168
+ "calculate_per_token_loss": true,
169
+ "use_flash_attn": false,
170
+ "attention_backend": "flash",
171
+ "optimizer": "adam",
172
+ "optimizer_cpu_offload": false,
173
+ "optimizer_offload_fraction": 1.0,
174
+ "use_precision_aware_optimizer": false,
175
+ "main_grads_dtype": "fp32",
176
+ "main_params_dtype": "fp32",
177
+ "exp_avg_dtype": "fp32",
178
+ "exp_avg_sq_dtype": "fp32",
179
+ "dataloader_type": "cyclic",
180
+ "manual_gc": false,
181
+ "manual_gc_interval": 0,
182
+ "lr": 6e-06,
183
+ "lr_decay_style": "cosine",
184
+ "lr_decay_iters": null,
185
+ "lr_warmup_iters": 0,
186
+ "lr_warmup_fraction": 0.05,
187
+ "min_lr": 6e-07,
188
+ "weight_decay": 0.1,
189
+ "clip_grad": 1.0,
190
+ "adam_beta1": 0.9,
191
+ "adam_beta2": 0.95,
192
+ "adam_eps": 1e-08,
193
+ "sgd_momentum": 0.9,
194
+ "save": "/workspace/megatron_output/Iceblink-v3-SFT-3/v0-20251101-214719",
195
+ "save_interval": 25,
196
+ "save_retain_interval": null,
197
+ "no_save_optim": true,
198
+ "no_save_rng": true,
199
+ "load": "/workspace/glm-4.5-air-mcore",
200
+ "no_load_optim": false,
201
+ "no_load_rng": false,
202
+ "finetune": true,
203
+ "ckpt_format": "torch_dist",
204
+ "no_initialization": true,
205
+ "auto_detect_ckpt_format": true,
206
+ "exit_on_missing_checkpoint": true,
207
+ "async_save": false,
208
+ "use_persistent_ckpt_worker": false,
209
+ "ckpt_fully_parallel_load": false,
210
+ "ckpt_assume_constant_structure": false,
211
+ "distributed_backend": "nccl",
212
+ "local_rank": 0,
213
+ "use_distributed_optimizer": true,
214
+ "tensor_model_parallel_size": 8,
215
+ "pipeline_model_parallel_size": 1,
216
+ "decoder_first_pipeline_num_layers": null,
217
+ "decoder_last_pipeline_num_layers": null,
218
+ "sequence_parallel": true,
219
+ "context_parallel_size": 1,
220
+ "tp_comm_overlap": false,
221
+ "overlap_grad_reduce": false,
222
+ "overlap_param_gather": false,
223
+ "distributed_timeout_minutes": 300000,
224
+ "num_layers_per_virtual_pipeline_stage": null,
225
+ "num_virtual_stages_per_pipeline_rank": null,
226
+ "microbatch_group_size_per_virtual_pipeline_stage": null,
227
+ "pipeline_model_parallel_layout": null,
228
+ "num_layers": 46,
229
+ "hidden_size": 4096,
230
+ "ffn_hidden_size": 10944,
231
+ "num_attention_heads": 96,
232
+ "group_query_attention": true,
233
+ "num_query_groups": 8,
234
+ "max_position_embeddings": 131072,
235
+ "position_embedding_type": "rope",
236
+ "mrope_section": null,
237
+ "rotary_base": 1000000,
238
+ "rotary_percent": 1.0,
239
+ "rotary_interleaved": false,
240
+ "normalization": "RMSNorm",
241
+ "norm_epsilon": 1e-05,
242
+ "swiglu": true,
243
+ "untie_embeddings_and_output_weights": true,
244
+ "disable_bias_linear": true,
245
+ "add_qkv_bias": true,
246
+ "attention_dropout": 0.0,
247
+ "hidden_dropout": 0.0,
248
+ "kv_channels": 128,
249
+ "qk_layernorm": false,
250
+ "transformer_impl": "transformer_engine",
251
+ "num_experts": 128,
252
+ "moe_layer_freq": "[0]*1+[1]*45",
253
+ "moe_ffn_hidden_size": 1408,
254
+ "moe_shared_expert_intermediate_size": 1408,
255
+ "moe_router_topk": 8,
256
+ "moe_router_pre_softmax": false,
257
+ "moe_router_dtype": "fp32",
258
+ "moe_router_score_function": "sigmoid",
259
+ "moe_router_bias_update_rate": 0.001,
260
+ "moe_router_enable_expert_bias": true,
261
+ "moe_router_topk_scaling_factor": 1.0,
262
+ "moe_router_load_balancing_type": "aux_loss",
263
+ "expert_model_parallel_size": 8,
264
+ "expert_tensor_parallel_size": 1,
265
+ "moe_token_dispatcher_type": null,
266
+ "moe_enable_deepep": false,
267
+ "moe_grouped_gemm": true,
268
+ "moe_permute_fusion": true,
269
+ "moe_aux_loss_coeff": 6e-05,
270
+ "moe_z_loss_coeff": null,
271
+ "moe_shared_expert_overlap": true,
272
+ "moe_layer_recompute": false,
273
+ "moe_expert_capacity_factor": null,
274
+ "moe_pad_expert_input_to_capacity": false,
275
+ "moe_token_drop_policy": null,
276
+ "multi_latent_attention": false,
277
+ "q_lora_rank": null,
278
+ "kv_lora_rank": 32,
279
+ "qk_head_dim": 128,
280
+ "qk_pos_emb_head_dim": 64,
281
+ "fp8_format": null,
282
+ "fp8_recipe": "delayed",
283
+ "fp8_amax_history_len": 1024,
284
+ "fp8_amax_compute_algo": "max",
285
+ "fp8_param_gather": false,
286
+ "fp16": false,
287
+ "bf16": true,
288
+ "apply_query_key_layer_scaling": false,
289
+ "attention_softmax_in_fp32": true,
290
+ "log_params_norm": false,
291
+ "log_throughput": false,
292
+ "tensorboard_log_interval": 1,
293
+ "tensorboard_queue_size": 50,
294
+ "log_timers_to_tensorboard": true,
295
+ "no_log_learning_rate_to_tensorboard": false,
296
+ "log_validation_ppl_to_tensorboard": true,
297
+ "log_memory_to_tensorboard": true,
298
+ "logging_level": null,
299
+ "wandb_project": "Megatron-Air-SFT",
300
+ "wandb_exp_name": "Iceblink-v3-SFT-3",
301
+ "wandb_save_dir": null,
302
+ "eval_iters": -1,
303
+ "eval_interval": 20,
304
+ "seq_length": 10280,
305
+ "num_workers": 8,
306
+ "megatron_extra_kwargs": {},
307
+ "add_version": true,
308
+ "rank": 0,
309
+ "global_world_size": 8,
310
+ "local_world_size": 8,
311
+ "model_suffix": "GLM-4.5-Air",
312
+ "model_info": "ModelInfo(model_type='glm4_5', model_dir='/root/.cache/modelscope/hub/models/ZhipuAI/GLM-4___5-Air', torch_dtype=torch.bfloat16, max_model_len=131072, quant_method=None, quant_bits=None, rope_scaling=None, is_moe_model=True, config=None, task_type='causal_lm', num_labels=None)",
313
+ "model_meta": "ModelMeta(model_type='glm4_5', model_groups=[ModelGroup(models=[Model(ms_model_id='ZhipuAI/GLM-4.5-Air-Base', hf_model_id='zai-org/GLM-4.5-Air-Base', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='ZhipuAI/GLM-4.5-Air', hf_model_id='zai-org/GLM-4.5-Air', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='ZhipuAI/GLM-4.5-Air-FP8', hf_model_id='zai-org/GLM-4.5-Air-FP8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='ZhipuAI/GLM-4.5-Base', hf_model_id='zai-org/GLM-4.5-Base', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='ZhipuAI/GLM-4.5', hf_model_id='zai-org/GLM-4.5', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='ZhipuAI/GLM-4.5-FP8', hf_model_id='zai-org/GLM-4.5-FP8', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[]), ModelGroup(models=[Model(ms_model_id='ZhipuAI/GLM-4.6', hf_model_id='zai-org/GLM-4.6', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[])], template='glm4_5', get_function=<function get_model_tokenizer_with_flash_attn at 0x76c9ab052520>, model_arch=None, architectures=['Glm4MoeForCausalLM'], additional_saved_files=[], torch_dtype=None, is_multimodal=False, is_reward=False, is_reranker=False, task_type=None, ignore_patterns=None, requires=['transformers>=4.54'], tags=[])",
314
+ "model_dir": "/root/.cache/modelscope/hub/models/ZhipuAI/GLM-4___5-Air",
315
+ "hub": "<class 'swift.hub.hub.MSHub'>",
316
+ "megatron_model_meta": "MegatronModelMeta(megatron_model_type='gpt', model_types=['qwen2', 'qwen2_5', 'qwq', 'qwq_preview', 'qwen2_5_math', 'llama', 'llama3', 'llama3_1', 'llama3_2', 'longwriter_llama3_1', 'codefuse_codellama', 'marco_o1', 'deepseek', 'deepseek_r1_distill', 'yi', 'yi_coder', 'sus', 'skywork_o1', 'openbuddy_llama', 'openbuddy_llama3', 'megrez', 'reflection', 'numina', 'ziya', 'mengzi3', 'qwen3', 'qwen3_thinking', 'qwen3_nothinking', 'qwen2_moe', 'qwen3_moe', 'qwen3_moe_thinking', 'qwen3_coder', 'internlm3', 'mimo', 'mimo_rl', 'moonlight', 'deepseek_moe', 'deepseek_v2', 'deepseek_v2_5', 'deepseek_r1', 'dots1', 'ernie', 'glm4_5', 'deepseek_v3_1', 'ernie_thinking'], convert_mcore2hf=<function convert_mcore2hf at 0x76c91dcb0fe0>, convert_hf2mcore=<function convert_hf2mcore at 0x76c91dcb0b80>, model_cls=<class 'swift.megatron.model.gpt_model.GPTModel'>, convert_hf_config=<function convert_gpt_hf_config at 0x76c91de6c9a0>, get_transformer_layer_spec=None, model_provider=<function model_provider at 0x76c91dddb880>, visual_cls=None, extra_args_provider=None)",
317
+ "extra_args": {
318
+ "use_ray": false,
319
+ "ray_exp_name": null,
320
+ "device_groups": null,
321
+ "model": "ZhipuAI/GLM-4.5-Air",
322
+ "model_type": "glm4_5",
323
+ "model_revision": null,
324
+ "task_type": "causal_lm",
325
+ "torch_dtype": "bfloat16",
326
+ "attn_impl": null,
327
+ "new_special_tokens": [],
328
+ "num_labels": null,
329
+ "problem_type": null,
330
+ "rope_scaling": null,
331
+ "device_map": null,
332
+ "max_memory": {},
333
+ "max_model_len": null,
334
+ "local_repo_path": null,
335
+ "init_strategy": null,
336
+ "template": "glm4_5",
337
+ "system": null,
338
+ "max_length": 10280,
339
+ "truncation_strategy": "delete",
340
+ "max_pixels": null,
341
+ "agent_template": null,
342
+ "norm_bbox": null,
343
+ "use_chat_template": true,
344
+ "padding_free": true,
345
+ "padding_side": "right",
346
+ "sequence_parallel_size": 1,
347
+ "response_prefix": null,
348
+ "template_backend": "swift",
349
+ "dataset": [
350
+ "/workspace/joined_dataset_cleaned_modified.jsonl"
351
+ ],
352
+ "val_dataset": [],
353
+ "split_dataset_ratio": 0.01,
354
+ "data_seed": 42,
355
+ "dataset_num_proc": 8,
356
+ "load_from_cache_file": true,
357
+ "dataset_shuffle": true,
358
+ "val_dataset_shuffle": false,
359
+ "streaming": false,
360
+ "interleave_prob": null,
361
+ "stopping_strategy": "first_exhausted",
362
+ "shuffle_buffer_size": 1000,
363
+ "download_mode": "reuse_dataset_if_exists",
364
+ "columns": {},
365
+ "strict": false,
366
+ "remove_unused_columns": true,
367
+ "model_name": null,
368
+ "model_author": null,
369
+ "custom_dataset_info": [],
370
+ "quant_method": null,
371
+ "quant_bits": null,
372
+ "hqq_axis": null,
373
+ "bnb_4bit_compute_dtype": "bfloat16",
374
+ "bnb_4bit_quant_type": "nf4",
375
+ "bnb_4bit_use_double_quant": true,
376
+ "bnb_4bit_quant_storage": null,
377
+ "max_new_tokens": null,
378
+ "temperature": null,
379
+ "top_k": null,
380
+ "top_p": null,
381
+ "repetition_penalty": null,
382
+ "num_beams": 1,
383
+ "stream": false,
384
+ "stop_words": [],
385
+ "logprobs": false,
386
+ "top_logprobs": null,
387
+ "ckpt_dir": "/workspace/glm-4.5-air-mcore",
388
+ "lora_modules": [],
389
+ "tuner_backend": "peft",
390
+ "train_type": "lora",
391
+ "adapters": [],
392
+ "external_plugins": [],
393
+ "model_kwargs": {},
394
+ "load_args": false,
395
+ "load_data_args": false,
396
+ "packing": true,
397
+ "packing_length": 10280,
398
+ "lazy_tokenize": false,
399
+ "cached_dataset": [],
400
+ "custom_register_path": [],
401
+ "use_hf": false,
402
+ "hub_token": null,
403
+ "ddp_timeout": 18000000,
404
+ "ddp_backend": null,
405
+ "ignore_args_error": false,
406
+ "use_swift_lora": false,
407
+ "freeze_llm": false,
408
+ "freeze_vit": true,
409
+ "freeze_aligner": true,
410
+ "freeze_parameters": [],
411
+ "freeze_parameters_regex": null,
412
+ "freeze_parameters_ratio": 0.0,
413
+ "trainable_parameters": [],
414
+ "trainable_parameters_regex": null,
415
+ "adapter_load": null,
416
+ "target_modules": [
417
+ "all-linear"
418
+ ],
419
+ "target_regex": null,
420
+ "modules_to_save": [],
421
+ "lora_rank": 256,
422
+ "lora_alpha": 16,
423
+ "lora_dropout": 0.05,
424
+ "lora_bias": "none",
425
+ "lora_dtype": null,
426
+ "use_rslora": true,
427
+ "rlhf_type": null,
428
+ "ref_load": null,
429
+ "ref_adapter_load": null,
430
+ "beta": 0.1,
431
+ "rpo_alpha": null,
432
+ "reference_free": false,
433
+ "label_smoothing": 0.0,
434
+ "f_divergence_type": "reverse_kl",
435
+ "loss_type": null,
436
+ "desirable_weight": 1.0,
437
+ "undesirable_weight": 1.0,
438
+ "calculate_KL": null,
439
+ "center_rewards_coefficient": null,
440
+ "padded_vocab_size": 151552,
441
+ "initialize_embedding": false,
442
+ "mlp_padding_free": false,
443
+ "dataloader_persistent_workers": true,
444
+ "dataloader_prefetch_factor": 10,
445
+ "architectures": "Glm4MoeForCausalLM",
446
+ "llm_architectures": null,
447
+ "max_epochs": 2,
448
+ "enable_dft_loss": false,
449
+ "enable_channel_loss": false,
450
+ "original_max_position_embeddings": null,
451
+ "partial_rotary_factor": 0.5,
452
+ "use_shared_expert_gate": false,
453
+ "vit_gradient_checkpointing": true,
454
+ "gradient_checkpointing_kwargs": null,
455
+ "linear_num_value_heads": null,
456
+ "linear_num_key_heads": null,
457
+ "linear_key_head_dim": null,
458
+ "linear_value_head_dim": null,
459
+ "linear_conv_kernel_dim": null,
460
+ "layer_types": null,
461
+ "mrope_interleaved": false,
462
+ "add_version": true,
463
+ "model_info": "ModelInfo(model_type='glm4_5', model_dir='/root/.cache/modelscope/hub/models/ZhipuAI/GLM-4___5-Air', torch_dtype=torch.bfloat16, max_model_len=131072, quant_method=None, quant_bits=None, rope_scaling=None, is_moe_model=True, config=None, task_type='causal_lm', num_labels=None)",
464
+ "model_meta": "ModelMeta(model_type='glm4_5', model_groups=[ModelGroup(models=[Model(ms_model_id='ZhipuAI/GLM-4.5-Air-Base', hf_model_id='zai-org/GLM-4.5-Air-Base', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='ZhipuAI/GLM-4.5-Air', hf_model_id='zai-org/GLM-4.5-Air', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='ZhipuAI/GLM-4.5-Air-FP8', hf_model_id='zai-org/GLM-4.5-Air-FP8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='ZhipuAI/GLM-4.5-Base', hf_model_id='zai-org/GLM-4.5-Base', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='ZhipuAI/GLM-4.5', hf_model_id='zai-org/GLM-4.5', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='ZhipuAI/GLM-4.5-FP8', hf_model_id='zai-org/GLM-4.5-FP8', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[]), ModelGroup(models=[Model(ms_model_id='ZhipuAI/GLM-4.6', hf_model_id='zai-org/GLM-4.6', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[])], template='glm4_5', get_function=<function get_model_tokenizer_with_flash_attn at 0x76c9ab052520>, model_arch=None, architectures=['Glm4MoeForCausalLM'], additional_saved_files=[], torch_dtype=None, is_multimodal=False, is_reward=False, is_reranker=False, task_type=None, ignore_patterns=None, requires=['transformers>=4.54'], tags=[])",
465
+ "megatron_model_meta": "MegatronModelMeta(megatron_model_type='gpt', model_types=['qwen2', 'qwen2_5', 'qwq', 'qwq_preview', 'qwen2_5_math', 'llama', 'llama3', 'llama3_1', 'llama3_2', 'longwriter_llama3_1', 'codefuse_codellama', 'marco_o1', 'deepseek', 'deepseek_r1_distill', 'yi', 'yi_coder', 'sus', 'skywork_o1', 'openbuddy_llama', 'openbuddy_llama3', 'megrez', 'reflection', 'numina', 'ziya', 'mengzi3', 'qwen3', 'qwen3_thinking', 'qwen3_nothinking', 'qwen2_moe', 'qwen3_moe', 'qwen3_moe_thinking', 'qwen3_coder', 'internlm3', 'mimo', 'mimo_rl', 'moonlight', 'deepseek_moe', 'deepseek_v2', 'deepseek_v2_5', 'deepseek_r1', 'dots1', 'ernie', 'glm4_5', 'deepseek_v3_1', 'ernie_thinking'], convert_mcore2hf=<function convert_mcore2hf at 0x76c91dcb0fe0>, convert_hf2mcore=<function convert_hf2mcore at 0x76c91dcb0b80>, model_cls=<class 'swift.megatron.model.gpt_model.GPTModel'>, convert_hf_config=<function convert_gpt_hf_config at 0x76c91de6c9a0>, get_transformer_layer_spec=None, model_provider=<function model_provider at 0x76c91dddb880>, visual_cls=None, extra_args_provider=None)"
466
+ }
467
+ }
chat_template.jinja ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [gMASK]<sop>
2
+ {%- if tools -%}
3
+ <|system|>
4
+ # Tools
5
+
6
+ You may call one or more functions to assist with the user query.
7
+
8
+ You are provided with function signatures within <tools></tools> XML tags:
9
+ <tools>
10
+ {% for tool in tools %}
11
+ {{ tool | tojson(ensure_ascii=False) }}
12
+ {% endfor %}
13
+ </tools>
14
+
15
+ For each function call, output the function name and arguments within the following XML format:
16
+ <tool_call>{function-name}
17
+ <arg_key>{arg-key-1}</arg_key>
18
+ <arg_value>{arg-value-1}</arg_value>
19
+ <arg_key>{arg-key-2}</arg_key>
20
+ <arg_value>{arg-value-2}</arg_value>
21
+ ...
22
+ </tool_call>{%- endif -%}
23
+ {%- macro visible_text(content) -%}
24
+ {%- if content is string -%}
25
+ {{- content }}
26
+ {%- elif content is iterable and content is not mapping -%}
27
+ {%- for item in content -%}
28
+ {%- if item is mapping and item.type == 'text' -%}
29
+ {{- item.text }}
30
+ {%- elif item is string -%}
31
+ {{- item }}
32
+ {%- endif -%}
33
+ {%- endfor -%}
34
+ {%- else -%}
35
+ {{- content }}
36
+ {%- endif -%}
37
+ {%- endmacro -%}
38
+ {%- set ns = namespace(last_user_index=-1) %}
39
+ {%- for m in messages %}
40
+ {%- if m.role == 'user' %}
41
+ {% set ns.last_user_index = loop.index0 -%}
42
+ {%- endif %}
43
+ {%- endfor %}
44
+ {% for m in messages %}
45
+ {%- if m.role == 'user' -%}<|user|>
46
+ {{ visible_text(m.content) }}
47
+ {{- '/nothink' if (enable_thinking is defined and not enable_thinking and not visible_text(m.content).endswith("/nothink")) else '' -}}
48
+ {%- elif m.role == 'assistant' -%}
49
+ <|assistant|>
50
+ {%- set reasoning_content = '' %}
51
+ {%- set content = visible_text(m.content) %}
52
+ {%- if m.reasoning_content is string %}
53
+ {%- set reasoning_content = m.reasoning_content %}
54
+ {%- else %}
55
+ {%- if '</think>' in content %}
56
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
57
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
58
+ {%- endif %}
59
+ {%- endif %}
60
+ {%- if loop.index0 > ns.last_user_index and reasoning_content -%}
61
+ {{ '\n<think>' + reasoning_content.strip() + '</think>'}}
62
+ {%- else -%}
63
+ {{ '\n<think></think>' }}
64
+ {%- endif -%}
65
+ {%- if content.strip() -%}
66
+ {{ '\n' + content.strip() }}
67
+ {%- endif -%}
68
+ {% if m.tool_calls %}
69
+ {% for tc in m.tool_calls %}
70
+ {%- if tc.function %}
71
+ {%- set tc = tc.function %}
72
+ {%- endif %}
73
+ {{ '\n<tool_call>' + tc.name }}
74
+ {% set _args = tc.arguments %}
75
+ {% for k, v in _args.items() %}
76
+ <arg_key>{{ k }}</arg_key>
77
+ <arg_value>{{ v | tojson(ensure_ascii=False) if v is not string else v }}</arg_value>
78
+ {% endfor %}
79
+ </tool_call>{% endfor %}
80
+ {% endif %}
81
+ {%- elif m.role == 'tool' -%}
82
+ {%- if m.content is string -%}
83
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
84
+ {{- '<|observation|>' }}
85
+ {%- endif %}
86
+ {{- '\n<tool_response>\n' }}
87
+ {{- m.content }}
88
+ {{- '\n</tool_response>' }}
89
+ {%- else -%}
90
+ <|observation|>{% for tr in m.content %}
91
+
92
+ <tool_response>
93
+ {{ tr.output if tr.output is defined else tr }}
94
+ </tool_response>{% endfor -%}
95
+ {% endif -%}
96
+ {%- elif m.role == 'system' -%}
97
+ <|system|>
98
+ {{ visible_text(m.content) }}
99
+ {%- endif -%}
100
+ {%- endfor -%}
101
+ {%- if add_generation_prompt -%}
102
+ <|assistant|>{{- '\n<think></think>' if (enable_thinking is defined and not enable_thinking) else '' -}}
103
+ {%- endif -%}
config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Glm4MoeForCausalLM"
4
+ ],
5
+ "attention_bias": true,
6
+ "attention_dropout": 0.0,
7
+ "dtype": "bfloat16",
8
+ "eos_token_id": [
9
+ 151329,
10
+ 151336,
11
+ 151338
12
+ ],
13
+ "first_k_dense_replace": 1,
14
+ "head_dim": 128,
15
+ "hidden_act": "silu",
16
+ "hidden_size": 4096,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 10944,
19
+ "max_position_embeddings": 131072,
20
+ "model_type": "glm4_moe",
21
+ "moe_intermediate_size": 1408,
22
+ "n_group": 1,
23
+ "n_routed_experts": 128,
24
+ "n_shared_experts": 1,
25
+ "norm_topk_prob": true,
26
+ "num_attention_heads": 96,
27
+ "num_experts_per_tok": 8,
28
+ "num_hidden_layers": 46,
29
+ "num_key_value_heads": 8,
30
+ "num_nextn_predict_layers": 1,
31
+ "pad_token_id": 151329,
32
+ "partial_rotary_factor": 0.5,
33
+ "rms_norm_eps": 1e-05,
34
+ "rope_scaling": null,
35
+ "rope_theta": 1000000,
36
+ "routed_scaling_factor": 1.0,
37
+ "tie_word_embeddings": false,
38
+ "topk_group": 1,
39
+ "transformers_version": "4.57.1",
40
+ "use_cache": true,
41
+ "use_qk_norm": false,
42
+ "vocab_size": 151552
43
+ }
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "eos_token_id": [
4
+ 151329,
5
+ 151336,
6
+ 151338
7
+ ],
8
+ "pad_token_id": 151329,
9
+ "transformers_version": "4.57.1"
10
+ }
model-00001-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:609da4f9b14647e32913e605c6d3c1b724a584cc5d9721961bad3258fd8582be
3
+ size 4991854888
model-00002-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d95fcc93464c8ab05738898664d908e4c1b2956efe1122dd87dca65ee329b357
3
+ size 4994465672
model-00003-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1008d9c8af5b170782337197d8a5d1830b7eed8b773612778ab7e52ec4b04149
3
+ size 4994465696
model-00004-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3b4f107977552119c73649dbc608ee974f578e41926b35f3f20193ad93267f2
3
+ size 4994465696
model-00005-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e01c70beb6782316a3486707b2657938b56665346aae63d4caae899daf2b8fe3
3
+ size 4994465696
model-00006-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6a424615128be0a49c5e2492798716cfdadd262bdb52d89f71260aba3f1e2ce
3
+ size 4974587952
model-00007-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8e145ca170cdcedb9ec180f2cba1e456a288f70a0de9e72df8759ebd2906372
3
+ size 4991318928
model-00008-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daef0e9340911e95cd682de0ae46830956d824db6e2e55ed10ad0f92b2f158d9
3
+ size 4994465664
model-00009-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:614ed88f2e646a963d232f07124086b4f051acf05f73b5c1510f65e15f1340e5
3
+ size 4994465784
model-00010-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afa94ee1f61d3cf4d2934f82a4831b18b67b5cff5a347e8e3ad370e06fe4f4b8
3
+ size 4994466104
model-00011-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a103f658f9cadc68474f761ade2a2e68641d248a868daca5cb6ea90b4c404d7e
3
+ size 4994466104
model-00012-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43aec1f6efe5c34e02836fca52a113a7e0c1785c2ef0929059b6ed3e1b80deac
3
+ size 4994466104
model-00013-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:152a4ef9aa2e6d35ee32b810f34af0e1b02a68f60e564b0765fce44530f17fc3
3
+ size 4994466104
model-00014-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bb3e7527afb28474e66ad7bac133f80953383a578d93b3c80be5ba84a5419fe
3
+ size 4994466104
model-00015-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c945b6280b7fd5482e30d24bc41932510966d39e25dc45dc92173bef919b7b34
3
+ size 4994466104
model-00016-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fad48b5ac703127ce1070b9f4d086917e13d8dccfa0bdee4181643cc9ba4ac4
3
+ size 4994466104
model-00017-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f65c79372dd149d25278d3c7c46e6d52545452d62d7a0a40016c32139058e6c
3
+ size 4994466104
model-00018-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c0f291786d380a490575b26fa6f4dcbaf8cb7e5caeee5437093fd8cbe39bc99
3
+ size 4994466112
model-00019-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bb7bddceeabdc7e91fedc5e2f22f5ac2a4f7e32421f0267339ec725d34561d1
3
+ size 4994466120
model-00020-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0e359451b5092b6bd1eaf60c656279a12209fa6e88852922e694dba9fcdc1bd
3
+ size 4994466120
model-00021-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:698a21c9ea6a431622706b478f5a301bd71096b659bff5772b2ce38de2681fdd
3
+ size 4992409992
model-00022-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a363c5395bdeeccde7cabda8508aafa675153c09f104e9d27858edf3933a69f
3
+ size 4996566656
model-00023-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33c7dfdef1339f56ad43e680a28e5d5ebd491ce358dac16034253402a6ebeb23
3
+ size 4994466088
model-00024-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a8c0eea22e23e3567bf06320bbe5e695a0a374d7288bacf284d22d40f07ff9a
3
+ size 4994466104
model-00025-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77977296accd10b00b600324b0c1789e974aa86adba53c6159a57e613a759ba8
3
+ size 4994466104
model-00026-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:323ff292bc29b7e62950b17e12a2c61a093b047e6b1e7402717123668f1684a0
3
+ size 4994466104
model-00027-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1898220f2ee578f36e0789c6cdfc3e950db59a0908b2ef46c9358179d87f4ece
3
+ size 4994466104
model-00028-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:935ae5a102baa4b7a330aa80b1351229f32cbba63a2c6f0ac0fb436146db27ff
3
+ size 4994466104
model-00029-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e066f0d65f1ab025218d43cb704740cd08c739546bae1fe4aeaab185c0d39b2
3
+ size 4994466104
model-00030-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f15f5f71a0bc28377a43375a0861aa672feaef5526968201c300803bf461849
3
+ size 4994466104
model-00031-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:753900185f0327a533d7e4675f7d61383995ce6a51ca8dd0e6665e0e2da4c81c
3
+ size 4994466104
model-00032-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3f9b1733908e3373011cbce7425ed29902174d448bb78f4ab1abe59f0451d83
3
+ size 4994466104
model-00033-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e3b98923f454f6e2e2e7490557a6ad2cba8eb1736b7c1703e6ac6d07dc934e5
3
+ size 4994466112
model-00034-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26a32b1f64f8aee5f3c08018f3cf468eb162612018ba59dd792162f2b3181148
3
+ size 4994466120
model-00035-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a342ec523cb7b7908c6099564212553e3ca245f48d3d2855e3deb2a0965c54bf
3
+ size 4994466120
model-00036-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d0e8f1e76f05196e8b913d05cdc5ec761117b6c47ccdc0ad0818826ebe62d97
3
+ size 4903256352
model-00037-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9806be2b30d75c6c5eb50f4f6259160788fc827973be43315036d4bc534ef8ce
3
+ size 4993444616
model-00038-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbcd6707495368da430c3f3349166b8e503691802b7dadd071073007c85b48a4
3
+ size 4994466080
model-00039-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5458a42e822b230966e733768bd222cee00470cbe1ff768062c177ade1b64653
3
+ size 4994466104
model-00040-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c6ca5db48d0ec47f7c9948184e88e97a533a94d002e0e82673be745c2a59c7
3
+ size 4994466104
model-00041-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c998055b9b4c6adb458301afa2388089ab1704c04db1ea1384a17e9bbea50f50
3
+ size 4994466104
model-00042-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c886171e3a17f9954ca74bf0ae50873b8da44c10346530378cbe83e418a0317
3
+ size 4994466104
model-00043-of-00043.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8160cd6ebd2457954eeefd046db58d6c5a837bf62884a9bbe78662140f0ebf9
3
+ size 11290194872
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|endoftext|>",
4
+ "[MASK]",
5
+ "[gMASK]",
6
+ "[sMASK]",
7
+ "<sop>",
8
+ "<eop>",
9
+ "<|system|>",
10
+ "<|user|>",
11
+ "<|assistant|>",
12
+ "<|observation|>",
13
+ "<|begin_of_image|>",
14
+ "<|end_of_image|>",
15
+ "<|begin_of_video|>",
16
+ "<|end_of_video|>",
17
+ "<|begin_of_audio|>",
18
+ "<|end_of_audio|>",
19
+ "<|begin_of_transcription|>",
20
+ "<|end_of_transcription|>",
21
+ "<|code_prefix|>",
22
+ "<|code_middle|>",
23
+ "<|code_suffix|>",
24
+ "/nothink"
25
+ ],
26
+ "eos_token": {
27
+ "content": "<|endoftext|>",
28
+ "lstrip": false,
29
+ "normalized": false,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ },
33
+ "pad_token": {
34
+ "content": "<|endoftext|>",
35
+ "lstrip": false,
36
+ "normalized": false,
37
+ "rstrip": false,
38
+ "single_word": false
39
+ }
40
+ }