ydshieh HF Staff commited on
Commit
e460201
·
verified ·
1 Parent(s): 5bebf1e

Upload model

Browse files
Files changed (4) hide show
  1. .gitattributes +1 -0
  2. config.json +11 -104
  3. generation_config.json +8 -0
  4. model.safetensors +3 -0
.gitattributes CHANGED
@@ -25,3 +25,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
28
+ model.safetensors filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -4,41 +4,18 @@
4
  ],
5
  "bos_token_id": 50256,
6
  "decoder": {
7
- "_name_or_path": "",
8
  "activation_function": "gelu_new",
9
  "add_cross_attention": true,
10
  "architectures": [
11
  "GPT2LMHeadModel"
12
  ],
13
  "attn_pdrop": 0.1,
14
- "bad_words_ids": null,
15
- "bos_token_id": 50256,
16
- "chunk_size_feed_forward": 0,
17
  "decoder_start_token_id": 50256,
18
- "diversity_penalty": 0.0,
19
- "do_sample": false,
20
- "early_stopping": false,
21
  "embd_pdrop": 0.1,
22
- "encoder_no_repeat_ngram_size": 0,
23
- "eos_token_id": 50256,
24
- "finetuning_task": null,
25
- "forced_bos_token_id": null,
26
- "forced_eos_token_id": null,
27
- "id2label": {
28
- "0": "LABEL_0",
29
- "1": "LABEL_1"
30
- },
31
  "initializer_range": 0.02,
32
  "is_decoder": true,
33
- "is_encoder_decoder": false,
34
- "label2id": {
35
- "LABEL_0": 0,
36
- "LABEL_1": 1
37
- },
38
  "layer_norm_epsilon": 1e-05,
39
- "length_penalty": 1.0,
40
- "max_length": 20,
41
- "min_length": 0,
42
  "model_type": "gpt2",
43
  "n_ctx": 1024,
44
  "n_embd": 768,
@@ -46,24 +23,11 @@
46
  "n_inner": null,
47
  "n_layer": 12,
48
  "n_positions": 1024,
49
- "no_repeat_ngram_size": 0,
50
- "num_beam_groups": 1,
51
- "num_beams": 1,
52
- "num_return_sequences": 1,
53
- "output_attentions": false,
54
- "output_hidden_states": false,
55
- "output_scores": false,
56
  "pad_token_id": 50256,
57
- "prefix": null,
58
- "problem_type": null,
59
- "pruned_heads": {},
60
- "remove_invalid_values": false,
61
- "repetition_penalty": 1.0,
62
  "resid_pdrop": 0.1,
63
- "return_dict": true,
64
- "return_dict_in_generate": false,
65
  "scale_attn_weights": true,
66
- "sep_token_id": null,
67
  "summary_activation": null,
68
  "summary_first_dropout": 0.1,
69
  "summary_proj_to_labels": true,
@@ -75,95 +39,38 @@
75
  "max_length": 50
76
  }
77
  },
78
- "temperature": 1.0,
79
- "tie_encoder_decoder": false,
80
- "tie_word_embeddings": true,
81
- "tokenizer_class": null,
82
- "top_k": 50,
83
- "top_p": 1.0,
84
- "torch_dtype": null,
85
- "torchscript": false,
86
- "transformers_version": "4.11.0.dev0",
87
- "use_bfloat16": false,
88
  "use_cache": true,
89
  "vocab_size": 50257
90
  },
91
  "decoder_start_token_id": 50256,
 
92
  "encoder": {
93
- "_name_or_path": "",
94
- "add_cross_attention": false,
95
  "architectures": [
96
  "ViTModel"
97
  ],
98
  "attention_probs_dropout_prob": 0.0,
99
- "bad_words_ids": null,
100
- "bos_token_id": null,
101
- "chunk_size_feed_forward": 0,
102
- "decoder_start_token_id": null,
103
- "diversity_penalty": 0.0,
104
- "do_sample": false,
105
- "early_stopping": false,
106
- "encoder_no_repeat_ngram_size": 0,
107
- "eos_token_id": null,
108
- "finetuning_task": null,
109
- "forced_bos_token_id": null,
110
- "forced_eos_token_id": null,
111
  "hidden_act": "gelu",
112
  "hidden_dropout_prob": 0.0,
113
  "hidden_size": 768,
114
- "id2label": {
115
- "0": "LABEL_0",
116
- "1": "LABEL_1"
117
- },
118
  "image_size": 224,
119
  "initializer_range": 0.02,
120
  "intermediate_size": 3072,
121
- "is_decoder": false,
122
- "is_encoder_decoder": false,
123
- "label2id": {
124
- "LABEL_0": 0,
125
- "LABEL_1": 1
126
- },
127
  "layer_norm_eps": 1e-12,
128
- "length_penalty": 1.0,
129
- "max_length": 20,
130
- "min_length": 0,
131
  "model_type": "vit",
132
- "no_repeat_ngram_size": 0,
133
  "num_attention_heads": 12,
134
- "num_beam_groups": 1,
135
- "num_beams": 1,
136
  "num_channels": 3,
137
  "num_hidden_layers": 12,
138
- "num_return_sequences": 1,
139
- "output_attentions": false,
140
- "output_hidden_states": false,
141
- "output_scores": false,
142
- "pad_token_id": null,
143
  "patch_size": 16,
144
- "prefix": null,
145
- "problem_type": null,
146
- "pruned_heads": {},
147
- "remove_invalid_values": false,
148
- "repetition_penalty": 1.0,
149
- "return_dict": true,
150
- "return_dict_in_generate": false,
151
- "sep_token_id": null,
152
- "task_specific_params": null,
153
- "temperature": 1.0,
154
- "tie_encoder_decoder": false,
155
- "tie_word_embeddings": true,
156
- "tokenizer_class": null,
157
- "top_k": 50,
158
- "top_p": 1.0,
159
- "torch_dtype": null,
160
- "torchscript": false,
161
- "transformers_version": "4.11.0.dev0",
162
- "use_bfloat16": false
163
  },
164
  "eos_token_id": 50256,
165
  "is_encoder_decoder": true,
166
  "model_type": "vision-encoder-decoder",
167
  "pad_token_id": 50256,
168
- "transformers_version": null
 
169
  }
 
4
  ],
5
  "bos_token_id": 50256,
6
  "decoder": {
 
7
  "activation_function": "gelu_new",
8
  "add_cross_attention": true,
9
  "architectures": [
10
  "GPT2LMHeadModel"
11
  ],
12
  "attn_pdrop": 0.1,
 
 
 
13
  "decoder_start_token_id": 50256,
14
+ "dtype": "float32",
 
 
15
  "embd_pdrop": 0.1,
 
 
 
 
 
 
 
 
 
16
  "initializer_range": 0.02,
17
  "is_decoder": true,
 
 
 
 
 
18
  "layer_norm_epsilon": 1e-05,
 
 
 
19
  "model_type": "gpt2",
20
  "n_ctx": 1024,
21
  "n_embd": 768,
 
23
  "n_inner": null,
24
  "n_layer": 12,
25
  "n_positions": 1024,
 
 
 
 
 
 
 
26
  "pad_token_id": 50256,
27
+ "reorder_and_upcast_attn": false,
 
 
 
 
28
  "resid_pdrop": 0.1,
29
+ "scale_attn_by_inverse_layer_idx": false,
 
30
  "scale_attn_weights": true,
 
31
  "summary_activation": null,
32
  "summary_first_dropout": 0.1,
33
  "summary_proj_to_labels": true,
 
39
  "max_length": 50
40
  }
41
  },
 
 
 
 
 
 
 
 
 
 
42
  "use_cache": true,
43
  "vocab_size": 50257
44
  },
45
  "decoder_start_token_id": 50256,
46
+ "dtype": "float32",
47
  "encoder": {
 
 
48
  "architectures": [
49
  "ViTModel"
50
  ],
51
  "attention_probs_dropout_prob": 0.0,
52
+ "dtype": "float32",
53
+ "encoder_stride": 16,
 
 
 
 
 
 
 
 
 
 
54
  "hidden_act": "gelu",
55
  "hidden_dropout_prob": 0.0,
56
  "hidden_size": 768,
 
 
 
 
57
  "image_size": 224,
58
  "initializer_range": 0.02,
59
  "intermediate_size": 3072,
 
 
 
 
 
 
60
  "layer_norm_eps": 1e-12,
 
 
 
61
  "model_type": "vit",
 
62
  "num_attention_heads": 12,
 
 
63
  "num_channels": 3,
64
  "num_hidden_layers": 12,
 
 
 
 
 
65
  "patch_size": 16,
66
+ "pooler_act": "tanh",
67
+ "pooler_output_size": 768,
68
+ "qkv_bias": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
69
  },
70
  "eos_token_id": 50256,
71
  "is_encoder_decoder": true,
72
  "model_type": "vision-encoder-decoder",
73
  "pad_token_id": 50256,
74
+ "tie_word_embeddings": false,
75
+ "transformers_version": "4.57.3"
76
  }
generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "decoder_start_token_id": 50256,
5
+ "eos_token_id": 50256,
6
+ "pad_token_id": 50256,
7
+ "transformers_version": "4.57.3"
8
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6ac1fdab7481c560cc103a2387a7ab45581b8f3330b0731eff4167cdff27ab7
3
+ size 956835520