Xenova HF Staff commited on
Commit
4004872
·
verified ·
0 Parent(s):

Super-squash branch 'main' using huggingface_hub

Browse files
.gitattributes ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,263 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: gemma
3
+ library_name: transformers
4
+ pipeline_tag: image-text-to-text
5
+ tags:
6
+ - transformers.js
7
+ ---
8
+
9
+ # Model Card for Model ID
10
+
11
+ <!-- Provide a quick summary of what the model is/does. -->
12
+ Code to generate this model:
13
+
14
+ ```py
15
+ from transformers import PaliGemmaForConditionalGeneration, PaliGemmaConfig, SiglipVisionConfig, GemmaConfig
16
+
17
+ shared_dim = 16
18
+
19
+ # Initializing a Siglip-like vision config
20
+ vision_config = SiglipVisionConfig(
21
+ hidden_size=16,
22
+ intermediate_size=32,
23
+ num_attention_heads=16,
24
+ num_hidden_layers=2,
25
+ num_image_tokens=256,
26
+ patch_size=14,
27
+ projection_dim=shared_dim,
28
+ projector_hidden_act="gelu_fast",
29
+ vision_use_head=False,
30
+ )
31
+
32
+ # Initializing a Gemma config
33
+ text_config = GemmaConfig(
34
+ hidden_size=shared_dim,
35
+ intermediate_size=16,
36
+ num_attention_heads=4,
37
+ num_hidden_layers=2,
38
+ num_key_value_heads=1,
39
+ vocab_size=257216,
40
+ )
41
+
42
+ # Initializing a PaliGemma paligemma-3b-224 style configuration
43
+ configuration = PaliGemmaConfig(
44
+ vision_config,
45
+ text_config,
46
+ bos_token_id=2,
47
+ eos_token_id=1,
48
+ hidden_size=shared_dim,
49
+ ignore_index=-100,
50
+ image_token_index=257152,
51
+ pad_token_id=0,
52
+ projection_dim=shared_dim,
53
+ )
54
+
55
+ # Initializing a model from the paligemma-3b-224 style configuration
56
+ model = PaliGemmaForConditionalGeneration(configuration)
57
+
58
+ # Randomize weights
59
+ import torch
60
+ torch.manual_seed(0)
61
+ for name, param in model.named_parameters():
62
+ param.data = torch.randn_like(param)
63
+
64
+ # Push to the Hub
65
+ model.push_to_hub('Xenova/tiny-random-PaliGemmaForConditionalGeneration')
66
+ ```
67
+
68
+ Followed by:
69
+ ```py
70
+ from transformers import AutoProcessor
71
+
72
+ processor = AutoProcessor.from_pretrained("google/paligemma-3b-mix-224")
73
+ processor.push_to_hub('Xenova/tiny-random-PaliGemmaForConditionalGeneration')
74
+ ```
75
+
76
+ ## Model Details
77
+
78
+ ### Model Description
79
+
80
+ <!-- Provide a longer summary of what this model is. -->
81
+
82
+ This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
83
+
84
+ - **Developed by:** [More Information Needed]
85
+ - **Funded by [optional]:** [More Information Needed]
86
+ - **Shared by [optional]:** [More Information Needed]
87
+ - **Model type:** [More Information Needed]
88
+ - **Language(s) (NLP):** [More Information Needed]
89
+ - **License:** [More Information Needed]
90
+ - **Finetuned from model [optional]:** [More Information Needed]
91
+
92
+ ### Model Sources [optional]
93
+
94
+ <!-- Provide the basic links for the model. -->
95
+
96
+ - **Repository:** [More Information Needed]
97
+ - **Paper [optional]:** [More Information Needed]
98
+ - **Demo [optional]:** [More Information Needed]
99
+
100
+ ## Uses
101
+
102
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
103
+
104
+ ### Direct Use
105
+
106
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
107
+
108
+ [More Information Needed]
109
+
110
+ ### Downstream Use [optional]
111
+
112
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
113
+
114
+ [More Information Needed]
115
+
116
+ ### Out-of-Scope Use
117
+
118
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
119
+
120
+ [More Information Needed]
121
+
122
+ ## Bias, Risks, and Limitations
123
+
124
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
125
+
126
+ [More Information Needed]
127
+
128
+ ### Recommendations
129
+
130
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
131
+
132
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
133
+
134
+ ## How to Get Started with the Model
135
+
136
+ Use the code below to get started with the model.
137
+
138
+ [More Information Needed]
139
+
140
+ ## Training Details
141
+
142
+ ### Training Data
143
+
144
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
145
+
146
+ [More Information Needed]
147
+
148
+ ### Training Procedure
149
+
150
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
151
+
152
+ #### Preprocessing [optional]
153
+
154
+ [More Information Needed]
155
+
156
+
157
+ #### Training Hyperparameters
158
+
159
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
160
+
161
+ #### Speeds, Sizes, Times [optional]
162
+
163
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
164
+
165
+ [More Information Needed]
166
+
167
+ ## Evaluation
168
+
169
+ <!-- This section describes the evaluation protocols and provides the results. -->
170
+
171
+ ### Testing Data, Factors & Metrics
172
+
173
+ #### Testing Data
174
+
175
+ <!-- This should link to a Dataset Card if possible. -->
176
+
177
+ [More Information Needed]
178
+
179
+ #### Factors
180
+
181
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
182
+
183
+ [More Information Needed]
184
+
185
+ #### Metrics
186
+
187
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
188
+
189
+ [More Information Needed]
190
+
191
+ ### Results
192
+
193
+ [More Information Needed]
194
+
195
+ #### Summary
196
+
197
+
198
+
199
+ ## Model Examination [optional]
200
+
201
+ <!-- Relevant interpretability work for the model goes here -->
202
+
203
+ [More Information Needed]
204
+
205
+ ## Environmental Impact
206
+
207
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
208
+
209
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
210
+
211
+ - **Hardware Type:** [More Information Needed]
212
+ - **Hours used:** [More Information Needed]
213
+ - **Cloud Provider:** [More Information Needed]
214
+ - **Compute Region:** [More Information Needed]
215
+ - **Carbon Emitted:** [More Information Needed]
216
+
217
+ ## Technical Specifications [optional]
218
+
219
+ ### Model Architecture and Objective
220
+
221
+ [More Information Needed]
222
+
223
+ ### Compute Infrastructure
224
+
225
+ [More Information Needed]
226
+
227
+ #### Hardware
228
+
229
+ [More Information Needed]
230
+
231
+ #### Software
232
+
233
+ [More Information Needed]
234
+
235
+ ## Citation [optional]
236
+
237
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
238
+
239
+ **BibTeX:**
240
+
241
+ [More Information Needed]
242
+
243
+ **APA:**
244
+
245
+ [More Information Needed]
246
+
247
+ ## Glossary [optional]
248
+
249
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
250
+
251
+ [More Information Needed]
252
+
253
+ ## More Information [optional]
254
+
255
+ [More Information Needed]
256
+
257
+ ## Model Card Authors [optional]
258
+
259
+ [More Information Needed]
260
+
261
+ ## Model Card Contact
262
+
263
+ [More Information Needed]
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "<image>": 257152
3
+ }
config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_vocab_size": 257152,
3
+ "architectures": [
4
+ "PaliGemmaForConditionalGeneration"
5
+ ],
6
+ "bos_token_id": 2,
7
+ "eos_token_id": 1,
8
+ "hidden_size": 16,
9
+ "image_token_index": 257152,
10
+ "model_type": "paligemma",
11
+ "pad_token_id": 0,
12
+ "projection_dim": 16,
13
+ "text_config": {
14
+ "hidden_size": 16,
15
+ "intermediate_size": 16,
16
+ "model_type": "gemma",
17
+ "num_attention_heads": 4,
18
+ "num_hidden_layers": 2,
19
+ "num_image_tokens": 256,
20
+ "num_key_value_heads": 1,
21
+ "vocab_size": 257216
22
+ },
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.47.0.dev0",
25
+ "vision_config": {
26
+ "hidden_size": 16,
27
+ "intermediate_size": 32,
28
+ "model_type": "siglip_vision_model",
29
+ "num_attention_heads": 16,
30
+ "num_hidden_layers": 2,
31
+ "num_image_tokens": 256,
32
+ "patch_size": 14,
33
+ "projection_dim": 16,
34
+ "projector_hidden_act": "gelu_fast",
35
+ "vision_use_head": false
36
+ }
37
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 2,
4
+ "eos_token_id": 1,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.47.0.dev0"
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:299b82026905d003dd58a212cf35c6e35fe3d04e46b9639e9ad2abef831efe89
3
+ size 16876488
preprocessor_config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": null,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "SiglipImageProcessor",
12
+ "image_seq_length": 256,
13
+ "image_std": [
14
+ 0.5,
15
+ 0.5,
16
+ 0.5
17
+ ],
18
+ "processor_class": "PaliGemmaProcessor",
19
+ "resample": 3,
20
+ "rescale_factor": 0.00392156862745098,
21
+ "size": {
22
+ "height": 224,
23
+ "width": 224
24
+ }
25
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<image>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ }
10
+ ],
11
+ "bos_token": {
12
+ "content": "<bos>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false
17
+ },
18
+ "eos_token": {
19
+ "content": "<eos>",
20
+ "lstrip": false,
21
+ "normalized": false,
22
+ "rstrip": false,
23
+ "single_word": false
24
+ },
25
+ "pad_token": {
26
+ "content": "<pad>",
27
+ "lstrip": false,
28
+ "normalized": false,
29
+ "rstrip": false,
30
+ "single_word": false
31
+ },
32
+ "unk_token": {
33
+ "content": "<unk>",
34
+ "lstrip": false,
35
+ "normalized": false,
36
+ "rstrip": false,
37
+ "single_word": false
38
+ }
39
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e3c3a19458f6053993ad11f9cf1ff4590fa1649cd3d04344233385c08c2cbb8
3
+ size 18650602
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8986bb4f423f07f8c7f70d0dbe3526fb2316056c17bae71b1ea975e77a168fc6
3
+ size 4264023
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff