zhiyang1 commited on
Commit
9102c71
·
verified ·
1 Parent(s): 03e2534

Upload folder using huggingface_hub

Browse files
Files changed (47) hide show
  1. .gitattributes +2 -0
  2. checkpoint-168000/config.json +79 -0
  3. checkpoint-168000/generation_config.json +6 -0
  4. checkpoint-168000/model.safetensors +3 -0
  5. checkpoint-168000/optimizer.pt +3 -0
  6. checkpoint-168000/rng_state_0.pth +3 -0
  7. checkpoint-168000/rng_state_1.pth +3 -0
  8. checkpoint-168000/rng_state_10.pth +3 -0
  9. checkpoint-168000/rng_state_11.pth +3 -0
  10. checkpoint-168000/rng_state_12.pth +3 -0
  11. checkpoint-168000/rng_state_13.pth +3 -0
  12. checkpoint-168000/rng_state_14.pth +3 -0
  13. checkpoint-168000/rng_state_15.pth +3 -0
  14. checkpoint-168000/rng_state_2.pth +3 -0
  15. checkpoint-168000/rng_state_3.pth +3 -0
  16. checkpoint-168000/rng_state_4.pth +3 -0
  17. checkpoint-168000/rng_state_5.pth +3 -0
  18. checkpoint-168000/rng_state_6.pth +3 -0
  19. checkpoint-168000/rng_state_7.pth +3 -0
  20. checkpoint-168000/rng_state_8.pth +3 -0
  21. checkpoint-168000/rng_state_9.pth +3 -0
  22. checkpoint-168000/scheduler.pt +3 -0
  23. checkpoint-168000/trainer_state.json +3 -0
  24. checkpoint-168000/training_args.bin +3 -0
  25. checkpoint-169000/config.json +79 -0
  26. checkpoint-169000/generation_config.json +6 -0
  27. checkpoint-169000/model.safetensors +3 -0
  28. checkpoint-169000/optimizer.pt +3 -0
  29. checkpoint-169000/rng_state_0.pth +3 -0
  30. checkpoint-169000/rng_state_1.pth +3 -0
  31. checkpoint-169000/rng_state_10.pth +3 -0
  32. checkpoint-169000/rng_state_11.pth +3 -0
  33. checkpoint-169000/rng_state_12.pth +3 -0
  34. checkpoint-169000/rng_state_13.pth +3 -0
  35. checkpoint-169000/rng_state_14.pth +3 -0
  36. checkpoint-169000/rng_state_15.pth +3 -0
  37. checkpoint-169000/rng_state_2.pth +3 -0
  38. checkpoint-169000/rng_state_3.pth +3 -0
  39. checkpoint-169000/rng_state_4.pth +3 -0
  40. checkpoint-169000/rng_state_5.pth +3 -0
  41. checkpoint-169000/rng_state_6.pth +3 -0
  42. checkpoint-169000/rng_state_7.pth +3 -0
  43. checkpoint-169000/rng_state_8.pth +3 -0
  44. checkpoint-169000/rng_state_9.pth +3 -0
  45. checkpoint-169000/scheduler.pt +3 -0
  46. checkpoint-169000/trainer_state.json +3 -0
  47. checkpoint-169000/training_args.bin +3 -0
.gitattributes CHANGED
@@ -76,3 +76,5 @@ checkpoint-153000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
76
  checkpoint-159000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
77
  checkpoint-160000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
78
  checkpoint-161000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
 
 
 
76
  checkpoint-159000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
77
  checkpoint-160000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
78
  checkpoint-161000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
79
+ checkpoint-168000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
80
+ checkpoint-169000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
checkpoint-168000/config.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ar_steps": 1,
3
+ "architectures": [
4
+ "DiffVLMDiffusion"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "condition_layer": -1,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1536,
12
+ "image_token_id": 151655,
13
+ "img_cross_attention_dim": 2048,
14
+ "img_diffuser_depth": 2,
15
+ "img_ffn_dim_multiplier": null,
16
+ "img_hidden_size": 1536,
17
+ "img_multiple_of": 256,
18
+ "img_norm_eps": 1e-05,
19
+ "img_num_attention_heads": 12,
20
+ "img_num_kv_heads": 12,
21
+ "img_qk_norm": true,
22
+ "in_channels": 32,
23
+ "initializer_range": 0.02,
24
+ "inject_img_diffuser": false,
25
+ "input_size": 32,
26
+ "intermediate_size": 8960,
27
+ "layer_group_size": 7,
28
+ "layerwise_start_idx": 0,
29
+ "lora_alpha": 128,
30
+ "lora_bias": "none",
31
+ "lora_dropout": 0.05,
32
+ "lora_enable": false,
33
+ "lora_r": 64,
34
+ "max_position_embeddings": 32768,
35
+ "max_window_layers": 28,
36
+ "model_type": "qwen2_vl",
37
+ "non_linearity": 1,
38
+ "norm_elementwise_affine": true,
39
+ "num_attention_heads": 12,
40
+ "num_hidden_layers": 28,
41
+ "num_key_value_heads": 2,
42
+ "patch_size": 2,
43
+ "repa_coeff": 0.1,
44
+ "repa_layers": "2",
45
+ "repa_shared": false,
46
+ "rms_norm_eps": 1e-06,
47
+ "rope_scaling": {
48
+ "mrope_section": [
49
+ 16,
50
+ 24,
51
+ 24
52
+ ],
53
+ "rope_type": "default",
54
+ "type": "default"
55
+ },
56
+ "rope_theta": 1000000.0,
57
+ "sample_size": 128,
58
+ "sampling_steps": 28,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.47.0",
63
+ "use_cache": true,
64
+ "use_repa": false,
65
+ "use_residual_attn": true,
66
+ "use_sliding_window": false,
67
+ "vae_path": "mit-han-lab/dc-ae-f32c32-sana-1.1-diffusers",
68
+ "video_token_id": 151656,
69
+ "vision_config": {
70
+ "hidden_size": 1536,
71
+ "in_chans": 3,
72
+ "model_type": "qwen2_vl",
73
+ "spatial_patch_size": 14
74
+ },
75
+ "vision_end_token_id": 151653,
76
+ "vision_start_token_id": 151652,
77
+ "vision_token_id": 151654,
78
+ "vocab_size": 151936
79
+ }
checkpoint-168000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.47.0"
6
+ }
checkpoint-168000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7552451b709beb0ebe133997d4462caae411121fe4e086d24aaad88f391e8818
3
+ size 4539487822
checkpoint-168000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f211cb8ce9106f548b53efa2814dced2f859afd56500ff9bf7df38836efa9438
3
+ size 6632345482
checkpoint-168000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47095d6e8479a067cac7482ef288bf54a66e516ded03c30ff03e6acf8e1500f5
3
+ size 15984
checkpoint-168000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f07e225f0098c5093a9cbca5ced6822ce1d9d9c24122e9b6a368f92dac02bad
3
+ size 15984
checkpoint-168000/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb6d04163a9e9ebf1314638e66b1d04f5182c2acde8990479be19e3afc0693e9
3
+ size 15997
checkpoint-168000/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d5e12ee656146ce58dfde9f6ff57978bdd527b956d841f4e48f505ad77f0b0e
3
+ size 15997
checkpoint-168000/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8eb3f9dcfda93b020db012dce6076d3a7bdcb2ed69e977021d5ab536e195707
3
+ size 15997
checkpoint-168000/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:916510491b7a743b57a312d0962db59a81257ea65b604e9e19d20a08498cf493
3
+ size 15997
checkpoint-168000/rng_state_14.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b583a68322fe9c27f1ab0e008d788f07c7f509a156b3b59da3b70e7624c6646
3
+ size 15997
checkpoint-168000/rng_state_15.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f683cda61dda7f7dddf62a6d688862e66e4e60e4d65c40d15b2e84194cb949e6
3
+ size 15997
checkpoint-168000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5f79780d6df7013c9e7f775b7bb3b4ec060ef6abb535830865ba1357dbbe43d
3
+ size 15984
checkpoint-168000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6fc18045879c2f1dd240e699e3311e6f476b0f9dafd2236e199844cb31b8806
3
+ size 15984
checkpoint-168000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e123e1b364813383db271ef899a27885984eeaf0fa8d32feec44a563c1fc23f5
3
+ size 15984
checkpoint-168000/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40b8dff393c971fbc29b05239a5d13a36df9ff1882d91ba07a563226b8804619
3
+ size 15984
checkpoint-168000/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5ec502da037b7c886d48eb92486c74895890e2cd510969c6d718dc3c437a8e7
3
+ size 15984
checkpoint-168000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f25652fca99c6963a237039fa675c4e56ac2dacf9ff5d78c4d578b1a0a7d8bae
3
+ size 15984
checkpoint-168000/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8be49603e9b72379f67a1483d50f12af670d3432aac55cd99c53f64b96744ec
3
+ size 15984
checkpoint-168000/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1faab8d8ceacdcaf31613861b7ef8f7592644dc59b1272fa397f703d5209fd6
3
+ size 15984
checkpoint-168000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f27035c7b2b1e1dbc4c9dafc26d00214857b52178d2506157f12f4c410a289e1
3
+ size 1064
checkpoint-168000/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6963963f9bcdd71b7c633c456b50ff770aff91ab41e296157e8c8cf1178b91b8
3
+ size 25981337
checkpoint-168000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27eb213c3f39bb6fcc57bd148721fd8c5e5d9c5746b727befb50af945420a150
3
+ size 6008
checkpoint-169000/config.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ar_steps": 1,
3
+ "architectures": [
4
+ "DiffVLMDiffusion"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "condition_layer": -1,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1536,
12
+ "image_token_id": 151655,
13
+ "img_cross_attention_dim": 2048,
14
+ "img_diffuser_depth": 2,
15
+ "img_ffn_dim_multiplier": null,
16
+ "img_hidden_size": 1536,
17
+ "img_multiple_of": 256,
18
+ "img_norm_eps": 1e-05,
19
+ "img_num_attention_heads": 12,
20
+ "img_num_kv_heads": 12,
21
+ "img_qk_norm": true,
22
+ "in_channels": 32,
23
+ "initializer_range": 0.02,
24
+ "inject_img_diffuser": false,
25
+ "input_size": 32,
26
+ "intermediate_size": 8960,
27
+ "layer_group_size": 7,
28
+ "layerwise_start_idx": 0,
29
+ "lora_alpha": 128,
30
+ "lora_bias": "none",
31
+ "lora_dropout": 0.05,
32
+ "lora_enable": false,
33
+ "lora_r": 64,
34
+ "max_position_embeddings": 32768,
35
+ "max_window_layers": 28,
36
+ "model_type": "qwen2_vl",
37
+ "non_linearity": 1,
38
+ "norm_elementwise_affine": true,
39
+ "num_attention_heads": 12,
40
+ "num_hidden_layers": 28,
41
+ "num_key_value_heads": 2,
42
+ "patch_size": 2,
43
+ "repa_coeff": 0.1,
44
+ "repa_layers": "2",
45
+ "repa_shared": false,
46
+ "rms_norm_eps": 1e-06,
47
+ "rope_scaling": {
48
+ "mrope_section": [
49
+ 16,
50
+ 24,
51
+ 24
52
+ ],
53
+ "rope_type": "default",
54
+ "type": "default"
55
+ },
56
+ "rope_theta": 1000000.0,
57
+ "sample_size": 128,
58
+ "sampling_steps": 28,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.47.0",
63
+ "use_cache": true,
64
+ "use_repa": false,
65
+ "use_residual_attn": true,
66
+ "use_sliding_window": false,
67
+ "vae_path": "mit-han-lab/dc-ae-f32c32-sana-1.1-diffusers",
68
+ "video_token_id": 151656,
69
+ "vision_config": {
70
+ "hidden_size": 1536,
71
+ "in_chans": 3,
72
+ "model_type": "qwen2_vl",
73
+ "spatial_patch_size": 14
74
+ },
75
+ "vision_end_token_id": 151653,
76
+ "vision_start_token_id": 151652,
77
+ "vision_token_id": 151654,
78
+ "vocab_size": 151936
79
+ }
checkpoint-169000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.47.0"
6
+ }
checkpoint-169000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74e340fafcbb08f9c561fa5a0a1fb6d3a1f7e8448f547fbdb61aab58720cf98a
3
+ size 4539487822
checkpoint-169000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74690add5b9f897413c7c4eb083b4b8b3638143cf4729e305cec996c0049c737
3
+ size 6632345482
checkpoint-169000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f551771e80e92465576f7ec6bb35d58089bb2b45e83b1001694b8332fef91b85
3
+ size 15984
checkpoint-169000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cabb49adb8e39286eb2902a41d4cbe5110e6a3884450ed44c63d18504f5f007b
3
+ size 15984
checkpoint-169000/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0caa5443d0c97e6e0ba405626749bcb0682df4c39bfda6464235399583f02f86
3
+ size 15997
checkpoint-169000/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f0822ac9a788c8cfdb2107eb9ab708d539dfef927fd0a77af683541b8e59021
3
+ size 15997
checkpoint-169000/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2add92aa99b8b4b690be9492ce4dd0f94b1857c4232b92e27163488b40a8fb6c
3
+ size 15997
checkpoint-169000/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b4e559fe570d9faca41f0d6a802bb982092af461b91507908c61d2e44468e3b
3
+ size 15997
checkpoint-169000/rng_state_14.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c7857a235405f86842a4337711c77d17b06483e75e1256ca57ea5574e07c92f
3
+ size 15997
checkpoint-169000/rng_state_15.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6580c3ec681ffd4cb780d146eed1c3052366e130856f3a6261cdb63abed09c65
3
+ size 15997
checkpoint-169000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ecf0ffd35eb31bfc43e438e7bef25682bea76c1f00268e7af78f6d4713fdea1
3
+ size 15984
checkpoint-169000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:917821797d34f5426793870f322e9d86a0fb156c7417cbf25fe5dc32a4f4136a
3
+ size 15984
checkpoint-169000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0d14a6ee57f3efc30803fcceb69b152de401133d39cd0709e0c0e3d887590d1
3
+ size 15984
checkpoint-169000/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43d695c5e6e8ef224abab165068a3b8b2af72261edc60b70063299e8e4c3c505
3
+ size 15984
checkpoint-169000/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bb152494d05b146533ab103ef66e38c4993178bd674f1a286f9c7fb975cc6e4
3
+ size 15984
checkpoint-169000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b09033ee095a81610890754827a2e46fbeb383be2e0f336fac7381fbc4fda2e
3
+ size 15984
checkpoint-169000/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8458f4337b2ffb058902f712a0d87a2e042a32cbaf1f214919ba5af7f71b0327
3
+ size 15984
checkpoint-169000/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3d88ba58d64800c063ae79265e89f3a700b6d02cb20506546e20a6dacb7aab3
3
+ size 15984
checkpoint-169000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b634d20033dde4dae08c987d92b857e6b66854a87dd2d72f4b9a792e04d32a02
3
+ size 1064
checkpoint-169000/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fda4bbaecec7609e1fb8e9ee6447f240450192880c15273aec8bb9d292680dd2
3
+ size 26135995
checkpoint-169000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aab5617b307d003f6641f76d898732dba59f92a0fa1afabd3981effd9cd2c94
3
+ size 6008