miscalculation's picture
Upload folder using huggingface_hub
f237283 verified
adapter_args:
activation_class: SILU
adapter_type: LINEAR
hidden_dim: output_dim
input_norm: false
norm_instance:
norm_class: LN
norm_eps: 1.0e-06
output_norm: false
pooling_factor_per_dim: 1
pre_project_norm: false
use_biases: true
use_gating: false
decoder_args:
activation_class: GELU_TANH
checkpoint: false
dim: 2048
dropout: 0.0
head_dim: 256
hidden_dim: 16384
n_heads: 8
n_kv_heads: 1
n_layers: 18
norm_args:
attention_norm: true
ffn_norm: true
norm_instance:
norm_class: RMS
norm_eps: 1.0e-06
post_attention_norm: false
post_ffn_norm: false
qk_norm: false
pos_embed: ROPE_1D
rope_freqs_split: null
rope_theta: 10000.0
total_parameter_count: 1981882368
use_attn_wk_biases: false
use_attn_wo_biases: false
use_attn_wq_biases: false
use_attn_wv_biases: false
use_ffn_biases: false
use_ffn_gating: true
image_encoder_args:
image_pooling_args:
pooling_type: NONE
max_image_size: 224
num_register_tokens: 0
output_norm: true
patch_norm: false
patch_size: 14
total_parameter_count: 412442352
transformer_args:
activation_class: GELU_TANH
checkpoint: false
dim: 1152
dropout: 0.0
head_dim: 72
hidden_dim: 4304
n_heads: 16
n_kv_heads: 16
n_layers: 27
norm_args:
attention_norm: true
ffn_norm: true
norm_instance:
norm_class: LN
norm_eps: 1.0e-06
post_attention_norm: false
post_ffn_norm: false
qk_norm: false
pos_embed: LEARNED
rope_freqs_split: null
rope_theta: null
total_parameter_count: 411466608
use_attn_wk_biases: true
use_attn_wo_biases: true
use_attn_wq_biases: true
use_attn_wv_biases: true
use_ffn_biases: true
use_ffn_gating: false
use_cls_token: false
use_patch_conv_biases: true
image_token_id: 257152
img_encoding_tokens:
- 257152
model_parallel: 1
model_type: vla_autoregressive
processor_args:
action_tokenizer_args:
fault_tolerant_decoding: true
max_action: 2.0
min_action: -2.0
n_bins: null
tokenizer_type: FAST
hardware_processor_args:
FRANKA_LIBERO:
action_space:
format: default
horizon: 10
names: null
tensors:
action:
dtype: float32
high: .inf
low: -.inf
names:
- action_{i}
- action_{i}
- action_{i}
- action_{i}
- action_{i}
- action_{i}
- action_{i}
tensor_decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
video_info:
has_audio: false
video.codec: h264
video.fps: 10.0
video.is_depth_map: false
video.pix_fmt: yuv420p
input_action_transforms:
- action_stats:
stats:
action:
max:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIAAAAAAAAO4/AAAAAAAA7j8AAAAAAADuPwAAAOAFxNY/AAAAAAAA2D8AAAAAAADYPwAAAAAAAPA/
mean:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIDoVfuNzErA/fg6uBzM7tj+6fG5WsCK3v/hkJ+sUuEE/y706MIMddz+d2v9xG2t1v7uhTTiCaqm/
min:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIAAAAAAAAO6/AAAAAAAA7r8AAAAAAADuvwAAAECVhtC/AAAAAAAA2L8AAADAHoXXvwAAAAAAAPC/
std:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gID61+XE4edU/R7zQtXk42D8OMzLxbnbcP4PD/LW6F6Q/DQt/c4U6sD8uF2Au3PWzP8wXQYjm9e8/
observation.state:
max:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIAAAACCw68o/AAAA4NQK2T8AAADgLdv1PwAAAGAUXw1AAAAAYDZ8DEAAAABgci72PwAAAACMraU/AAAAoBpWVj8=
mean:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIBNme1VP0ae/GQwTsRCeoT9+Wp/PNnfoP0o858UVxwdAPv83m1o4zL9NbNdi/BLAvxfFWPZoj5s/rKiMmebXm78=
min:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIAAAACCH5t6/AAAAQBHV1L8AAADggaWAPwAAAADVk9Y/AAAAQKYhDcAAAABA23v9vwAAAOCNQla/AAAAQGCGpb8=
std:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIHpiAWeb3bo/tHB6GhNtwz9rUe4jnjnYP0P+b3mTCNY/jjhbZ7UF7T86wtSNONPUPy7+dppBCI0/kcnB2ejKjD8=
norm_type: MIN_MAX
transform_type: normalize_action
input_observation_transforms:
- norm_type: MIN_MAX
state_stats:
stats:
action:
max:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIAAAAAAAAO4/AAAAAAAA7j8AAAAAAADuPwAAAOAFxNY/AAAAAAAA2D8AAAAAAADYPwAAAAAAAPA/
mean:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIDoVfuNzErA/fg6uBzM7tj+6fG5WsCK3v/hkJ+sUuEE/y706MIMddz+d2v9xG2t1v7uhTTiCaqm/
min:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIAAAAAAAAO6/AAAAAAAA7r8AAAAAAADuvwAAAECVhtC/AAAAAAAA2L8AAADAHoXXvwAAAAAAAPC/
std:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gID61+XE4edU/R7zQtXk42D8OMzLxbnbcP4PD/LW6F6Q/DQt/c4U6sD8uF2Au3PWzP8wXQYjm9e8/
observation.state:
max:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIAAAACCw68o/AAAA4NQK2T8AAADgLdv1PwAAAGAUXw1AAAAAYDZ8DEAAAABgci72PwAAAACMraU/AAAAoBpWVj8=
mean:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIBNme1VP0ae/GQwTsRCeoT9+Wp/PNnfoP0o858UVxwdAPv83m1o4zL9NbNdi/BLAvxfFWPZoj5s/rKiMmebXm78=
min:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIAAAACCH5t6/AAAAQBHV1L8AAADggaWAPwAAAADVk9Y/AAAAQKYhDcAAAABA23v9vwAAAOCNQla/AAAAQGCGpb8=
std:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIHpiAWeb3bo/tHB6GhNtwz9rUe4jnjnYP0P+b3mTCNY/jjhbZ7UF7T86wtSNONPUPy7+dppBCI0/kcnB2ejKjD8=
transform_type: normalize_observation
observation_space:
format: default
images:
observation.images.image:
height: 256
n_channels: 3
video_info:
has_audio: false
video.codec: h264
video.fps: 10.0
video.is_depth_map: false
video.pix_fmt: yuv420p
width: 256
observation.images.wrist_image:
height: 256
n_channels: 3
video_info:
has_audio: false
video.codec: h264
video.fps: 10.0
video.is_depth_map: false
video.pix_fmt: yuv420p
width: 256
names: null
state:
observation.state:
dtype: float64
high: .inf
low: -.inf
names:
- observation.state_{i}
- observation.state_{i}
- observation.state_{i}
- observation.state_{i}
- observation.state_{i}
- observation.state_{i}
- observation.state_{i}
- observation.state_{i}
tensor_decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
video_info:
has_audio: false
video.codec: h264
video.fps: 10.0
video.is_depth_map: false
video.pix_fmt: yuv420p
output_action_transforms:
- action_stats:
stats:
action:
max:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIAAAAAAAAO4/AAAAAAAA7j8AAAAAAADuPwAAAOAFxNY/AAAAAAAA2D8AAAAAAADYPwAAAAAAAPA/
mean:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIDoVfuNzErA/fg6uBzM7tj+6fG5WsCK3v/hkJ+sUuEE/y706MIMddz+d2v9xG2t1v7uhTTiCaqm/
min:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIAAAAAAAAO6/AAAAAAAA7r8AAAAAAADuvwAAAECVhtC/AAAAAAAA2L8AAADAHoXXvwAAAAAAAPC/
std:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gID61+XE4edU/R7zQtXk42D8OMzLxbnbcP4PD/LW6F6Q/DQt/c4U6sD8uF2Au3PWzP8wXQYjm9e8/
observation.state:
max:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIAAAACCw68o/AAAA4NQK2T8AAADgLdv1PwAAAGAUXw1AAAAAYDZ8DEAAAABgci72PwAAAACMraU/AAAAoBpWVj8=
mean:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIBNme1VP0ae/GQwTsRCeoT9+Wp/PNnfoP0o858UVxwdAPv83m1o4zL9NbNdi/BLAvxfFWPZoj5s/rKiMmebXm78=
min:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIAAAACCH5t6/AAAAQBHV1L8AAADggaWAPwAAAADVk9Y/AAAAQKYhDcAAAABA23v9vwAAAOCNQla/AAAAQGCGpb8=
std:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIHpiAWeb3bo/tHB6GhNtwz9rUe4jnjnYP0P+b3mTCNY/jjhbZ7UF7T86wtSNONPUPy7+dppBCI0/kcnB2ejKjD8=
norm_type: MIN_MAX
transform_type: denormalize_action
output_observation_transforms:
- norm_type: MIN_MAX
state_stats:
stats:
action:
max:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIAAAAAAAAO4/AAAAAAAA7j8AAAAAAADuPwAAAOAFxNY/AAAAAAAA2D8AAAAAAADYPwAAAAAAAPA/
mean:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIDoVfuNzErA/fg6uBzM7tj+6fG5WsCK3v/hkJ+sUuEE/y706MIMddz+d2v9xG2t1v7uhTTiCaqm/
min:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gIAAAAAAAAO6/AAAAAAAA7r8AAAAAAADuvwAAAECVhtC/AAAAAAAA2L8AAADAHoXXvwAAAAAAAPC/
std:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoAction
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbN10sImRhdGFfb2Zmc2V0cyI6WzAsNTZdfX0gID61+XE4edU/R7zQtXk42D8OMzLxbnbcP4PD/LW6F6Q/DQt/c4U6sD8uF2Au3PWzP8wXQYjm9e8/
observation.state:
max:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIAAAACCw68o/AAAA4NQK2T8AAADgLdv1PwAAAGAUXw1AAAAAYDZ8DEAAAABgci72PwAAAACMraU/AAAAoBpWVj8=
mean:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIBNme1VP0ae/GQwTsRCeoT9+Wp/PNnfoP0o858UVxwdAPv83m1o4zL9NbNdi/BLAvxfFWPZoj5s/rKiMmebXm78=
min:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIAAAACCH5t6/AAAAQBHV1L8AAADggaWAPwAAAADVk9Y/AAAAQKYhDcAAAABA23v9vwAAAOCNQla/AAAAQGCGpb8=
std:
batch_dims: []
decoder:
tensor_decoder_type: default
type: gs_vla.third_party.libero.data.LiberoObservation
tensor: OAAAAAAAAAB7IiI6eyJkdHlwZSI6IkY2NCIsInNoYXBlIjpbOF0sImRhdGFfb2Zmc2V0cyI6WzAsNjRdfX0gIHpiAWeb3bo/tHB6GhNtwz9rUe4jnjnYP0P+b3mTCNY/jjhbZ7UF7T86wtSNONPUPy7+dppBCI0/kcnB2ejKjD8=
transform_type: denormalize_observation
processor_type: vla
vlm_processor_args:
annotation_formatter_args:
formatter_type: COMPACT_DISCRETE
bottleneck: null
decoder_attention_args:
pattern: PREFIX_LM
image_preprocess_args:
interpolation_mode: BICUBIC
max_image_size: 224
patch_size: 14
pooling_factor_per_dim: 1
resize_mode: FIXED
rgb_normalization: SIGLIP
instruct_template_args:
template_type: GENESIS
processor_type: vlm
tokenizer_args:
tokenizer_type: PALIGEMMA
total_parameter_count: 3450244848
use_output_biases: false
vocab_size: 257216