Ethan-pooh commited on
Commit
4454d14
1 Parent(s): 6ab993b

End of training

Browse files
README.md CHANGED
@@ -2,10 +2,17 @@
2
  ---
3
  license: mit
4
  base_model: robotics-diffusion-transformer/rdt-1b
 
 
 
 
5
  tags:
6
  - robotics
7
  - pytorch
8
- - transformers
 
 
 
9
  - rdt
10
  ---
11
 
 
2
  ---
3
  license: mit
4
  base_model: robotics-diffusion-transformer/rdt-1b
5
+ language:
6
+ - en
7
+ pipeline_tag: robotics
8
+ library_name: transformers
9
  tags:
10
  - robotics
11
  - pytorch
12
+ - multimodal
13
+ - pretraining
14
+ - vla
15
+ - diffusion
16
  - rdt
17
  ---
18
 
checkpoint-10/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - model_hub_mixin
4
+ - pytorch_model_hub_mixin
5
+ ---
6
+
7
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
+ - Library: https://huggingface.co/robotics-diffusion-transformer/rdt-1b
9
+ - Docs: [More Information Needed]
checkpoint-10/config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "action_dim": 128,
3
+ "ema": {
4
+ "inv_gamma": 1.0,
5
+ "max_value": 0.9999,
6
+ "min_value": 0.0,
7
+ "power": 0.75,
8
+ "update_after_step": 0
9
+ },
10
+ "img_adaptor": "mlp2x_gelu",
11
+ "img_cond_len": 4374,
12
+ "img_pos_embed_config": [
13
+ [
14
+ "image",
15
+ [
16
+ 2,
17
+ 3,
18
+ -729
19
+ ]
20
+ ]
21
+ ],
22
+ "img_token_dim": 1152,
23
+ "lang_adaptor": "mlp2x_gelu",
24
+ "lang_pos_embed_config": [
25
+ [
26
+ "lang",
27
+ -1024
28
+ ]
29
+ ],
30
+ "lang_token_dim": 4096,
31
+ "max_lang_cond_len": 1024,
32
+ "noise_scheduler": {
33
+ "beta_schedule": "squaredcos_cap_v2",
34
+ "clip_sample": false,
35
+ "num_inference_timesteps": 5,
36
+ "num_train_timesteps": 1000,
37
+ "prediction_type": "sample",
38
+ "type": "ddpm"
39
+ },
40
+ "pred_horizon": 64,
41
+ "rdt": {
42
+ "cond_pos_embed_type": "multimodal",
43
+ "depth": 28,
44
+ "hidden_size": 2048,
45
+ "num_heads": 32
46
+ },
47
+ "state_adaptor": "mlp3x_gelu",
48
+ "state_token_dim": 128
49
+ }
checkpoint-5/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - model_hub_mixin
4
+ - pytorch_model_hub_mixin
5
+ ---
6
+
7
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
+ - Library: https://huggingface.co/robotics-diffusion-transformer/rdt-1b
9
+ - Docs: [More Information Needed]
checkpoint-5/config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "action_dim": 128,
3
+ "ema": {
4
+ "inv_gamma": 1.0,
5
+ "max_value": 0.9999,
6
+ "min_value": 0.0,
7
+ "power": 0.75,
8
+ "update_after_step": 0
9
+ },
10
+ "img_adaptor": "mlp2x_gelu",
11
+ "img_cond_len": 4374,
12
+ "img_pos_embed_config": [
13
+ [
14
+ "image",
15
+ [
16
+ 2,
17
+ 3,
18
+ -729
19
+ ]
20
+ ]
21
+ ],
22
+ "img_token_dim": 1152,
23
+ "lang_adaptor": "mlp2x_gelu",
24
+ "lang_pos_embed_config": [
25
+ [
26
+ "lang",
27
+ -1024
28
+ ]
29
+ ],
30
+ "lang_token_dim": 4096,
31
+ "max_lang_cond_len": 1024,
32
+ "noise_scheduler": {
33
+ "beta_schedule": "squaredcos_cap_v2",
34
+ "clip_sample": false,
35
+ "num_inference_timesteps": 5,
36
+ "num_train_timesteps": 1000,
37
+ "prediction_type": "sample",
38
+ "type": "ddpm"
39
+ },
40
+ "pred_horizon": 64,
41
+ "rdt": {
42
+ "cond_pos_embed_type": "multimodal",
43
+ "depth": 28,
44
+ "hidden_size": 2048,
45
+ "num_heads": 32
46
+ },
47
+ "state_adaptor": "mlp3x_gelu",
48
+ "state_token_dim": 128
49
+ }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe217f4491ea882b0b52df1cb23ae4e8a11c1328ed29f2ed712e02aad2c02102
3
  size 2456755578
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aafd1b2b11e457fdf5fec0e762ee07053d39384a178b6ff068a8564f8b6ee801
3
  size 2456755578