8000 GitHub - lafmdp/ReViWo: Code for ICLR 25 paper: Learning View-invariant World Models for Visual Robotic Manipulation
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content
/ ReViWo Public
forked from Trevor-emt/Reviwo

Code for ICLR 25 paper: Learning View-invariant World Models for Visual Robotic Manipulation

Notifications You must be signed in to change notification settings

lafmdp/ReViWo

 
 

Repository files navigation

[ICLR 2025] Learning View-invariant World Models for Visual Robotic Manipulation

🔧 Python Environment Configuration

  1. Update the prefix parameter in environment.yml
  2. Build Python environment with following command
conda env create -f environment.yml

🚀 View-invariant Encoder Training

  1. Collect the multi-view data from Metaworld with the following command, make sure you have installed mujoco, and we recommend using mujoco-210.
python collect_data/collect_multi_view_data.py
  1. Train the view-invariant encoder by running, the configs of training is referred to pathconfigs/config.yaml:
python tokenizer_main.py

🦾 Running COMBO with the learnt view-invariant encoder

  1. Collect the single-view data for COMBO with the following command:
python collect_data/collect_world_model_training_data.py --env_name ${your_metaworld_env_name}
  1. Running COMBO with the following command. A self-trained checkpoint can be found in "checkpoints/multiview_v0/model.pth" with the default model config in "configs/config.yaml". We provide three settings for evaluation:
  • Training View:
python rl_main.py --env_name ${your_metaworld_env_name} --env_mode "normal"
  • Novel View(CIP):
python rl_main.py --env_name ${your_metaworld_env_name} --env_mode "novel" --camera_change ${change_of_azimuth}
  • Shaking View(CSH):
python rl_main.py --env_name ${your_metaworld_env_name} --env_mode "shake"

😊 Acknowledgement

We would like to thank the authors of OfflineRLKit for their great work and generously providing source codes, which inspired our work and helped us a lot in the implementation.

📚 Citation

If you find our work helpful, please consider citing:

@inproceedings{pang2025reviwo,
  title={Learning View-invariant World Models for Visual Robotic Manipulation},
  author={Jing-Cheng Pang and 
	  Nan Tang and 
	  kaiyuan Li and 
	  Yuting Tang and 
	  Xin-Qiang Cai and 
	  Zhen-Yu Zhang and 
	  Gang Niu and 
	  Sugiyama Masashi and 
	  Yang Yu},
  booktitle={International Conference on Learning Representations (ICLR)},
  year={2025}
}

About

Code for ICLR 25 paper: Learning View-invariant World Models for Visual Robotic Manipulation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%
0