mshab_checkpoints / README.md
arth-shukla's picture
Update README.md
09b3039 verified
---
license: cc-by-4.0
viewer: false
---
# Model Checkpoints for ManiSkill-HAB
**[Paper](https://arxiv.org/abs/2412.13211)**
| **[Website](https://arth-shukla.github.io/mshab)**
| **[Code](https://github.com/arth-shukla/mshab)**
| **[Models](https://huggingface.co/arth-shukla/mshab_checkpoints)**
| **[Dataset](https://arth-shukla.github.io/mshab/#dataset-section)**
| **[Supplementary](https://sites.google.com/view/maniskill-hab)**
RL (SAC, PPO) and IL (BC, DP) baselines for ManiSkill-HAB. Each checkpoint includes a torch checkpoint `policy.pt` (model, optimizer/scheduler state, other trainable parameters) and a train config `config.yml` with hyperparemeters and env kwargs.
RL Pick/Place policies are trained using SAC due to improved performance, while Open/Close is trained with PPO for wall-time efficiency (see Appendix A.4.3). All-object RL policies are under `all/` directories, while per-object policies are under directories labeled by the object name. IL policies do not have per-object Pick/Place variants.
To download these policies, run the following:
```
huggingface-cli download arth-shukla/mshab_checkpoints --local-dir mshab_checkpoints
```
If you use ManiSkill-HAB in your work, please consider citing the following:
```
@inproceedings{shukla2025maniskillhab,
author = {Arth Shukla and
Stone Tao and
Hao Su},
title = {ManiSkill-HAB: {A} Benchmark for Low-Level Manipulation in Home Rearrangement
Tasks},
booktitle = {The Thirteenth International Conference on Learning Representations,
{ICLR} 2025, Singapore, April 24-28, 2025},
publisher = {OpenReview.net},
year = {2025},
url = {https://openreview.net/forum?id=6bKEWevgSd},
timestamp = {Thu, 15 May 2025 17:19:05 +0200},
biburl = {https://dblp.org/rec/conf/iclr/ShuklaTS25.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
```