Giter Club home page Giter Club logo

omnihang's Introduction

OmniHang: Learning to Hang Arbitrary Objects using Contact Point Correspondences and Neural Collision Estimation

This is the implementation of ICRA 2021 paper "OmniHang: Learning to Hang Arbitrary Objects using Contact Point Correspondences and Neural Collision Estimation" created by Yifan You*, Lin Shao*, Toki Migimatsu, and Jeannette Bohg.

image1

Hanging objects is a common daily task. Our system helps robots learn to hang arbitrary objects onto a diverse set of supporting items such as racks and hooks. All hanging poses rendered here are outputs of our proposed pipeline on object-supporting item pairs unseen during training.

File Structure

This repository provides data and code as follows.

    data/                       # contains all data. Details explained in Dataset Structure section
        
    src/    
        scripts/                        # contains code used to generate data
                ...
                collect_pose_data_vary_scale.py # generates hanging poses
                generate_takeoff_v2.py          # checks if an object in the hanging pose can be taken off
                generate_cp_acc_soft.py         # generates soft contact points, given hanging poses
                generate_partial_pc_soft.py     # generates partial point clouds, given soft contact points

    	utils/					    # something useful

    	lin_my/	                    # training/evaluation 
            runs/                   # contains pretrained models, also where models/tensorboards/debugging info are saved during training/evaluation

            pointnet4/              # code adapted from PointNet++ (https://github.com/charlesq34/pointnet2)
            ...
            simple_dataset.py       # simple dataset that loads supporting items, objects, and successful hanging poses
            hang_dataset.py         # dataset that loads supporting items, objects, successful hanging poses,
                                    # contact points, and contact point correspondences
            ...
            s1_train_matching.py    # stage 1 training/evaluation
            s2a_train.py            # stage 2a training/evaluation
            s2b_train_discrete.py   # stage 2b training/evaluation
            s3_rl_collect.py        # stage 3 online data collection. also used for stage 3 evaluation
            s3_rl_train.py          # stage 3 online training


This code has been tested on Ubuntu 16.04 with Cuda 9.0, Python 3.6, and TensorFlow 1.12.

Dataset Structure

The dataset is organized as follows.

    data/
        ...
        geo_data/                   # urdfs/meshes for objects and supporting items
        geo_data_partial_cp_pad/    # partial point clouds for objects and supporting items
        collection_result/          # successful hanging pose
        collection_result_more/     # more successful hanging pose
        collection_result_neg/      # unsuccessful hanging pose
        collection_result_pene_big_neg_new # object poses w/o collision
        collection_result_pene_big_pos_new # object poses w/ collision
        dataset_cp/                 # contact points and contact point correspondences for poses in collection_result/
        dataset_cp_more/            # contact points and contact point correspondences for poses in collection_result_more/

Dependencies

This repo requires building PointNet++(https://github.com/charlesq34/pointnet2) in src/lin_my/pointnet4/. Please refer to PointNet++'s repo for building instructions.

Downloading Pretrained Models

To download the pretrained models, run

wget http://download.cs.stanford.edu/juno/omnihang/zipdir/runs.zip

unzip the downloaded runs.zip, and move runs/ to src/lin_my/runs.

Downloading Dataset

We split the dataset into several zip files available for download. For all zip files downloaded, unzip the contents and move them under the data/ folder described in Dataset Structure section.

wget http://download.cs.stanford.edu/juno/omnihang/data/zipdir/geo_data.zip # contains the geo_data folder
wget http://download.cs.stanford.edu/juno/omnihang/data/zipdir/collection_result.zip # contains all data related to hanging poses (collection_result/ collection_result_more/, collection_result_neg/, collection_result_pene_big_neg_new/, collection_result_pene_big_pos_new/)
wget http://download.cs.stanford.edu/juno/omnihang/data/zipdir/dataset_cp.zip # contains all data related to contact points (dataset_cp, dataset_cp_more, and some other auxiliary files)
wget http://download.cs.stanford.edu/juno/omnihang/data/zipdir/geo_data_partial_cp_pad.zip # contains the geo_data_partial_cp_pad folder

Questions

Please post issues for questions and more helps on this Github repo page. We encourage using Github issues instead of sending us emails since your questions may benefit others.

Maintainers

@yifan-you-37 @linsats

Citation

@inproceedings{you2021omnihang,
   title={OmniHang: Learning to Hang Arbitrary Objects Using Contact Point Correspondences and Neural Collision Estimation},
   author={You, Yifan and Shao, Lin and Migimatsu, Toki and Bohg, Jeannette},
   booktitle={2021 IEEE International Conference on Robotics and Automation (ICRA)},
   year={2021},
   organization={IEEE}}

License

MIT License

Todos

Please request in Github Issue for more code to release.

omnihang's People

Contributors

yifan-you-37 avatar

Stargazers

SangjunNoh avatar Mingdong Wu avatar  avatar  avatar

Watchers

James Cloos avatar Lin avatar  avatar

omnihang's Issues

Visualize the results

Hi
I'm very interested in your work and would like to reproduce your results.
I downloaded all the data and code, but due to the lack in instructions and documentation, I find it difficult to figure out which scripts should be run in which order and with which arguments. I'd very much appreciate any additional documentation on
-how to visualize the final result, given the pretrained model (similar to the gifs on the project webpage)
-how to train a new model from scratch given the data
-how to generate additional training data

Thank you very much for your consideration.

Question about output of s3_rl_collect

Dear Yifan,

I am very interested in this project and tried to replicate your code by running the file 's3_rl_collect.py' and got a .json file as output. I am not sure about the meaning of each elements in the output. I want to get the final relative pose between hanger and supporting item and success rate. I think that "succ" means whether the hanging of this pair objects is successful, "final_pred_transl" means the translation between two objects and "final_pred_aa" is the orientation between two objects (as marked in the image below)?

Thank you very much!
0B836C20-DD6A-4EFA-A6C9-62A346894E99

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.