Giter Club home page Giter Club logo

sdfusion's Introduction

👋 Hi, I’m @yccyenchicheng, a CS PhD student at UIUC.

👀 I’m interested in generative models.

sdfusion's People

Contributors

yccyenchicheng avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

sdfusion's Issues

No module named 'utils.qual_util'"

Hi, in shape completion code I get a "No module named 'utils.qual_util'" error or the line "SDFusion = create_model(opt)". What could be the cause of this?

Code Problems

Thanks for your great. I would like to know when the code will be released?

Missing file bpe_simple_vocab_16e6.txt.gz

Hi,
While trying to run demo_img2shape.ipynb I encountered the following error, in the cell nr 3 (creating the model):

FileNotFoundError: [Errno 2] No such file or directory: 'random_path/SDFusion/external/clip/bpe_simple_vocab_16e6.txt.gz'

Any help with this one?
Thanks!

About the time and resources consumed during training

Hello, thank you for sharing your great work. I would like to know which categories you trained on shapenet when doing multi-modal generation. How much GPU memory does training take up? How long did it take to train?

about text-to-shape synthesis

Hi,
Thank you for this great work! I have a question about the training process of text-to-shape. According to your paper, the model is trained to generate only the geometry of objects, so how do you handle the color/texture-related texts in the text-to-shape dataset? Did you just let the conditioned diffusion model learns to ignore such information and just capture the geometry-related information? Or did you apply other preprocess to eliminate color-related words in a given text?

Experiment details

Hello @yccyenchicheng,

I apologize for the inconvenience and hope this message finds you well. I find your research to be highly intriguing and I have some inquiries regarding the experimental details.

Specifically, I would like to inquire about the device(s) used when training the Sdfusion-mm2shape. Additionally, I am curious to know how long it took to complete the training process.

Furthermore, I would like to kindly request if you could consider releasing the training code for Sdfusion-mm2shape as soon as possible.

Thank you very much for your time and consideration. I appreciate any assistance you can provide!

Best regards

BuldingNet SDF

Hi,
Impressive work. Is it possible to give the dataset collection and the preprocessing step that went into SDF creation. Because most of the buildings in BuildingNet are open and have holes in them. When i tried SDF with them for a different project of mine, it did not work well. If you already have a subset of buildings that you have collected and preprocessed to make it a closed surface would like to use it for one of my projects.

missing file

./launchers/launch_create_sdf_pix3d.sh: No such file or directory

Error in installation

Hi,
I am trying to install dependency packages for SDFusion, but getting an error in the pytorch3D installation.

I am using cuda 11.7 and nvidia driver 515.
What environment did you use for inference? like cuda, nvidia-driver version etc.

AutoSDF checkpoint of text2shape

Hi,
Appreciate your great work! I notice that you compare the text-guided generation results with AutoSDF. Can you provide your checkpoint of AutoSDF on the text2shape dataset? It seems that AutoSDF only provide checkpoint on a different text-guided dataset (ShapeGlot?). Thank you!

video memory with vqvae resolution

Great work and thanks for sharing! I am trying to increase the resolution in vqvae_snet.yaml, but it seems that setting the resolution more than 64 quickly explode the memory and caused out of video memory. Is there any suggestion to increase the resolution for higher quality models? thanks.

Storage format of dataset

I want to use your model to train my dataset.Can you tell me how to transfer my dataset to the format in the model,liking the format in the ShapeNet_filelists.Thank you

Why do not use clip image and text encoder?

Why do not use clip image and text encoder? In Stable diffusion, text encoder used clip text encoder, and freeze condtion model while training diffusion. But in SDFusion, text encoder do not use clip-text-encoder and text-encoder params do not freeze.

Could you explain it ? Thanks.

set_input self.img = input['img'] KeyError: 'img'

Hi, hope you can help in solving this error:
when running the third step in training for single-view reconstruction:
./launchers/train_sdfusion_img2shape.sh

I get this error:
Traceback (most recent call last):
File "train.py", line 172, in
train_main_worker(opt, model, train_dl, test_dl, test_dl_for_eval, visualizer, device)
File "train.py", line 51, in train_main_worker
model.set_input(data)
File "/home/paperspace/sdf/SDFusion/models/sdfusion_img2shape_model.py", line 240, in set_input
self.img = input['img']
KeyError: 'img'

The conclusion I've come to so far is that the code is written for training on pix3d (in step 3 of training) while I'm training on ShapeNet v1. The dictionary in the file snet_dataset.py does not contain a key called 'img' like the one in pix3d_dataset.py but the error shows it's still expected as an argument.

I tried adding the img in the dictionary but it didn't work.The previous 2 steps in training work fine, but I'm having issues here. Thanks in advance...

No module named 'models.networks.resnet_v1'

in models/sdfusion_mm_model.py
File "/root/h/SDFusion/models/sdfusion_mm_model.py", line 25, in <module> from models.networks.resnet_v1 import resnet18 ModuleNotFoundError: No module named 'models.networks.resnet_v1'

can you provide the related codes? thanks so much!

stride=(1, 2, 2) in Downsample() and Upsample()?

Appreciate your great work!
When I try to run your text2shape model, I notice that stride=(1, 2, 2) in Downsample() and Upsample().
Could you tell me more reasons why you chose stride=(1, 2, 2)? Why not stride=(2, 2, 2)?

eval_metric code not implement.

It seems that there is no evaluation code available. The indicators in the paper cannot be tested for alignment.
Could you provide evaluation code?

Very thanks.

About Multiple Modalities

Hi @yccyenchicheng ,

I have some question about multi-modalities. Which dataset is this portion of the function implemented on? and When the code for this part will be release ?

Best regards,
Yingjie CAI

About Pix3D preprocess~~

I find no match file with launch_create_sdf_pix3d.sh. Could you please re-upload this file?
Thank you very much

Missing model vqvae-snet-all.pth

Thank you for your work and for open-sourcing the code! I am trying to run the demo_txt2shape notebook, but it seems like it's missing the model for vqvae-snet-all.pth. Did I miss something? Thanks for the help!

Pix3D per-class results and data splits

Dear Yen-Chi,

thanks for this great work and open-sourcing the code and model checkpoints!
As there are no official data splits for Pix3D, except for the chair category, could you provide the data splits which were used in the paper?

Additionally, would it be possible for you to provide the per-class results of the single-view reconstruction results (Tab. 2) or even an evaluation script for that part?

Thanks a lot!

Will the ./launchers/train_vqvae-bnet.sh file be provided?

The result in this work is amazing and I'm very interested in the building reconstruction.
I try to train the vqvae and sdfusion models on the BuildingNet dataset, but the related scripts or configs needed by the buildingnet dataset is not provided.
Could you please provide the training files related to the buildingnet dataset?

Questions about get more geometric details

Hi, @yccyenchicheng.
Appreciate your great work! I want to get more geometric details of the generated ShapeNet models. One direct way is to increase the resolution of the SDF to 128, but this requires more memory and computational resources. To reduce memory consumption, it may be necessary to reduce the dimensionality of the latent space. However, I am not sure if reducing the dimensionality of the latent space can result in a good VQVAE model? Do you have any suggestions on this? Or are there other ways to get more geometric details of the generated model?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.