Giter Club home page Giter Club logo

idm's People

Contributors

ree1s avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar

idm's Issues

About training on multi-gpus

Hello! Thanks for sharing your excellent work! I tried the training code and have the following two problems.

  1. I have tried training with 2 GPUs using the command in run.sh. But the training stoped after about several hours with errors about l_pixel.backward:

/IDM-main/model/model.py", line 86,in optimize parameters:
l pix.backward()
anaconda3/envs/idm3/lib/python3.9/site-packages/torch/autograd/init .py", in line 200, in backward
Variable. execution engine.run backward #Calls into C++ engine to run the backward pass
RuntimeError:[../third party/gloo/gloo/transport/tcp/pair.cc:598] Connection closed by peer [127.0.0.1]: 47290
ERROR:torch.distributed.elastic.multiprocessing.api:failed (exitcode: 1) local_rank: 0 (pid:2176380) of binary:

Have you ever met this problem?

  1. Regarding the main function, since I met the problem of 1), I tried to use sr.py (in SR3 repository) instead of idm_main.py. idm_main.py uses nn.parallel.DistributedDataParallel while SR3 uses nn.DataParallel. However, when I tried sr.py to use 2 gpus for training, only one single GPU was used. I wonder if you also found this problem and thus use nn.parallel.DistributedDataParallel. Is it caused by basicsr?

Thanks for your patience in reading this issue! Looking forward to your reply.

Applying the pretrained model to other scales

Hi,
Thanks for sharing your excellent work!
I'm wondering if the released pretrained model (x8 SR) can be directly applied to other scales (e.g., x4) by modifying the output resolution. Or should we change the config file first, then train from scratch to obtain models of specific scales?

pre-trained checkpoints

I would like to ask if the size of pre-trained checkpoints/face_gen.pth is suitable for any scale between 16-128 of the face data set or any starting point from *1 to *8, such as 32-256 of the face data set.
My understanding is that the dimensions of face_gen.pth are any scale between 16-128 applicable to the face dataset

About training on a gpus

So is that the start of training? But after an hour, how could it still be like this. How long does it take to train an epoch?
屏幕截图 2023-12-06 163622

23-12-06 07:29:46.861 - INFO: Model [DDPM] is created.
23-12-06 07:29:46.861 - INFO: Initial Model Finished
23-12-06 07:29:46.861 - INFO: Resuming training from epoch: 0, iter: 0.

cuda version

I use cuda toolkit version 11.1; then there will be an issue about "torch.distributed.elastic.multiprocessing.api failed". Do you know if this issue is related to the cuda version?

It seems that the part of SR on general images for DIV2K dataset has not been realized ??

According to the current code for config file and dataset, 'lr_dataroot' and 'hr_dataroot' keys in config files are not used for creating dataset.
And current dataset-related codes, such as 'init.py' and 'LRHR_dataset.py', are only fitted to FFHQ and CelebAHQ dataset.
Besides, the cropped patches of lr_size 16 and hr_size 128 for DIV2K confuse me.
In this way, I wonder whether the part of SR on general images for DIV2K dataset has been realized.

About arbitrary size image upsampling

When I try to upsample arbitrary size image (e.g., 228x344 Set5 woman.png), dimension mismatching error occurs at line 400 in unet.py file.
The problem happens in the else clause at line 404 in the same file, where the argument h is set to the width of feats (e.g., h=feats[-1].shape[-1]).
After I change it to h=feats[-1].shape[-2], arbitrary size image upsampling becomes possible.
If the code is not intended, then it would be nice to fix it.

x = rearrange(x, 'b (h w) c -> b c h w', h=feats[-1].shape[-1])

关于模型训练的问题

@Ree1s 您好,我想请问您一下。能不能在readme里面更新一下详细的训练步骤,我不太清楚您的模型是怎样进行训练的。

About prepare_data.py

Thanks for your great work!
After I ran 'prepare_data.py', I only got two '.mdb' files other than images. Does that right?

Mistake in readme

IDM/README.md

Line 24 in 735e836

conda install pytorch==1.11.0 torchvision==0.12.0 torchaudio==0.11.0 cudatoolkit=11.3

Here in order to install pytorch you need postfix -c pytorch
conda install pytorch==1.11.0 torchvision==0.12.0 torchaudio==0.11.0 cudatoolkit=11.3 -c pytorch

License Query

Can you add a LICENSE to your repository?

MIT would be cool, but CC BY-NC 2.0 is understandable.

You do interesting work!

distributed

I have a question about distributed training, how to run the idm_main.py file on my single graphics card window computer.My question is that "RuntimeError: Default process group has not been initialized, please make sure to call init_process_group."

关于预训练模型

关于预训练模型,我在您给出的链接里只看到了_gen.pth文件却没有_opt.pth文件,我的模型无法训练

Missing checkpoint?

Hey,

For training, I'm missing the checkpoint: 'best_psnr_opt.pth'. Will this be uploaded?

Could you release the train.py and test.py?

Thanks!
I would like to train your model on other tasks.

After reading your excellent work, I would like to know more about the actual training parameters and process of your model.

About training epoch and time

What is the batch size and eopch when training 1616->128128 SRs on two 24GB NVIDIA RTX A5000 GPUs on the FFHQ dataset? How many days did you train at a time? Thanks.

error: [WinError 2] 系统找不到指定的文件

(idm) PS C:\wav2lip\IDM-main> python setup.py develop
C:\ProgramData\Anaconda3\lib\site-packages\setuptools_init_.py:85: _DeprecatedInstaller: setuptools.installer and fetch_build_eggs are deprecated. Requirements should be satisfied by a PEP 517 installer. If you are using pip, you can try pip install --use-pep517.
dist.fetch_build_eggs(dist.setup_requires)
running develop
C:\ProgramData\Anaconda3\lib\site-packages\setuptools\command\easy_install.py:144: EasyInstallDeprecationWarning: easy_install command is deprecated. Use build and pip and other standards-based tools.
warnings.warn(
C:\ProgramData\Anaconda3\lib\site-packages\setuptools\command\install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools.
warnings.warn(
running egg_info
writing basicsr.egg-info\PKG-INFO
writing dependency_links to basicsr.egg-info\dependency_links.txt
writing requirements to basicsr.egg-info\requires.txt
writing top-level names to basicsr.egg-info\top_level.txt
reading manifest file 'basicsr.egg-info\SOURCES.txt'
writing manifest file 'basicsr.egg-info\SOURCES.txt'
running build_ext
error: [WinError 2] 系统找不到指定的文件。

a bug of distributed

when i run this command:
CUDA_VISIBLE_DEVICES=5 python -m torch.distributed.launch idm_main.py -p train -c config/ffhq_liifsr3_scaler_16_128.json -r /home/sx/sx_data/IDM-main/checkpoints/face/home/sx/sx_data/IDM-main/checkpoints/face
this error occurs, how can i fix it?
image
and when i run this command:
CUDA_VISIBLE_DEVICES=5 python -m torch.distributed.launch idm_main.py --local_rank=0 -p train -c config/ffhq_liifsr3_scaler_16_128.json -r /home/sx/sx_data/IDM-main/checkpoints/face/home/sx/sx_data/IDM-main /checkpoints/face
this error occurs,
image

How to decide the performance?

Hi, Reel1s:

Thanks for the nice work. How do you decide the reported performance? Do you select the models from the iterations with best PSNR or other metrics? And do you take average over multiple runs?

Thanks
Jinxin

Image quality on different scales

I can't get a good image on a scale other than 8x even with pre-trained checkpoints. As explained by the README, if additional learning is performed, the results are okay at a scale around 8x, but the same results as the paper cannot be obtained.
How do I get the same results as the paper?

Regarding the resolution of the test image

It seems that the test image resolution size of the given code is all square, for example, 16×16, 128×128. If the resolution of the test image is rectangular, can it be successfully enlarged, for example, 512×256

Results on DIV2K validation set

Dear authors, I have some questions about the results on DIV2K validation set (Table 4):

  1. Did you use resized HR images to generate LR-HR pairs with predefined LR/HR resolutions, rather than the original DIV2K validation set? Could this potentially lead to biased evaluation results by omitting finer details from the original HR images?

  2. How about the LPIPS scores, since all papers in Table 4 are generative SR methods.

  3. Can your model produce full-resolution images on the DIV2K validation set without cropping them into squared dimensions? I'm curious because other works in the field seem to adopt this validation setting.

  4. If not, could you share your specific validation setting, including LR resolution, HR resolution, and any pertinent details?

Thank you! If I made some mistakes, please correct me.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.