Giter Club home page Giter Club logo

deepfillv2's Introduction

Hi ๐Ÿ‘‹, I'm Yuzhi @zhaoyuzhi

Researcher in Computer Vision | Ph.D., City University of Hong Kong

โœจ Quick Facts

  • ๐Ÿ”ญ I got the Ph.D. degree from City University of Hong Kong, Hong Kong SAR, China and the B.Eng degree from Huazhong University of Science of Technology, Wuhan, China

  • ๐ŸŒฑ My research interests include image and video processing and generative models. Recently, I focus on AI-Generated Content (AIGC) and Multimodal Large Language Model (MLLM)

  • ๐Ÿ’ฌ How to reach me: [email protected]

  • ๐Ÿ“ซ My personal webpage: https://zhaoyuzhi.github.io/

  • ๐Ÿ“„ My Google scholar webpage: https://scholar.google.com/citations?user=OtoqVTIAAAAJ&hl=zh-CN/

deepfillv2's People

Contributors

zhaoyuzhi avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar

deepfillv2's Issues

what is the value of opt.lr?

Wonderful work!
I am sorry to bother you,but I have a question about the value of opt.lr in train.py.

Traceback (most recent call last):
File "train.py", line 62, in
trainer.WGAN_trainer(opt)
File "/home/fl/deepfillv2/trainer.py", line 152, in WGAN_trainer
adjust_learning_rate(optimizer_g, (epoch + 1), opt)
File "/home/fl/deepfillv2/trainer.py", line 53, in adjust_learning_rate
lr = opt.lr * (opt.lr_decrease_factor ** (epoch // opt.lr_decrease_epoch))
AttributeError: 'Namespace' object has no attribute 'lr'

I would very appreciate if you could help me,thank you!

new model and code released

I have released all new code and models including colorful and grayscale image inpainting. Please refer to README

A question on block6 of discriminator

May I ask one question on discriminator output?
The code shows the discriminator are suppose to output 256 channels
x = self.block6(x) # out: [B, 256, 8, 8]

However, the definition of block6 are shown below
self.block6 = Conv2dLayer(opt.latent_channels * 4, 1, 4, 2, 1, pad_type = opt.pad_type, activation = 'none', norm = 'none', sn = True)
It seems it will output 1 channel instead of 256 channels

In the deepfillv2 paper, it is trying to output >1 channels. May I ask why the implementation changes this?

How long does it take to train the model?

Hello, I really appreciate your elegant implementation of the gated convolution and the coarse to fine structure. I am very curious about how long does it take to get the comparable result as the original paper?

Value of weight_decay for optimizers

In the trainer.py, line no. 53 and 54, you have weight_decay = opt.weight_decay. I tried to find the value of weight_decay in run.sh file, but I could not find it. Kindly let us know the value you used.

There are no pretrained weights for RGB?

Missing pretrained VGG16

Hey,

Great work and really appreciate for sharing the code.
I want to train deepfillv2 for RGB images. But, when I tried to train it on custom data it raises an error in line number 39 in utils.py file saying that NoFileFoundError for "./vgg16_pretrained.pth".
I looked for the file in your repository but was not able to find it. Can you please share the link for the file?
Thank you in advance.

Update README if possible

Please allow me expressing my gratitude for this beautiful re-implementation.

I suggest the author updates the README on section 1.1 about the parameter setting since I find 'perceptual_param' and 'gan_param' have been replaced with new names 'lambda_perceptual' and 'lambda_gan'.

Besides, may I ask why the 'lambda_gan' parameter has been setting that low(0.01) and only applying on generator loss instead of on both generator and discriminator loss? I am assuming it will make discriminator have little impact to guide generator training. The default value '1' shown in 'train.py' seems a more reasonable value for me. Exhibiting the exact parameters of training current model is very crucial for others to understand the system.

I will be more than appreciate if you can answer my question.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.