Comments (10)
@salvadog yea, i'll get the upsampler finished by mid this month, and we can figure out if it is just scale, or something else
from gigagan-pytorch.
@lucidrains Certainly, thank you! In my current work focused on image super resolution or restoration, the top priority is to preserve the faithfulness of the input data rather than solely pursuing visually pleasing outcomes. Although cGAN-based techniques are widely used, their capabilities in terms of generating high-quality images are still somewhat restricted. Considering that diffusion-based methods inevitably modify the input data to some extent, it would be highly advantageous to introduce an upsampler similar to gigagan that can mitigate this issue.
from gigagan-pytorch.
@salvadog i'm not sure... maybe it really is just stabilizing the attention via the l2 distance and then stack-more-layers-meme.jpeg
i honestly didn't even think i would ever have to build a GAN again after the DDPM revolution happened
from gigagan-pytorch.
@salvadog imo the multi-scale discrimination contributes the most, it somehow teaches the model how to do super resolution. And i feel it provides more supervison than "progressive growing", because the model would always receive gradients from all scales.
from gigagan-pytorch.
@salvadog we can leave this issue open and maybe someone who has been following GAN literature can chime in
from gigagan-pytorch.
"The power of GANs in super-resolution lies in their ability to navigate the complex and nuanced landscape of high-resolution images, weaving together the delicate threads of detail and structure to create a tapestry of realistic and stunning resolution."
from gigagan-pytorch.
"The power of GANs in super-resolution lies in their ability to navigate the complex and nuanced landscape of high-resolution images, weaving together the delicate threads of detail and structure to create a tapestry of realistic and stunning resolution."
Interesting. Where does this sentence come from?
from gigagan-pytorch.
moved
from gigagan-pytorch.
Hello, am I not totally lost in here thinking this "multi-scale discriminator" result in the model that could perhaps be considered "scale equivariant"?
I mean sure, in the SG3 it's actually the generator that is equivariant (rotation, translation).
However, in the limitations, discussion, & future directions section of the paper authors state:
"In this work we modified only the generator, but it seems likely that further benefits would be available
by making the discriminator equivariant as well."
from gigagan-pytorch.
@salvadog imo the multi-scale discrimination contributes the most, it somehow teaches the model how to do super resolution. And i feel it provides more supervison than "progressive growing", because the model would always receive gradients from all scales.
yes, I agree with this. perhaps the unet architecture for generator too
from gigagan-pytorch.
Related Issues (20)
- Possible Discrepancies HOT 3
- The training code not deal with paired data yet? HOT 2
- [Question] About the upscaler HOT 2
- Multi GPU training HOT 4
- Multi GPU with gradient accumulation
- [Request] Please provide a replicate.com version
- Confused about this project?
- NaN losses after hours of training (UPSAMPLER) HOT 16
- How to implement this model to enhance my input images? Do I have to train the model to use? HOT 2
- Weights of Gigagan Upscaler HOT 1
- Turn on/off gradients computation between generator/discriminator HOT 2
- Wrong order of resolutions list HOT 1
- to_rgb branch has only 1 learnable kernel HOT 7
- Gradient Penalty is very high in the start HOT 10
- How to use this model for SR ?
- Has Anyone Trained This Model Yet? HOT 2
- The text-to-image tasks
- Config to reproduce paper
- question about code in unet_upsampler.py HOT 1
- the loss became nan after a few train steps HOT 2
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from gigagan-pytorch.