Comments (10)
Could it be that this shows an underlying tension between having two different things: "examples" that should be read to learn, and a "model/sota zoo" that has enough so that someone can jump-start towards current SOTA?
Because regarding the two models: WRN is super widely used, but PyramidNet with Shake-shake is (close to) SOTA. If I want to learn about the framework, I'm happy reading WRN without many tricks, but if I want to see if my newest invention increases SOTA, I'd like to get a small code which gets close to SOTA where I can add my invention and get great numbers (or not).
from flax.
That link is now a 404. Where is the current CIFAR10 example code?
Do you still have a cifar10 example?
from flax.
Thanks for adding this issue Alexey, I agree with your observations.
Can we just keep one or two of the above combinations?
Which ones are more relevant / useful / educational?
Fine by me, I don't know much about the details of these architectures, but I would probably pick the one that is most common. I am not sure which one is most educational.
Should we turn some other combinations into HOW-TOs? Which ones?
HOWTOs are based on diffs, so they are most useful if they are being used for some added functionality. It seems that WideResNet is implemented without shake-shake and with shake-shake. Is it possible to create a HOWTO that shows how to add shake-shake to WideResNet?
from flax.
Good point Lucas! I think we are still figuring out which examples belong in Flax exactly (I agree with the tension), but for this specific example I probably lean towards keeping the simpler one (WRN), with possibly one or more HOWTOs, and having the most complex ones in a separate repository.
from flax.
FYI I chatted about this with Alexey, and he told me that he'd like to wait with finishing the regression testing until he will act on this issue. So I've linked that PR here.
from flax.
We have recently removed out official CIFAR example and instead linked to a much better open source repository (https://github.com/google-research/google-research/tree/master/flax_models/cifar)
from flax.
That link is now a 404. Where is the current CIFAR10 example code?
from flax.
We removed the example because it was using the old API and we had nobody actively working on it and willing to port it.
You can see more examples in our examples README:
https://github.com/google/flax/tree/main/examples
Note that for example the vision_transformer
codebase has example code to fine-tune a model on CIFAR10:
- https://github.com/google-research/vision_transformer
- https://colab.research.google.com/github/google-research/vision_transformer/blob/master/vit_jax.ipynb
from flax.
@andsteing Is there any example code showing a simple resnet running on CIFAR10 that you know of?
from flax.
You could switch the imagenet2012
dataset with cifar10
in our examples/imagenet
- or you could download the pre-trained imagenet checkpoint and fine-tune it on cifar10
I'm not aware of a repo that simply trains a resnet
on cifar10
.
from flax.
Related Issues (20)
- Error when calling `Module.tabulate` on normalization wrappers like `WeightNorm` and `SpectralNorm`
- Orbax checkpoint for LogicallyPartitioned params HOT 2
- For some reason these imports are elided on read the docs HOT 1
- Using variable declared at a broader scope in a function is bad form HOT 1
- Add `BatchRenorm` layer to `linen.normalization`
- GroupedConv distributed training failure
- In `MultiHeadAttention`, let `num_heads=1` by default
- Documentation/notebook errors HOT 2
- Remove `tree_map` deprecation filter after Flax upgrades minimum Python version to 3.10
- Unpickled modules with constructor arguments cannot be initialized
- Improve SEO for docs pages HOT 2
- Add ability to easily change documentation version
- Problem while using checkpoints.restore_checkpoint with gradio HOT 1
- nnx static fields not part of static tree structure HOT 1
- nn.remat_scan doesn't work with nn.with_partitioning HOT 1
- No way to call nnx.State.from_flat_path HOT 5
- Tutorial request HOT 2
- with_partitioning has surprising behavior with MultiHeadAttention and DenseGeneral HOT 1
- nnx.graph.split infinite recursion when used in a thread HOT 3
- Documentation links 404 HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from flax.