Giter Club home page Giter Club logo

depthfirstlearning.com's Introduction

Depth First Learning

This is the open source code repository running Depth First Learning, powered by Jekyll and the Tale theme.

Start this locally with bundle exec jekyll serve.

Contributing

We're looking for people to help make DFL better. Here are some ways you can help:

  • Run a study group based on one of our guides, and let us know about it. What worked well? What didn't?
  • Check out the open issues, where we keep track of improvements that we'd like to see.
  • Tell us what our guides are missing, or where they may be confusing, by filing issues.
  • Improve our guides, by filing pull requests. We are especially looking for Colabs with lots of explanatory text to explain and reproduce results. All of the content is stored in markdown under the posts/ directory.
  • Add guides for other papers! Reach out to us at [email protected] and let's talk.

depthfirstlearning.com's People

Contributors

aslanides avatar avital avatar bhairavmehta95 avatar chisness avatar cinjon avatar dependabot[bot] avatar dniku avatar hemildesai avatar jamesallingham avatar kumarkrishna avatar learnaifromscratch avatar luca-venturi avatar maxisawesome avatar micpie avatar p-patil avatar rileyedmunds avatar skroon avatar sschoenholz avatar suryabhupa avatar wgierke avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

depthfirstlearning.com's Issues

Additional materials/posts related to some of the common papers.

This looks like a great initiative, and it's something that was much needed.

But if you could also add the explanation/resources of some of the 'elementary' research papers (and by elementary I do not mean simple - rather the basic/foundation), like papers introducing CNNs or RNNs, or some regularization technique like dropout, it would be really helpful.

The reason for this is that if you have to learn about GAN (say a specific use-case of GAN like neural-style transfer), it's necessary that you should have the required knowledge regarding CNNs.

Although I know that explaining each and every paper out there in the domain of ML is not plausible, but maybe this could be done for some of the foundational papers.

Error in Problem Set 1 2e?

I think there may be an error in Problem Set 1 2e:

image

Specifically, x = phi(h), whereas the integrand term has phi(x). If this is a mistake, the problem is carried forward to

image

Add curriculum for Neural Mechanics paper

I'd like to add a curriculum for the recent paper Neural mechanics: symmetry and broken conservation laws in deep learning dynamics. There is a lot of depth to this paper, and the payoff (being able to calculate exact training dynamics of certain parameter combinations) makes that worth studying. Like the Resurrecting the Sigmoid curriculum, I'd probably structure it to be mostly based around problem sets, except with more interactive (coding/jupyter notebook) problems. If this sounds good, I can get started.

Add more solutions?

Thanks for the exceptional depthfirstlearning material!

I am currently working through the InfoGAN material and put together a small repo with my solutions.
If the solutions are of interest for the InfoGAN page I could add them to your repo.

I would be happy contribute to this great project.

Kind regards
Michael

InfoGAN: Looking for good explanation for relationship between JS divergence, Jensen's inequality and Shannon Entropy

Why is Jensen-Shannon divergence called this way?

The answer is something like this: https://dit.readthedocs.io/en/latest/measures/divergences/jensen_shannon_divergence.html#derivation (where "x" is a convex combination of P and Q), and the expectation is taken over the binary variable defining which of {P,Q} to take

We'd like a clear write-up of this definitely of the JS divergence, including a proof of the equivalence between this definition and the others.

New paper suggestion for study: "A correspondence between Random neural networks and statistical field theory"

Hey,

as I am strong supporter of "Depth First Learning" and am very intrigued by the material, I wanted
to suggest to include a curriculum on "A correspondence between Random neural networks and statistical field theory" (Pennington, Schoenholz, 2017).

Even though there are some more recent developments on statistical field theory for neural networks (e.g. Helias, Dahmen, 2020 or Grosvenor, Jefferson, 2021) I think, that the paper of Pennington and Schoenholz provides rich opportunities to create a curriculum which covers not only NN and statistical field theory but also some basic introduction in to replica methods.

As the publication is in the realm of materials which I study for my PhD, I could design a curriculum.
Greetings
Javed (Lindner), (PhD student, RWTH Aachen University)

Vanishing gradients of GANs?

I am wrapping my head around the explanation of the vanishing gradients problem of GANs for quite some time:

The current solution pdf document plots the input of the sigma functions over the loss function values to explain the (non-)saturating behavior. However, I am asking myself if that plot just captures the saturation of the sigma function and not the saturation behavior of the G loss function itself.

The NIPS 2016 GAN tutorial shows in figure 16 (p.26) an explanation of the saturating loss without taking the sigma function into account. With this explanation, I guess, the saturation behavior is explained through the gradients for G when G is not (yet) able to generate good fakes and D can easily identify them as fake (x = 0 or close to 0).
See a plot of the saturating and non-saturating loss function and their derivations. There, the saturating loss has a small gradient of around -1 and the saturating loss of -infinity at x = 0.
When I plot the gradients over the training for both loss functions I also get higher gradient means and higher standard deviations for the non-saturating loss when compared to the saturating loss (see notebook).

Maybe I am missing something?

I would be happy if somebody could point me in the right direction.

Typos in "About" page

http://www.depthfirstlearning.com/about/
Typo 1:
Second paragraph, third sentence, extraneous "is".
"We spent some time understanding each paper and writing down the core concepts on which they were is built."

Typo 2:
Fourth paragraph, fourth sentence, duplicate "out".
"Check out out github for examples and open an issue detailing which paper you want to study."

Broken links in Stein Variational Gradient Descent

Hey all!

I've just been taking a look at the SVGD curriculum and unfortunately, there are a number of broken links. Hopefully, some of these can be fixed without too much trouble! Here is a non-exhaustive list:

  1. https://www.win.tue.nl/~rvhassel/Onderwijs/Old-Onderwijs/2DE08-1011/ConTeXt-OWN-FA-201209-Bib/Literature/sigma-algebra/gc_06_measure_theory.pdf
  2. https://www.ims.nus.edu.sg/Programs/stein09/files/A%20Gentle%20Introduction%201.pdf
  3. https://www.cs.dartmouth.edu/~qliu/PDF/ksd_short.pdf
  4. https://www.cs.dartmouth.edu/~qliu/PDF/slides_ksd_icml2015.pdf
  5. https://www.cs.dartmouth.edu/~qliu/PDF/steinslides16.pdf

I found (I think) an alternative source for 1) here: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.124.3854&rep=rep1&type=pdf, but the rest I didn't manage. There could also be more missing links.

Cheers,

James

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.