Comments (3)
It would be nice if could also see which weights were used to generate each self-play game and feature. So I might want to add the generation field (can be null) to the database and extend the REST protocol to allow for some way to include the generation in there (maybe as a header?), we also need some way to retrieve the latest generation using the REST-ful API.
from dream-go.
How many features do we need to include during training. More is better, but being realistic, where does the diminishing returns start kicking in?
- DeepMind includes the 500,000 most recent games (but extract multiple features from each game?)
- Thinking Fast and Slow with Deep Learning and Tree Search use DAGGER which trains on the entire dataset.
The network that DeepMind were training is roughly twice as big as ours is, so we can probably get away with half of their number of games. But we could also follow the DAGGER approach and train on everything.
I have some concern with training on everything since some early games can be very bad and imitating them will not get us anything good. So we will probably go with the DeepMind approach and train on the 250,000 most recent games. This is a nice number also because that is about the number of games we can produce in a day using two GPU's.
from dream-go.
I checked in the mentioned images, see the the docker directory for all the details.
from dream-go.
Related Issues (20)
- Re-balance search tree size vs neural network size HOT 2
- Scoring and `kgs-genmove_cleanup` improvements
- About MCTSnet HOT 2
- Introduce a new self-play mode
- Poor GPU utilization observed during play HOT 2
- Re-factor MCTS code to use asynchronous framework
- Shape of the convolution in the policy head
- Monte-Carlo tree search as regularized policy optimization HOT 3
- Investigate MCTS parallelism degradation HOT 7
- Prune nodes from the search tree that are obviously bad HOT 1
- Re-implement `INT8x32_CONFIG` support during inference
- Investigate SWISH as activation function in cuDNN
- GPU vs CPU matrix multiplication HOT 1
- Sparse Quantized Model
- MLP-Mixer: An all-MLP Architecture for Vision HOT 7
- NNUE (ƎUИИ Efficiently Updatable Neural Network) for Go HOT 5
- Triton: Open-Source GPU Programming for Neural Networks
- Long startup times due to `cudnnBuildRNNDynamic`
- 2022 TCGA Computer Go Tournament is coming! HOT 1
- Unsound uninitialized array
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from dream-go.