Comments (10)
We haven't yet made a setting for it, but the other integral item with VRAM usage while training is the batch size. ATM, batch size is 128, and it could be 64 or even 32 and still produce decent results. Maybe we should also expose batch_size in settings.py.
from nmt-chatbot.
Did you change anything in settings? If yes, did you either uncomment 'override_loaded_hparams' or remove everything inside 'model' folder? hparams section from settings is saved with model with first train.py run.
from nmt-chatbot.
Everytime I change my settings I remove everything inside 'model' folder and re-run prepare_data.py ( if changed vocab_size )
from nmt-chatbot.
Okay I see that num_units is directly proportional to that, so I can reduce that number somehow and fit it. Should I be doing more epochs then If I go with a very low num_units value?
from nmt-chatbot.
One more question at the moment - did you clone that repo at it's current state or are you using your fork made at some point? There were a bug associated with training set creation.
from nmt-chatbot.
With 4GB of VRAM you should probably use defaults. I was training models using my GTX 970 with 4GB of VRAM with that settings (but with slightly bigger vocab of size of 17k tokens).
from nmt-chatbot.
I cloned it a few days back... 3 Dec I guess. The problem with defaults is that it won't work on my data set but works on sample data set. Does size of each line matter as well? for example, If all comments are less than , let's say, 1000 characters, would it help reduce how much memory is required?
How many training samples did you use in your data set?
from nmt-chatbot.
I fixed that bug in Decemeber 3rd, so please do a fresh clone and try again.
from nmt-chatbot.
@daniel-kukiela I'll do that and try again. Thank you :)
@Sentdex Thank you for the tip, I'll try using different batch_size(s). BTW, your videos are awesome :)
from nmt-chatbot.
@daniel-kukiela @Sentdex I tried both of the things at the same time and lowering batch_size works, also that about that bug, I no longer need to remove empty lines from vocab manually. Thank you for help.
Also, I've made a pull request #9 for adding batch_size in settings.py
from nmt-chatbot.
Related Issues (20)
- Chatbot only produces 1 answer despite beam width being 20 HOT 4
- Can't run train.py HOT 7
- Train.py says it's finished training after 1 second. HOT 2
- ERROR: cannot import name 'nmt' from 'nmt' HOT 2
- FileNotFoundError: [Errno 2] No such file or directory: 'data//corpus_size'
- nmt: [ deleted ] while training
- Trained Model ? HOT 1
- does this work with tensorflow 2.0?
- is it supposed to just be stopped on this
- 2021-05-06 22:03:17.947802: I tensorflow/core/kernels/shuffle_dataset_op.cc:121] Shuffle buffer filled.
- decoding to output F:\Nebula0.0.5\nmt-chatbot\model\output_dev.
- from nmt import nmt : error
- TypeError: 'encoding' is an invalid keyword argument for this function HOT 1
- Inference returning the same empty answer to everything i type HOT 1
- bot producing only one translation per input
- Cannot Downgrade Tensorflow 1.x on Google Colab
- how do i make it train more epoch like 10000
- cannot import name 'nmt' from 'nmt'
- Issue with tensorflow api and pip installation candidate missing before 2.x
- cannot import name 'lookup_ops' from 'tensorflow.python.ops'
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from nmt-chatbot.