Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training on GPU? #53

Open
jesseos opened this issue Feb 9, 2019 · 4 comments
Open

Training on GPU? #53

jesseos opened this issue Feb 9, 2019 · 4 comments

Comments

@jesseos
Copy link

jesseos commented Feb 9, 2019

What changes do I have to make to train.py to train using my GTX-1080 (I have tensorflow-gpu installed and recognizing it)?

@pender
Copy link
Owner

pender commented Feb 17, 2019

You'll probably need to scale down the size of the net or the batch size or sequence length to accommodate that the GTX-1080 has only 8gb of VRAM instead of the 11gb of the 1080Ti that I trained on. I'd probably recommend scaling down the size of the net, for example python3 train.py --block_size 1024.

@claudetheboof
Copy link

I am trying to train the Chatbot on a , Core I5-3570, 12 gb RAM, GTX 1060 6gb any idea what I need to change to makes it run effectively? I have tried change the block_size but still somehow still training on my CPU , if I can train using both CPU + GPU what parameters should I add to it?

@claudetheboof
Copy link

nvm i found a fix using Pip install --ignore-installed --upgrade tensorflow-gpu

@ghost
Copy link

ghost commented Apr 12, 2019

FWIW: I'm training a bot instance with a 100MB input-set with a GeForce GTX 1050 Ti and this succeeds if I use --num_blocks 1 for a train.py parameter.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants