Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fewer steps per second as training progresses #3

Open
TyJK opened this issue Dec 11, 2017 · 7 comments
Open

Fewer steps per second as training progresses #3

TyJK opened this issue Dec 11, 2017 · 7 comments

Comments

@TyJK
Copy link

TyJK commented Dec 11, 2017

Apologies if there is an obvious answer, but from the readme I gathered that when running properly, the steps per second should remain constant throughout training. Running on a GTX 970, I started out with ~90 steps per second and 25% GPU utilization. After leaving it to run overnight, I've found it's only run for 6 epochs and has slowed to about 46 steps per second, with about 15% GPU utilization. Everything runs perfectly otherwise, it takes several hours for the issue to appear, and restarting brings it back up to a normal rate. Is there a known cause/solution for this?

Thank you

@ShibiHe
Copy link
Owner

ShibiHe commented Dec 11, 2017

It is probably because when the epsilon-greedy is less likely to be greedy, the inference takes more time. I have also noticed different game behaves really different. Some games' speeds do not degenerate and others do.

@TyJK
Copy link
Author

TyJK commented Dec 12, 2017

Thanks for the quick response. This isn't an issue so I apologize, but I was wondering how you gathered the average score per episode, since the only output are the 3 csv files. Is it just an average of the episode reward over the number of episodes?

@ShibiHe
Copy link
Owner

ShibiHe commented Dec 12, 2017

I think I run a few random seeds and compute an average.

@TyJK
Copy link
Author

TyJK commented Dec 12, 2017

Alright, thanks for the confirmation 👍

@TyJK
Copy link
Author

TyJK commented Dec 12, 2017

Sorry (I'm doing this for a reproducibility project based partially on your code), but I just wanted to confirm that the --save-pkl parameter allows you to save the weights of the NN. And finally, how would I get the raw scores for the 30-op evaluation method once it's trained?

Thank you once again. I really appreciate it.

@ShibiHe
Copy link
Owner

ShibiHe commented Dec 12, 2017

I remember I wrote another code specially for doing 30 no-op test. This code base is quite old, so I do not remember the details. My new implementation was using tensorflow to save and load the model.

@TyJK
Copy link
Author

TyJK commented Dec 12, 2017

Alright, thank you very much :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants