-
Notifications
You must be signed in to change notification settings - Fork 800
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to understand the following implementation of loss in theano? #88
Comments
I don't understand your question. Do you mean it should be 1/N * ... instead of 1/NT * ... ? |
I think so. Not sure |
Generally you want your loss / gradient updates to be scale invariant. Also it does not really matter if you use rmsprop / adam etc. since they automatically rescale your gradients. |
jonashen
pushed a commit
to jonashen/rllab
that referenced
this issue
May 29, 2018
Add a customized tensor scalar to tensorboard by using the custom_scalar plugin in tensorboard. Each line in the scalar corresponds to an element in the tensor. Wrap the tensorboard logging module into a new class `Summary` in file rllab/misc/tensor_summary.py. It supports both the simple value and tensor logging. It also saves the computation graph created by rllab. To record the tensor into tensorboard, use the `record_tensor` function in file rllab/misc/logger.py. Refer to: rll#39, rll#38
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I don't see how the stochasticity of the actions 1/T*... is implemented in the theano line above. Isn't log_ likelihood_sym only computing one distribution? and not T of them and taking the average.
The text was updated successfully, but these errors were encountered: