You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Why you use softmax_cross_entropy_with_logits here, the first state is "[10.0, 128.0, 1.0, 1.0]*args.max_layers",so does the labels. The final output of RNN contributes to the action, why you use softmax on the action?
The text was updated successfully, but these errors were encountered:
for example:
state=[10.0, 128.0, 1.0, 1.0,10.0, 128.0, 1.0, 1.0], the final output with softmax is [0.1,0.1,0.1,0.1,0.1,0.1,0.2,0.2], then the loss is:
-(10log0.1+128log0.1+1log0.1+1log0.1+10log0.1+128log0.1+1log0.2+1log0.2)
what does this mean?
self.cross_entropy_loss = tf.nn.softmax_cross_entropy_with_logits(logits=self.logprobs[:, -1, :], labels=self.states)
Why you use softmax_cross_entropy_with_logits here, the first state is "[10.0, 128.0, 1.0, 1.0]*args.max_layers",so does the labels. The final output of RNN contributes to the action, why you use softmax on the action?
The text was updated successfully, but these errors were encountered: