Hey there,
I've got some questions concerning the DQN demo.
What does the average q-learning loss value actually mean? Is this the error term resulting from the training of the neural network in the DQN demo? I'm applying that approach to one of my games and value is usually greater than two. If this is related to the error property of training the neural network, then I'd assume that I have conflicting training data.
And another question:
Concerning calling the train function:
var loss = this.tdtrainer.train(x, ystruct);
Is this completly related to the neural network implementation? Like, could I just go ahead and just go on from that point with another neural net implementation?
Thanks in advance!
(I'm trying to make use of DQN for my game
BRO)