--
You received this message because you are subscribed to the Google Groups "gps-help" group.
To unsubscribe from this group and stop receiving emails from it, send an email to gps-help+unsubscribe@googlegroups.com.
To post to this group, send email to gps-...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/gps-help/054eabbd-ffad-4059-9a97-911d123d6722%40googlegroups.com.
Thank you for your replay!"changing the target is the same problem as changing the initial position." But, if I change the target point, I have to tell the policy network where the target is, so I make the target point and current state as inputs of policy network. In changing initial position problem, make the current state as inputs of policy network is enough. Am I right? Is it a good idea to make the target point as inputs of policy network and let the aircraft fellow the target point inputs?
在 2016年9月6日星期二 UTC+8下午11:32:53,Chelsea Finn写道:
Hi,I don't currently have the bandwidth to help figure out your problem, other than a few recommendations:- If the cost is decreasing appropriately but the policy isn't improving, then it is likely that something is wrong with the objective, e.g. the cost isn't defined correctly- If this cost *isn't* decreasing, then there is probably either an issue with the analytic derivatives of the cost or how the state is representedNote that, for the pointmass example, changing the position of the target relative to the initial position is the same as moving the initial position relative to the target (except in the reverse direction). So, algorithmically, changing the target is the same problem as changing the initial position.Chelsea
On Mon, Sep 5, 2016 at 8:36 PM, <cuik...@163.com> wrote:
ps: I rewite the box2d pointmass example using badmm
在 2016年9月5日星期一 UTC+8下午11:27:03,cuik...@163.com写道:I noticed that the moving aircraft can reach a fixed point from any initial state in box2d pointmass example. When I alter the target point, the training process has to be carried out again. So, how can I train the moving aircraft to reach any given points? I try to rewrite code to add mass target points and make the target point as inputs of policy network. Unfortunately, the result was bad. In the paper End-to-end Training of Deep Visuomotor Policies, the agent can adapt different target object position. How to do this. Thanks.
--
You received this message because you are subscribed to the Google Groups "gps-help" group.
To unsubscribe from this group and stop receiving emails from it, send an email to gps-help+u...@googlegroups.com.
To post to this group, send email to gps-...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/gps-help/054eabbd-ffad-4059-9a97-911d123d6722%40googlegroups.com.
--
You received this message because you are subscribed to the Google Groups "gps-help" group.
To unsubscribe from this group and stop receiving emails from it, send an email to gps-help+unsubscribe@googlegroups.com.
To post to this group, send email to gps-...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/gps-help/b57d30ba-4ce8-4096-bcff-7ee237dd008a%40googlegroups.com.