--
You received this message because you are subscribed to the Google Groups "lasagne-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to lasagne-user...@googlegroups.com.
To post to this group, send email to lasagn...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/lasagne-users/a9117aa8-7ce4-40a3-b7c0-40b642e45ba9%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
To view this discussion on the web visit https://groups.google.com/d/msgid/lasagne-users/7616aaf1-918e-400c-bd41-2ccff2b49bea%40googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/lasagne-users/05f91206-17bc-4989-87d3-05412cc188c2%40googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/lasagne-users/ad3ddd3f-da41-4bc4-b754-8b4682c1fe82%40googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/lasagne-users/A2119C99-91E6-424F-9C57-4CBFDC4EFCAF%40gmail.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/lasagne-users/CAMxMOAuB7BnDnLseFtqamoDB_A%3DbsYC-mGTOaywr29mCwZZ0-A%40mail.gmail.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/lasagne-users/2711c83b-a4e7-44ee-81e4-715d1e4c82b1%40googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/lasagne-users/ab83d8c6-8d17-48b7-b461-194afe5e0eee%40googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/lasagne-users/25f5110a-32ff-4011-bc90-dc7f33ab52a9%40googlegroups.com.
I think my theory makes sense because the last DenseLayer for localization network, see below, doesn't have any nonlinear function really (the intention should be to initialize an identity matrix I assume), so large pixel value could easily cause explosion problem.
2. I've checked with output for 'loc_out' layer,[[ 9.37732399e-01 -4.42903414e-02 -7.92445168e-02 -1.68894301e-04, 9.90774453e-01 -3.39208022e-02]]It's still very much the identity matrix we initialized, so I think the localization matrix didn't learn much how to find the object. Not sure about how to deal with problem yet.
I'm not too familiar with this, but it seems to me that getting a good localization network will be comparable in difficulty to training a good classifier.
1. I've tried with reduce the learning rate for localization network, but basically, the output parameters eventually go to vary large number.2. Try to increase learning rate for localization network, but it's not helping, the parameters are still roughly the identity matrix.
3. Will think about maybe using more complex localization network or insert it in the middle.
1) Explosion problem when training for localization parameters. The localization network I copied from Lasagne/Recipe doesn't have any non-linear function in between, which makes it's linear between conv layers. You can image the pixel value remain uint8 passing through layers in localization network, then I think it's possible that even using with very low learning rate (I've tried with 10-6), I still get large parameter values from localization network. A direct result with these large strange parameters, the transformer layer produce similar images with different input, within a narrow pixel value range.
What do you mean by "doesn't have non-linear function in between" ? By default each Conv2DLayer is using a relu
And the weird thing is the transformed images have uniform values, for example, red channel pixels are all the same values, ditto for the other color channels.