--
You received this message because you are subscribed to the Google Groups "open-nars" group.
To unsubscribe from this group and stop receiving emails from it, send an email to open-nars+...@googlegroups.com.
To view this discussion visit https://groups.google.com/d/msgid/open-nars/a3b40440-b6f5-4a1a-ab85-3bd828b41cabn%40googlegroups.com.
--
To view this discussion visit https://groups.google.com/d/msgid/open-nars/e072e899-fee7-4a9a-89a0-33ee62c2a873n%40googlegroups.com.
To view this discussion visit https://groups.google.com/d/msgid/open-nars/b20ac7aa-58f2-4767-bc1d-e4014dc1871an%40googlegroups.com.
--
You received this message because you are subscribed to the Google Groups "open-nars" group.
To unsubscribe from this group and stop receiving emails from it, send an email to open-nars+...@googlegroups.com.
To view this discussion visit https://groups.google.com/d/msgid/open-nars/CACSdSzrJx2w9A-KWf5zdyLvH7BbFx4TEzEgtif%3D40LdfoAoY5A%40mail.gmail.com.
Hello Christopher,
Thanks for interacting in this thread, I like observing the various points of views.
Personally I make a clear distinction between what I call "standard" probability as formally defined by Kolomogorov and what is called a probability distribution in neural nets, usually a categorical distributions made out of softmax or equivalent but based on axioms from information theory. My article is clearly about the former, not specifically the later. However, softmax for instance might have the same kind of issues since the set of categories is not supposed to evolve during training either. In my article I try to analyze the situation with an as unbiased perspective as possible and that's why I will not take perceptron as the foundation of my analysis. Not to deny all the great things that NN has brought to the table, it seems to me that this concept was forced by the observations we have made of biological neurons, but it's not a guarantee that it would be the best solution eventually and therefore I rather try to approach the problem from first principles and necessary conditions instead.
Having said that, the important bit that I'm most interested in and that is shared by both frameworks is the concept of amount of information gathered after an observation (also called surprise or entropy). This is what I focus on atm. Also note that there is a third interpretation of probabilities that I also consider very interesting, which is the perspective taken in quantum physics which sees the wave function as representing a probability of observing a given state of the particle in a given configuration.
Regards,
Clément