YKY, you may some valuable insight in Michael Leyton's "A Generative Theory of Shape" <http://www.springer.com/computer/image+processing/book/978-3-540-42717-9>.
Semantic vectors sort-of-ish work because the mathematical structure of the tensor product, and the structure of grammar are both described by the same underlying device: the so-called "non-symmetric compact closed monoidal category". The difference is that tensors are also symmetric, and so forcing this symmetry then forces a kind-of straight-jacket onto the language.
References:see also work by Bob CoeckeFWIW, I believe that dependency grammars, and link-grammar in particular, are isomorphic to categorical grammars. Its almost obvious if you stare at the above wikipedia article long enough: the expressions are just link-grammar links. The categorical grammar notation is rather unwieldy, that's the big difference.
On Thu, Jun 19, 2014 at 5:29 AM, Linas Vepstas <linasv...@gmail.com> wrote:
Semantic vectors sort-of-ish work because the mathematical structure of the tensor product, and the structure of grammar are both described by the same underlying device: the so-called "non-symmetric compact closed monoidal category". The difference is that tensors are also symmetric, and so forcing this symmetry then forces a kind-of straight-jacket onto the language.According to the Wikipedia page, symmetric means that the dual is replaced by left and right adjoints. I can only vaguely understand this...Since the tensor product is non-commutative, I was thinking it suffices to represent word-order differences, such as
John loves Mary != Mary loves John.Perhaps the non-symmetric category is a better model for sentences with word order differences, than merely using a (tensor) product that is non-commutative?
References:
see also work by Bob CoeckeFWIW, I believe that dependency grammars, and link-grammar in particular, are isomorphic to categorical grammars. Its almost obvious if you stare at the above wikipedia article long enough: the expressions are just link-grammar links. The categorical grammar notation is rather unwieldy, that's the big difference.I don't see a clear correspondence between link grammar and Lambek (categorical grammar). What do the "links" correspond to, in the categorical formation?
By the way, the book "Quantum physics and linguistics" seems very good for this type of things:
http://ukcatalogue.oup.com/product/9780199646296.do
Perhaps the non-symmetric category is a better model for sentences with word order differences, than merely using a (tensor) product that is non-commutative?yes exactly.
On Fri, Jun 20, 2014 at 6:08 AM, Linas Vepstas <linasv...@gmail.com> wrote:
Perhaps the non-symmetric category is a better model for sentences with word order differences, than merely using a (tensor) product that is non-commutative?yes exactly.But what exactly is the advantage of the categorical approach regarding capturing word order, over a non-commutative (say tensor) product?
The way I see it, the unrestricted tensor product has a problem: there is no guarantee that the location of "John loves Mary" wouldn't clash with another unrelated product such as "Peter ate pizza", simply by accident. That is clearly bad as a model of semantic space.
The remedy is to enforce a distance metric among word products. This metric can be defined as the graph distance between nodes in the Cayley graph of the free group generated by the words. This metric makes semantic sense. Any other embedding that do not severely distort this metric would also be OK.
I don't know how the categorical approach deals with this problem. Perhaps it doesn't, because the problem has to do with metric space. Perhaps the categorical approach makes different sentences "mutually exclusive", as you mentioned?