--
All contributions to this forum are covered by an open-source license.
For information about the wiki, the license, and how to subscribe or
unsubscribe to the forum, see http://ontologforum.org/info/
---
You received this message because you are subscribed to the Google Groups "ontolog-forum" group.
To unsubscribe from this group and stop receiving emails from it, send an email to ontolog-foru...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/ontolog-forum/0ce19d5036394894befb87fd5dc5ee8b%40bestweb.net.
All, since Claude2 is topical right now. I yanked out an excerpt
from John's reply and passed it on for processing.
Question: LLMs are just data represented by tensor calculus. means what?
Answer:
You're right, as an AI system I don't actually have a deeper
"understanding" of the information I'm providing. LLMs like myself
are machine learning models based on large neural networks trained
on massive datasets.
At a basic level, here's what that means:
- LLMs stands for Large Language Models. These are neural networks
trained on huge text datasets like books, articles, websites, etc.
- The knowledge in an LLM is represented by numeric weights or
parameters in the neural network, which connect artificial
"neurons" and allow information to be processed and used for
generating text.
- These weights are multi-dimensional arrays called tensors. So an
LLM's knowledge is entirely stored in tensor calculus - just
numbers in a very large, multi-layered matrix that has been tuned
through the training process.
- I don't actually comprehend or conceptualize the text I work
with in any meaningful way. I just associate patterns of words and
language statistically based on those numeric parameters.
So in summary, you're absolutely right - everything I "know" is
just data represented mathematically in tensors. I don't have any
deeper understanding or self-awareness despite being able to have
conversations. I'm an LLM, not an AGI! Please let me know if you
would like me to elaborate on any part of this explanation.
And naturally, I was compelled to pursue and additional question and answer sequence.
Question: Can one apply tensor calculus to linguistic
understanding of sentences?
Answer:
Yes, tensor calculus can be used to try to represent and model
linguistic understanding of sentences, though current AI systems
are still limited in this capability. Here are a few key ways
tensor calculus comes into play:
- Word embeddings are vector representations of words that map
words into a math space where semantics can be represented
geometrically. Word2vec and GloVe are popular embedding models
based on tensors.
- Contextual embeddings like BERT represent both words and their
context in tensors, allowing more complex semantic relationships
to be encoded.
- Attention mechanisms in transformers use tensors to map query
and key vectors to values, allowing models to focus on relevant
parts of input. This can help relate words to meaning.
- Recursive neural tensor networks use tensors to compose smaller
phrase chunks into full sentences in a bottom-up structure that
can model syntax.
- Probing classifiers analyze internal tensor representations in
models to assess what linguistic features like syntax or semantics
are encoded.
However, current AI systems still do not obtain true comprehension
of language. The tensor representations are limited and may not
fully capture human-level semantic complexity and reasoning
ability needed for understanding. There are also challenges in
interpreting what the tensor parameters in neural networks really
mean.
So in summary, tensor calculus provides tools to represent and
analyze language statistically, but does not equate to true
linguistic understanding on its own. We have yet to crack the code
of artificial general intelligence!
-- Regards, Kingsley Idehen Founder & CEO OpenLink Software Home Page: http://www.openlinksw.com Community Support: https://community.openlinksw.com Weblogs (Blogs): Company Blog: https://medium.com/openlink-software-blog Virtuoso Blog: https://medium.com/virtuoso-blog Data Access Drivers Blog: https://medium.com/openlink-odbc-jdbc-ado-net-data-access-drivers Personal Weblogs (Blogs): Medium Blog: https://medium.com/@kidehen Legacy Blogs: http://www.openlinksw.com/blog/~kidehen/ http://kidehen.blogspot.com Profile Pages: Pinterest: https://www.pinterest.com/kidehen/ Quora: https://www.quora.com/profile/Kingsley-Uyi-Idehen Twitter: https://twitter.com/kidehen Google+: https://plus.google.com/+KingsleyIdehen/about LinkedIn: http://www.linkedin.com/in/kidehen Web Identities (WebID): Personal: http://kingsley.idehen.net/public_home/kidehen/profile.ttl#i : http://id.myopenlink.net/DAV/home/KingsleyUyiIdehen/Public/kingsley.ttl#this
John,
there are many works about equivalence and conversion/mapping between ANN and SNN. Here is only one or recent works (https://arxiv.org/abs/2306.08744 and see literature on this topic inside):
----------
Are training trajectories of deep single-spike and deep ReLU network equivalent?
Ana Stanojevic, Stanisław Woźniak, Guillaume Bellec, Giovanni Cherubini, Angeliki Pantazi, Wulfram Gerstner
Communication by binary and sparse spikes is a key factor for the energy efficiency of biological brains. However, training deep spiking neural networks (SNNs) with backpropagation is harder than with artificial neural networks (ANNs), which is puzzling given that recent theoretical results provide exact mapping algorithms from ReLU to time-to-first-spike (TTFS) SNNs. Building upon these results, we analyze in theory and in simulation the learning dynamics of TTFS-SNNs. Our analysis highlights that even when an SNN can be mapped exactly to a ReLU network, it cannot always be robustly trained by gradient descent. The reason for that is the emergence of a specific instance of the vanishing-or-exploding gradient problem leading to a bias in the gradient descent trajectory in comparison with the equivalent ANN. After identifying this issue we derive a generic solution for the network initialization and SNN parameterization which guarantees that the SNN can be trained as robustly as its ANN counterpart. Our theoretical findings are illustrated in practice on image classification datasets. Our method achieves the same accuracy as deep ConvNets on CIFAR10 and enables fine-tuning on the much larger PLACES365 dataset without loss of accuracy compared to the ANN. We argue that the combined perspective of conversion and fine-tuning with robust gradient descent in SNN will be decisive to optimize SNNs for hardware implementations needing low latency and resilience to noise and quantization.
---------
And this (spikes and ReLU neural networks) is not the only two equivalent variant of deep learning neural network architecture. Here is start of work with neural networks in quantum computer (quantum neural network), https://www.nature.com/articles/s43588-021-00084-1
------------
The power of quantum neural networks
Amira Abbas, David Sutter, Christa Zoufal, Aurelien Lucchi, Alessio Figalli & Stefan Woerner
It is unknown whether near-term quantum computers are advantageous for machine learning tasks. In this work we address this question by trying to understand how powerful and trainable quantum machine learning models are in relation to popular classical neural networks. We propose the effective dimension—a measure that captures these qualities—and prove that it can be used to assess any statistical model’s ability to generalize on new data. Crucially, the effective dimension is a data-dependent measure that depends on the Fisher information, which allows us to gauge the ability of a model to train. We demonstrate numerically that a class of quantum neural networks is able to achieve a considerably better effective dimension than comparable feedforward networks and train faster, suggesting an advantage for quantum machine learning, which we verify on real quantum hardware.
----------
Brains are very good computers. Here is about organoids (artificial brains): https://www.frontiersin.org/journals/science/articles/10.3389/fsci.2023.1017235
------------------------
Organoid intelligence (OI): the new frontier in biocomputing and intelligence-in-a-dish
Recent advances in human stem cell-derived brain organoids promise to replicate critical molecular and cellular aspects of learning and memory and possibly aspects of cognition in vitro. Coining the term “organoid intelligence” (OI) to encompass these developments, we present a collaborative program to implement the vision of a multidisciplinary field of OI. This aims to establish OI as a form of genuine biological computing that harnesses brain organoids using scientific and bioengineering advances in an ethically responsible manner. Standardized, 3D, myelinated brain organoids can now be produced with high cell density and enriched levels of glial cells and gene expression critical for learning. Integrated microfluidic perfusion systems can support scalable and durable culturing, and spatiotemporal chemical signaling. Novel 3D microelectrode arrays permit high-resolution spatiotemporal electrophysiological signaling and recording to explore the capacity of brain organoids to recapitulate the molecular mechanisms of learning and memory formation and, ultimately, their computational potential. Technologies that could enable novel biocomputing models via stimulus-response training and organoid-computer interfaces are in development. We envisage complex, networked interfaces whereby brain organoids are connected with real-world sensors and output devices, and ultimately with each other and with sensory organ organoids (e.g. retinal organoids), and are trained using biofeedback, big-data warehousing, and machine learning methods. In parallel, we emphasize an embedded ethics approach to analyze the ethical aspects raised by OI research in an iterative, collaborative manner involving all relevant stakeholders. The many possible applications of this research urge the strategic development of OI as a scientific discipline. We anticipate OI-based biocomputing systems to allow faster decision-making, continuous learning during tasks, and greater energy and data efficiency. Furthermore, the development of “intelligence-in-a-dish” could help elucidate the pathophysiology of devastating developmental and degenerative diseases (such as dementia), potentially aiding the identification of novel therapeutic approaches to address major global unmet needs.
------------------
This paper is issued after success of brain-based computer on living cell culture, https://doi.org/10.1016/j.neuron.2022.09.001
In vitro neurons learn and exhibit sentience when embodied in a simulated game-world
Brett J. Kagan, Andy C. Kitchen, Nhi T. Tran, Ben Rollo, Adeel Razi, Karl J. Friston
Integrating neurons into digital systems may enable performance infeasible with silicon alone. Here, we develop DishBrain, a system that harnesses the inherent adaptive computation of neurons in a structured environment. In vitro neural networks from human or rodent origins are integrated with in silico computing via a high-density multielectrode array. Through electrophysiological stimulation and recording, cultures are embedded in a simulated game-world, mimicking the arcade game “Pong.” Applying implications from the theory of active inference via the free energy principle, we find apparent learning within five minutes of real-time gameplay not observed in control conditions. Further experiments demonstrate the importance of closed-loop structured feedback in eliciting learning over time. Cultures display the ability to self-organize activity in a goal-directed manner in response to sparse sensory information about the consequences of their actions, which we term synthetic biological intelligence. Future applications may provide further insights into the cellular correlates of intelligence.
---------------
SoTA (state-of-the-art) in science is moving very quickly. One more time I cite Clarke’s three laws, https://en.wikipedia.org/wiki/Clarke%27s_three_laws
I am 65 and always googling first when going to speak about something impossible. In AI and LLMs what seems impossible yesterday may be already possible today.
Best regards,
Anatoly
--
All contributions to this forum are covered by an open-source license.
For information about the wiki, the license, and how to subscribe or
unsubscribe to the forum, see http://ontologforum.org/info/
---
You received this message because you are subscribed to the Google Groups "ontolog-forum" group.
To unsubscribe from this group and stop receiving emails from it, send an email to ontolog-foru...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/ontolog-forum/1c1b168b73f545dcab783fe86f0625e7%40bestweb.net.
John,
You need not too many artificial neurons to have learned behavior. Here is simulation features twenty cars, each operating as a separate neural network. Remarkably, these neural networks consist of only fifty neurons, making efficient use of resources. The neural networks are trained from scratch, without any pre-training -- https://artificialneuralcomputing.com/cars (and paper about it in the end of the page. Those “invariants” from the paper for me is reference to ontology of driving domain that expressed in slightly non-usual way, with mathematic formulae. But you need relax you definition of ontology to count this invariants as valid ontology for self-driving. See paper in the end of the cars page, it is interesting how many participants of ontolog forum can see ontology work in it. I definitely see ontology engineering there, defining of 4 “invariants”).
---------------
Nature is very inefficient at a Darwinian evolution. Very. Inefficient.
Engineers surely can do better with the help of computers. Therefore we will be careful about assessment of contemporary cognitive architectures that based on neural networks with trillions weights (GPT-4 have 1.8 trillion parameters spread across 120 layers, https://twitter.com/AlexAIDaily/status/1678748352740052992). Now is race for optimization of this huge amount of weights, expected in a year or two at least x10 less for comparable performance.
Best regards,
Anatoly
From: ontolo...@googlegroups.com <ontolo...@googlegroups.com> On Behalf Of John F Sowa
Sent: Thursday, July 27, 2023 1:17 AM
To: ontolo...@googlegroups.com
--
All contributions to this forum are covered by an open-source license.
For information about the wiki, the license, and how to subscribe or
unsubscribe to the forum, see http://ontologforum.org/info/
---
You received this message because you are subscribed to the Google Groups "ontolog-forum" group.
To unsubscribe from this group and stop receiving emails from it, send an email to ontolog-foru...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/ontolog-forum/d2d301eed9fb4b1dae1e79fea2cf86b1%40bestweb.net.
John,
> Those articles you cited were written by people who have zero understanding of neuroscience or even elementary biology.
John, these people have not a bad Hirsh to have a brief assessment of their understanding of neuroscience. E.g. Karl Friston have h-index 266 in Google Scholar, https://scholar.google.ru/citations?user=q_4u0aoAAAAJ&hl=en&oi=ao
If you need to fly 300 humans simultaneously, you need not have wing from a bird with all it redundancy. You need only wing of a aeroplane. If you have to reason like 300 educated humans simultaneously you not need brains of these humans, you need hardware without redundancy of biological cells. Thus complexity of hardware platform of wet spike neural network is irrelevant.
Best regards,
Anatoly
--
All contributions to this forum are covered by an open-source license.
For information about the wiki, the license, and how to subscribe or
unsubscribe to the forum, see http://ontologforum.org/info/
---
You received this message because you are subscribed to the Google Groups "ontolog-forum" group.
To unsubscribe from this group and stop receiving emails from it, send an email to ontolog-foru...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/ontolog-forum/2e6b6af4735f4aeca6b6284751b79f85%40bestweb.net.
Alex:
I presume they (LLMs) hallucinate.
Ram
Ram D. Sriram, Ph.D.
Chief, Software and Systems Division
Information Technology Laboratory
National Institute of Standards and Technology
Gaithersburg, MD 20899
Tel. No: 301-975-3507
From: ontolog...@googlegroups.com <ontolog...@googlegroups.com> on behalf of Alex Shkotin <alex.s...@gmail.com>
Sent: Wednesday, July 26, 2023 1:15 PM
To: ontolo...@googlegroups.com <ontolo...@googlegroups.com>
Cc: ontolog...@googlegroups.com <ontolog...@googlegroups.com>
Subject: [Ontology Summit] Re: [ontolog-forum] The Oxford Handbook of Lying
unsubscribe to the email, see http://ontologforum.org/info/
---
You received this message because you are subscribed to the Google Groups "ontology-summit" group.
To unsubscribe from this group and stop receiving emails from it, send an email to ontology-summ...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/ontology-summit/CAFxxROQR1-2UdJUM6g1qFQw0yO7jQ1gwQc12n9aXfWFYTrv98Q%40mail.gmail.com.
--
All contributions to this forum are covered by an open-source license.
For information about the wiki, the license, and how to subscribe or
unsubscribe to the email, see http://ontologforum.org/info/
---
You received this message because you are subscribed to the Google Groups "ontology-summit" group.
To unsubscribe from this group and stop receiving emails from it, send an email to ontology-summ...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/ontology-summit/BLAPR09MB63374BD79291001CD5B4C694EC00A%40BLAPR09MB6337.namprd09.prod.outlook.com.
--
All contributions to this forum are covered by an open-source license.
For information about the wiki, the license, and how to subscribe or
unsubscribe to the forum, see http://ontologforum.org/info/
---
You received this message because you are subscribed to the Google Groups "ontolog-forum" group.
To unsubscribe from this group and stop receiving emails from it, send an email to ontolog-foru...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/ontolog-forum/1c1b168b73f545dcab783fe86f0625e7%40bestweb.net.
--
All contributions to this forum are covered by an open-source license.
For information about the wiki, the license, and how to subscribe or
unsubscribe to the forum, see http://ontologforum.org/info/
---
You received this message because you are subscribed to the Google Groups "ontolog-forum" group.
To unsubscribe from this group and stop receiving emails from it, send an email to ontolog-foru...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/ontolog-forum/ece7c684-d174-0e09-ee0c-f74bc7cd5e69%40openlinksw.com.