I have the PNAS article, which I’ll try to look at. Note that there has been, at least, title drift between bioarxiv and published:
Bioarxiv: Artificial Neural Networks Accurately Predict Language Processing in the Brain.
PNAS: The neural architecture of language: Integrative modeling converges on predictive processing.
From: ontolo...@googlegroups.com <ontolo...@googlegroups.com>
On Behalf Of Michael DeBellis
Sent: 24 December 2021 00:59
To: ontolog-forum <ontolo...@googlegroups.com>
Subject: [ontolog-forum] The neural architecture of language
CAUTION: This email came from outside of the University. To keep your account safe, only click on links and open attachments if you know the person who sent the email, or you expected to receive this communication. |
I came across the following article that seemed to make some pretty strong claims for machine learning approaches as a model for human language comprehension: https://interestingengineering.com/ai-mimicking-the-brain-on-its-own
I've found that it is important to look at the actual research because things often get distorted and exaggerated when presented in the popular press. The actual journal paper is here: https://www.pnas.org/content/118/45/e2105646118 However, that seems to require an academic login which I don't have but I found an earlier version of the paper that was freely available and more or less the same: https://www.biorxiv.org/content/10.1101/2020.06.26.174482v3.full.pdf+html I've been trying to read this but I just can't get it. What really has me confused is that they seem to be saying that the Machine Learning models predict both NLP facts (e.g., given the first words of a sentence predicting the next word) but that those same models also predict (or correlate to or ???) neural states in people's brains when they look at similar language input. But I feel that I must be misunderstanding that because even if the neural net was analogous to a language processing neural net in the brain there is no way that fMRI or other measures of brain activity come anywhere close to measuring things like which neurons connect and are firing and inhibiting/triggering which other neurons (not to mention many of the ML models aren't even neural nets). I'm just wondering if anyone can take a look at this and provide some high level summary of what the claim is?
Michael
--
All contributions to this forum are covered by an open-source license.
For information about the wiki, the license, and how to subscribe or
unsubscribe to the forum, see
http://ontologforum.org/info/
---
You received this message because you are subscribed to the Google Groups "ontolog-forum" group.
To unsubscribe from this group and stop receiving emails from it, send an email to
ontolog-foru...@googlegroups.com.
To view this discussion on the web visit
https://groups.google.com/d/msgid/ontolog-forum/461bb270-3dcb-45b3-859b-35588c4ee414n%40googlegroups.com.
I may be wrong here, but I THINK what they are claiming is that, using the same structures as standard NLP ANNs, in particular GPT-2, they can build ANNs which do well at predicting the signals they CAN measure in the brain.
The key sentence for me (page 2) is this.
Specifically, we examined the relationships between 43 diverse
state-of-the-art ANN language models (henceforth “models”)
across three neural language comprehension datasets (two fMRI,
one electrocorticography [ECoG]), as well as behavioral signatures
of human language processing in the form of self-paced reading
times, and a range of linguistic functions assessed via standard
engineering tasks from NLP.
To view this discussion on the web visit https://groups.google.com/d/msgid/ontolog-forum/LO2P265MB44006E5CE6D01B8AAC3580BFE07F9%40LO2P265MB4400.GBRP265.PROD.OUTLOOK.COM.