--
You received this message because you are subscribed to the Google Groups "open-nars" group.
To unsubscribe from this group and stop receiving emails from it, send an email to open-nars+...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/open-nars/f645aa59-748d-42eb-9a84-21bb5343ca27n%40googlegroups.com.
I was also thinking about this, in the form of two questions:
"could a term logic dataset (NAL/Narsese) improve the reasoning abilities of LLMs?"
"could a NARS system be implemented as an LLM?"
"are human language(English) pairings necessary?"
We could build a LLM-NARS system around LangChain / LLamaIndex, given the extensions they have it would be immediately more practical agent.The LLM-NARS inference and control would be an LLM model:Trained/finetuned on a generated a NAL ruleset and derivations, and Narsese dataset (from the various implementations?)For control, we could align/finetune/rlhf on another dataset of prompts and desired end/state outputs ala chatgpt/alpaca.The buffer would be the prompt context length (standard 2048 tokens) and output.Memory would be a vector or tensor database.We could keep in alignment with AIKR by keeping the database fixed-size to system resources, ie by pruning the vector/tensor database by least accessed vectors, and/or pruning metadata based on truth values.
With 4bit quantised models, Alpaca.cpp has shown the 7b model memory/disk requirements are 4.21GB, can run on CPU, and can run on Raspberry Pis / Smartphones.With the latest mmap patches, even the 30b model can have a memory usage of 6GB during inference.
What's unclear to me is what would be the best way to go about generating a clean dataset?What exactly would the control dataset look like?Would it be problematic by introducing biases/idiosyncrasies of existing control mechanisms? (would that be an issue?)How would you measure the difference in reasoning ability of LLMs with/without the term logic dataset?
To view this discussion on the web visit https://groups.google.com/d/msgid/open-nars/b0f51228-67e1-4ea6-90b8-2fd75ed82b30n%40googlegroups.com.
I was also thinking about this, in the form of two questions:
"could a term logic dataset (NAL/Narsese) improve the reasoning abilities of LLMs?"
"could a NARS system be implemented as an LLM?"
"are human language(English) pairings necessary?"
The LLM-NARS inference and control would be an LLM model:Trained/finetuned on a generated a NAL ruleset and derivations, and Narsese dataset (from the various implementations?)For control, we could align/finetune/rlhf on another dataset of prompts and desired end/state outputs ala chatgpt/alpaca.The buffer would be the prompt context length (standard 2048 tokens) and output.Memory would be a vector or tensor database.We could keep in alignment with AIKR by keeping the database fixed-size to system resources, ie by pruning the vector/tensor database by least accessed vectors, and/or pruning metadata based on truth values.
With 4bit quantised models, Alpaca.cpp has shown the 7b model memory/disk requirements are 4.21GB, can run on CPU, and can run on Raspberry Pis / Smartphones.With the latest mmap patches, even the 30b model can have a memory usage of 6GB during inference.What's unclear to me is what would be the best way to go about generating a clean dataset?What exactly would the control dataset look like?Would it be problematic by introducing biases/idiosyncrasies of existing control mechanisms? (would that be an issue?)How would you measure the difference in reasoning ability of LLMs with/without the term logic dataset?
On Monday, March 13, 2023 at 4:16:29 PM UTC+1 Pei Wang wrote:
--
You received this message because you are subscribed to a topic in the Google Groups "open-nars" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/open-nars/-7fBHhdRFY8/unsubscribe.
To unsubscribe from this group and all its topics, send an email to open-nars+...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/open-nars/b0f51228-67e1-4ea6-90b8-2fd75ed82b30n%40googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/open-nars/0a32df03-65a3-476f-bb0f-679c68efe7abn%40googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/open-nars/CAL7RFHQT2qo48zKdyWyD3x2XMSU7Hh84tUJHruCWwgchh6O9Gg%40mail.gmail.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/open-nars/e8e2f386-c681-4afe-a278-70b92dac3c58n%40googlegroups.com.