Overcoming Incompleteness

1 view
Skip to first unread message

Mohsen Ravanbakhsh

unread,
May 22, 2007, 6:57:40 AM5/22/07
to everyth...@googlegroups.com
Hi everybody,
It seems Bruno's argument is a bit rich for some of us to digest, so I decided to keep talking by posing another issue.
By Godel's argument we know that every sufficiently powerful system of logic would be incomplete, and recently there has been much argument to make human an exception; that's because we see the truth of Godelian statements ( i.e. This sentence is unprovable in this system)
Let's call such a system S1, and call another (powerful enough in Godel's sense) system S2. and suppose S2 structurally is able to give statements about the statements in S1. What does it mean? Consider S2 as a being able examine some statements in S2 via some operators and get the result(like function calls).
 
My claim is:
 
1.
 S2, is able to see the truth of the Godelian statements in S1, and in some sense:
"S2 is complete against the statements of S1", because it can see that S1 at last wont be able to evaluate our Godelian statement and so the statement would be correct.
 
2.
We humans are vulnerable to the Godelian statements like all other logical systems. We have our paradoxes too.
Consider the same Godelian statement for yourself as a system (i.e. "You can't prove me" or some similar sentences like "This sentence is false")
 
3.in the first claim consider the first system to have the same attitude toward the second one, I mean let there be a loop (some how similar to Hofstadter's Strange loops as the foundation of self)
Is it complete of not? 
 
--

Mohsen Ravanbakhsh.

Bruno Marchal

unread,
May 23, 2007, 11:22:43 AM5/23/07
to everyth...@googlegroups.com

Le 22-mai-07, à 12:57, Mohsen Ravanbakhsh a écrit :

> Hi everybody,
> It seems Bruno's argument is a bit rich for some of us to digest, so I
> decided to keep talking by posing another issue.
> By Godel's argument we know that every sufficiently powerful system of
> logic would be incomplete, and recently there has been much argument
> to make human an exception;


Mainly Lucas 1960 and Penrose recent books. But Emil Post found in 1921
(!!!) both the "godelian" argument against mechanism AND the pitfal (it
is just an error) in that argument.
Those who read french can look at my "Conscience & Mecanisme" for a
thorough analysis of that "error", including the natural "machine's
answer" to that error.

> that's because we see the truth of Godelian statements ( i.e. This
> sentence is unprovable in this system)


Of course we "see" the truth only for simple machine. Not for machine
just a little more complex. This is used explicitly in my approach. A
rich lobian machine (like ZF) can derived the whole theology of PA
(whole means both the provable and unprovable part (at some level)).
But ZF has to make a leap of faith to lift that theology on herself.


> Let's call such a system S1, and call another (powerful enough in
> Godel's sense) system S2. and suppose S2 structurally is able to give
> statements about the statements in S1. What does it mean? Consider S2
> as a being able examine some statements in S2 via some operators and
> get the result(like function calls).
>  
> My claim is:
>  
> 1.
>  S2, is able to see the truth of the Godelian statements in S1, and in
> some sense:
> "S2 is complete against the statements of S1", because it can see that
> S1 at last wont be able to evaluate our Godelian statement and so the
> statement would be correct.

You are correct (up to some details I want bore you right now)

>  
> 2.
> We humans are vulnerable to the Godelian statements like all other
> logical systems.


Yes. As far as we are correct!


> We have our paradoxes too.
> Consider the same Godelian statement for yourself as a system
> (i.e. "You can't prove me" or some similar sentences like "This
> sentence is false")


Better: This sentence is not provable by you.
Is that sentence true? Can you prove it?

>  
> 3.in the first claim consider the first system to have the same
> attitude toward the second one, I mean let there be a loop (some how
> similar to Hofstadter's Strange loops as the foundation of self)
> Is it complete of not? 


S1 can simulate S2, but S1 has no reason to believe whatever S2 says.
There is no problem.
Hofstadter "strange loop" are more related to arithmetical
self-reference or general fixed point of recursive operator (imo).

I have to go. Don't hesitate to make any comments. If I am slow to
answer, it just means I'm a bit busy, not that I consider some question
don't have to be answered. Thanks for the interest,

Bruno

http://iridia.ulb.ac.be/~marchal/

Mohsen Ravanbakhsh

unread,
May 24, 2007, 1:32:29 PM5/24/07
to everyth...@googlegroups.com
Thanks for your patience! , I know that my arguments are somehow raw and immature in your view, but I'm just at the beginning.


S1 can simulate S2, but S1 has no reason to believe whatever S2 says.
There is no problem.
Hofstadter "strange loop" are more related to arithmetical
self-reference or general fixed point of recursive operator


OK then it, becomes my own idea!
Suppose S1 and S2 are the same systems, and both KNOW that the other one is a similar system. Then both have the reason to believe in each others statements, with the improvement that the new system is COMPLETE. We've not exploited any more powerful system to overcome the incompleteness in our system.
I think this is a great achievement!
It's actually like this: YOU believe in ME. THEY give you a godelian statement (You theoretically can not prove this statement) you give it to ME and then see that I can neither prove it nor disprove it, so you tell 
THEM that their statement is true.
But the wonder is in what we do just by ourselves. We have a THEORY OF MIND. You actually do not need to ask me about the truth of that statement, you just simulate me and that's why I can see the a godelian statement is at last true. But in the logical sense ONE system wont be able to overcome the incompleteness, 
so I 
might conclude:
I'M NOT ONE LOGICAL SYSTEM!
This is how we might rich a theory of self. A loopy(!) and multi(!) self.





Mohsen Ravanbakhsh

Russell Standish

unread,
May 24, 2007, 9:59:23 AM5/24/07
to everyth...@googlegroups.com
Sounds plausible that self-aware systems can manage this. I'd like to
see this done as a formal system though, as I have a natural mistrust
of handwaving arguments!

On Thu, May 24, 2007 at 10:32:29AM -0700, Mohsen Ravanbakhsh wrote:
> Thanks for your patience! , I know that my arguments are somehow
> raw and immature in your view, but I'm just at the beginning.
>

> *S1 can simulate S2, but S1 has no reason to believe whatever S2 says.
> There is no problem.
> **Hofstadter "strange loop" are more related to arithmetical
> self-reference or general fixed point of recursive operator*


>
> OK then it, becomes my own idea!
> Suppose S1 and S2 are the same systems, and both KNOW that the other one is
> a similar system. Then both have the reason to believe in each others
> statements, with the improvement that the new system is COMPLETE. We've not
> exploited any more powerful system to overcome the incompleteness in our
> system.
> I think this is a great achievement!
> It's actually like this: YOU believe in ME. THEY give
> you a godelian statement (You theoretically can not prove this
> statement) you give it to ME and then see that I can neither prove it
> nor disprove it, so you tell
> THEM that their statement is true.
> But the wonder is in what we do just by ourselves. We have a THEORY OF MIND.
> You actually do not need to ask me about the truth of that statement, you
> just simulate me and that's why I can see the a godelian statement is at
> last
> true. But in the logical sense ONE system wont be able to overcome the
> incompleteness,
> so I might conclude:
> I'M NOT ONE LOGICAL SYSTEM!
> This is how we might rich a theory of self. A loopy(!) and multi(!) self.
>
>
>

> *
>
> *Mohsen Ravanbakhsh
>
> >

--

----------------------------------------------------------------------------
A/Prof Russell Standish Phone 0425 253119 (mobile)
Mathematics
UNSW SYDNEY 2052 hpc...@hpcoders.com.au
Australia http://www.hpcoders.com.au
----------------------------------------------------------------------------

Jesse Mazer

unread,
May 24, 2007, 9:45:45 PM5/24/07
to everyth...@googlegroups.com
I definitely don't think the two systems could be complete, since (handwavey
argument follows) if you have two theorem-proving algorithms A and B, it's
trivial to just create a new algorithm that prints out the theorems that
either A or B could print out, and incompleteness should apply to this too.

Jesse

_________________________________________________________________
Like the way Microsoft Office Outlook works? You'll love Windows Live
Hotmail.
http://imagine-windowslive.com/hotmail/?locale=en-us&ocid=TXT_TAGHM_migration_HM_mini_outlook_0507

Russell Standish

unread,
May 24, 2007, 10:12:49 PM5/24/07
to everyth...@googlegroups.com
You are right when it comes to the combination of two independent
systems A and B. What the original poster's idea was a
self-simulating, or self-aware system. In this case, consider the liar
type paradox:

I cannot prove this statement

Whilst I cannot prove this statement, I do know it is true, simply
because if I could prove the statement it would be false.

To know that it is true, I am using self-reference about my own proof
capabilities.

I don't think anyone yet has managed a self aware formal system,
although self-reproducing systems have been known since the 1950s, and
are popularly encountered in the form of computer viruses. There has
to be some relationship between a self-reproducing system and a
self-aware system...

Cheers

Jesse Mazer

unread,
May 24, 2007, 10:31:33 PM5/24/07
to everyth...@googlegroups.com
Russell Standish:

>
>You are right when it comes to the combination of two independent
>systems A and B. What the original poster's idea was a
>self-simulating, or self-aware system. In this case, consider the liar
>type paradox:
>
> I cannot prove this statement
>
>Whilst I cannot prove this statement, I do know it is true, simply
>because if I could prove the statement it would be false.

Yes, but Godel statements are more complex than verbal statements like the
one above, they actually encode the complete rules of the theorem-proving
system into the statement. A better analogy might be if you were an upload
(see http://en.wikipedia.org/wiki/Mind_transfer) living in a self-contained
deterministic computer simulation, and the only messages you could send to
the outside world were judgments about whether particular mathematical
theorems were true (once you make a judgement, you can't take it back...for
any judgements your program makes, there must be a halting program that can
show that you'll definitely make that judgement after some finite number of
steps). Suppose you know the complete initial conditions X and dynamical
rules Y of the simulation. Then suppose you're given a mathematical theorem
Z which you can see qualifies as an "encoding" of the statement "the
deterministic computer simulation with initial conditions X and dynamical
rules Y will never output theorem Z as a true statement." You can see
intuitively that it should be true if your reasoning remains correct, but
you can't be sure that at some point after the simulation is running for a
million years or something you won't decide to output that statement in a
fit of perversity, nor can you actually come up with a rigorous *proof* that
you'll never do that, since you can't find any shortcut to predicting the
system's behavior aside from actually letting the simulation run and seeing
what happens.

The same thing would be true even if you replaced an individual in a
computer simulation with a giant simulated community of mathematicians who
could only output a given theorem if they had a unanimous vote, and where
the size of the community was constantly growing so the probability of
errors should be ever-diminishing...although they might hope that they might
never make an error even if the simulation ran forever, they couldn't
rigorously prove this unless they found some shortcut for predicting their
own community's behavior better than just letting the program run and seeing
what would happen (if they did find such a shortcut, this would have strange
implications for their own feeling of free will!)

Jesse

_________________________________________________________________
More photos, more messages, more storage揚et 2GB with Windows Live Hotmail.
http://imagine-windowslive.com/hotmail/?locale=en-us&ocid=TXT_TAGHM_migration_HM_mini_2G_0507

Stephen Paul King

unread,
May 24, 2007, 11:53:59 PM5/24/07
to everyth...@googlegroups.com
Dear Russell,

Isn't the key feature of a "self-aware system" the ability to generate
some form of representation of itself within itself? Would it not be a
simple matter of a system being able to generate some form of simulation of
itself such that there is both a similarity and a difference between the
system and its simulation of itself?
The similarity, it seems to me, would be merely a matter of faithful and
predictible simulation of the system; the difference might be a difference
in state, resolution or some other measure.

For me the question has always been how does one "overcome
Incompleteness" when it is impossible for a simulated system to be identical
to its simulator unless the two are one and the same. Given this identity,
how then does any notion of provability arise when all that exists is a
tautology, A=A.

BTW, have you ever taken a look at Jon Barwise's treatment of the Liar
Paradox?

http://en.wikipedia.org/wiki/Jon_Barwise

Kindest regards,

Stephen

Stephen Paul King

unread,
May 25, 2007, 12:06:22 AM5/25/07
to everyth...@googlegroups.com
Dear Jesse,

Hasn't Stephen Wolfram proven that it is impossible to "shortcut"
predictions for arbitrary behaviours of sufficienty complex systems?

http://www.stephenwolfram.com/publications/articles/physics/85-undecidability/


Stephen

----- Original Message -----
From: "Jesse Mazer" <laser...@hotmail.com>
To: <everyth...@googlegroups.com>
Sent: Thursday, May 24, 2007 10:31 PM
Subject: Re: Overcoming Incompleteness


snip

Russell Standish

unread,
May 25, 2007, 12:14:35 AM5/25/07
to everyth...@googlegroups.com
On Thu, May 24, 2007 at 11:53:59PM -0400, Stephen Paul King wrote:
>
> For me the question has always been how does one "overcome
> Incompleteness" when it is impossible for a simulated system to be identical
> to its simulator unless the two are one and the same.

Is it though? If the simulated system is different from the original,
then indeed I would agree with you.

In the case of human self-awareness, I thought it was implemented not
by simulation as such, but by decoupling the actual inputs and outputs
from the real world, and then feeding the scenario input into the
actual brain circuit, and examine the output _without_ actually moving
a muscle. It has something to do with the "mirror" neurons, and it
really is quite a neat trick (at least Dennett thinks so, and I tend
to agree).

Not being into supernatural explanations, I think a perfectly
mechanical, or formal model should be able to capture this
ability. But how to do it without running into infinite regress is the
challenge. And if and when we have this formal model, we can then see whether
this idea of solving incompleteness has any merit. I'm as sceptical as
anyone, but I do believe the case is more subtle than to be destroyed
by a couple of lines of handwaving argument :).

Cheers

Brent Meeker

unread,
May 25, 2007, 12:31:13 AM5/25/07
to everyth...@googlegroups.com
Russell Standish wrote:
> You are right when it comes to the combination of two independent
> systems A and B. What the original poster's idea was a
> self-simulating, or self-aware system. In this case, consider the liar
> type paradox:
>
> I cannot prove this statement
>
> Whilst I cannot prove this statement, I do know it is true, simply
> because if I could prove the statement it would be false.

But doesn't that depend on having adopted rules of inference and values, i.e. "the sentence is either true or false". Why shouldn't we suppose that self-referential sentences can have other values or that your informal reasoning above is a proof and hence there is contradiction.

ISTM that all this weirdness comes about because we have tried to formalize ordinary reasoning into LOGIC, which works wonderfully for many axiom sets and rules of inference, but not for all. That's why mathematicians generally get along just fine without worrying about completeness or the provability of consistency.

Brent Meeker

>
> To know that it is true, I am using self-reference about my own proof
> capabilities.
>
> I don't think anyone yet has managed a self aware formal system,
> although self-reproducing systems have been known since the 1950s, and
> are popularly encountered in the form of computer viruses. There has
> to be some relationship between a self-reproducing system and a
> self-aware system...

I think it would be almost trivial to create and AI system that would recognize the equivalent of "I cannot prove this statement" is true or mu. It would just be a trick of formal logic and would have very little to do with self-awareness.

Brent Meeker

Jesse Mazer

unread,
May 25, 2007, 5:05:58 AM5/25/07
to everyth...@googlegroups.com
Stephen Paul King wrote:

>
>Dear Jesse,
>
> Hasn't Stephen Wolfram proven that it is impossible to "shortcut"
>predictions for arbitrary behaviours of sufficienty complex systems?
>
>http://www.stephenwolfram.com/publications/articles/physics/85-undecidability/
>
>
>Stephen

The paper itself doesn't seem to prove it--he uses a lot of tentative
language about how certain problems "may" be computational irreducible or
are "expected" to be, as in this paragraph:

"Many complex or chaotic dynamical systems are expected to be
computationally irreducible, and their behavior effectively found only by
explicit simulation. Just as it is undecidable whether a particular initial
state in a CA leads to unbounded growth, to self-replication, or has some
other outcome, so it may be undecidable whether a particular solution to a
differential equation (studied say with symbolic dynamics) even enters a
certain region of phase space, and whether, say, a certain -body system is
ultimately stable. Similarly, the existence of an attractor, say, with a
dimension above some value, may be undecidable."

Still, I think it's plausible that he's correct, and that there are indeed
computations for which there is no "shortcut" to finding the program's state
after N steps except actually running it for N steps.

Jesse

_________________________________________________________________
Catch suspicious messages before you open them謡ith Windows Live Hotmail.
http://imagine-windowslive.com/hotmail/?locale=en-us&ocid=TXT_TAGHM_migration_HM_mini_protection_0507

Mohsen Ravanbakhsh

unread,
May 25, 2007, 7:42:50 AM5/25/07
to everyth...@googlegroups.com
Russell,
Sounds plausible that self-aware systems can manage this. I'd like to
see this done as a formal system though, as I have a natural mistrust
of handwaving arguments! 

 
I like it too :).
I think the computational view would help in construction.
 
Jesse,

I definitely don't think the two systems could be complete, since (handwavey
argument follows) if you have two theorem-proving algorithms A and B, it's
trivial to just create a new algorithm that prints out the theorems that
either A or B could print out, and incompleteness should apply to this too
 
They're not independent systems.putting that aside, I can't find the correspondence to my argument. It would be nice if you could clarify your point.
 
Brent,
But doesn't that depend on having adopted rules of inference and values, i.e. "the sentence is either true or false".  Why shouldn't we suppose that self-referential  sentences can have other values or that your informal reasoning above is a proof and hence there is contradiction

Actually this is the main advantage of making such loop. you're right only when we're talking about ONE system and that system has concluded the truth of such statement. But it can be avoided by two or more systems. They are reliable in each other's view, and have statements for that. maybe that is the only (symmetric!) difference of those systems. Except for that, both(all) are the same. 


Mohsen Ravanbakhsh

Bruno Marchal

unread,
May 25, 2007, 9:02:13 AM5/25/07
to everyth...@googlegroups.com

Le 24-mai-07, à 19:32, Mohsen Ravanbakhsh a écrit :


> Thanks for your patience! , I know that my arguments are somehow
> raw and immature in your view, but I'm just at the beginning.
>
> S1 can simulate S2, but S1 has no reason to believe whatever S2 says.
> There is no problem.
> Hofstadter "strange loop" are more related to arithmetical
> self-reference or general fixed point of recursive operator
>
> OK then it, becomes my own idea!
> Suppose S1 and S2 are the same systems, and both KNOW that the other
> one is a similar system.

They cannot *know* that. The first person associated to each system is
different of the other. Unless you mean something *very* general by
"similar".


> Then both have the reason to believe in each others statements, with
> the improvement that the new system is COMPLETE.


Why? Only if S2 is much more simple than S1, can S1 be complete on S2.
No system can be complete on itself or on anything similar to itself.

> We've not exploited any more powerful system to overcome the
> incompleteness in our system.
> I think this is a great achievement!
> It's actually like this: YOU believe in ME. THEY give
> you a godelian statement (You theoretically can not prove this
> statement) you give it to ME and then see that I can neither prove it n
> or disprove it, so you tell 
> THEM that their statement is true.


If that is a proof, then you are inconsistent. If it is an intuition,
or a prayer, a hope, a bet, or anything like that, then it is ok, but
you don't need to system for that. S1 can use the very similar system
S1 itself.
The apparent infinite regression is soleved by the traditional
diagonalisation technics. Look in the archive for both
diagonalisation and
diagonalization
in the archive. Or ask (and be patient :)


> But the wonder is in what we do just by ourselves. We have a THEORY OF
> MIND. You actually do not need to ask me about the truth of that
> statement, you just simulate me and that's why I can see the a
> godelian statement is at last true. 


But a simulation is not a proof, especially if the simulation doesn't
halt.


> But in the logical sense ONE system wont be able to overcome the incomp
> leteness, 
> so I might conclude:
> I'M NOT ONE LOGICAL SYSTEM!
> This is how we might rich a theory of self. A loopy(!) and multi(!)
> self.


Here I do agree (but from a different reasoning). See later, or,
meanwhile, search for "Plotinus" or "guardian angel" or "hypostases".

Very shortly, the lobian-godelian incompleteness forces the distinction
between:

p
Bp
Bp & p
Bp & Dt
(Bp & p) & Dt

which makes a total of eight notions of "self" (8, not 5, because 3 of
them splits in two different logic due to incompleteness. B is for
"beweisbar" and correspond to Goel arithmetical provability predicate.

You can read them as

truth
provability
knowability
observability
sensibility

or, with Plotinus (300 AD):

ONE
INTELLECT
SOUL
INTELLIGIBLE MATTER
SENSIBLE MATTER

With the provable versus true distinction (the [ ] / [ ]*
distinction), the hypostases are:


one
discursive Intellect divine Intellect
soul
intell. matter intell. matter
sensible mat. sensible mat

AND THEN, éadding the comp assumption leads to the comp physics, for
the soul and both matter hypostases, and it is enough to cmpare the
comp-physics with the empirical physics to evaluate the degree of
plausibility of comp. And that's my point.

Bruno


http://iridia.ulb.ac.be/~marchal/

Jesse Mazer

unread,
May 25, 2007, 9:24:54 AM5/25/07
to everyth...@googlegroups.com
Mohsen Ravanbakhsh
>
>*Jesse,

>I definitely don't think the two systems could be complete, since
>(handwavey
>argument follows) if you have two theorem-proving algorithms A and B, it's
>trivial to just create a new algorithm that prints out the theorems that
>either A or B could print out, and incompleteness should apply to this too*
>**

>They're not independent systems.putting that aside, I can't find the
>correspondence to my argument. It would be nice if you could clarify your
>point.

I didn't say they were independent--but each has a well-defined set of
theorems that they will judge to be true, no? My point was just that they
could not together be complete as you say, since the combination of the two
can always be treated as a *single* axiomatic system or theorem-proving
algorithm which proves every theorem in the union of the two sets A and B
prove individually, and this must necessarily be incomplete--there must be
true theorems of arithmetic which this single system cannot prove (meaning
that they don't belong to the set A can prove or the set B can prove).

Bruno Marchal

unread,
May 25, 2007, 10:00:40 AM5/25/07
to everyth...@googlegroups.com

Le 25-mai-07, à 04:12, Russell Standish a écrit :

> I don't think anyone yet has managed a self aware formal system,

I would say all my work is about that. You can interpret Godel's
theorem, or more exactly the fact that machine can prove their own
provability logic, and even guess correctly the true but non provable
part, as a general statement of self-awareness. Sometimes,
self-awareness is defined more precisely by the "4" modal formula: Bp
-> BBp. This is a theorem for PA, but not for LRA.
When LRA proves p, then soon or later, if not already, LRA will prove
Bp (that is LRA will prove that she prove p). So Bp -> BBp is true for
LRA, but only PA can prove such a proposition on itself.

> although self-reproducing systems have been known since the 1950s, and
> are popularly encountered in the form of computer viruses.

The principle is really the same. That is what I show in my paper
"amoeba, planaria, and dreaming machine". self-reproduction and
self-awareness are a consequence of the closure of the set of partial
computable function for the daigonalization procedure.


bruno

http://iridia.ulb.ac.be/~marchal/

James N Rose

unread,
May 25, 2007, 10:11:49 AM5/25/07
to everyth...@googlegroups.com
Bruno, et al.,

There is a CRITICAL FUNDAMENTAL ERROR in
Godel's papers and concept.

If a simpler 'less complete' system - which
-includes- its statements, attempts to make
-presumptive statements- about a 'more complete'
corresponding system ... and its relationship to
the simpler 'base of statements' system, then a
conclusion is: a system CANNOT accurately 'self
assess' but can accurately 'other assess' information
which may not in fact be present for assessment.

Generalized conclusion:

It is not possible to assess known information
whereas it is possible to assess unknown information.


James Rose

Stephen Paul King

unread,
May 26, 2007, 1:40:36 AM5/26/07
to everyth...@googlegroups.com
Hi Russell,

----- Original Message -----
From: "Russell Standish" <li...@hpcoders.com.au>
To: <everyth...@googlegroups.com>

Sent: Friday, May 25, 2007 12:14 AM
Subject: Re: Overcoming Incompleteness


>
> On Thu, May 24, 2007 at 11:53:59PM -0400, Stephen Paul King wrote:
>>
>> For me the question has always been how does one "overcome
>> Incompleteness" when it is impossible for a simulated system to be
>> identical
>> to its simulator unless the two are one and the same.
>
> Is it though? If the simulated system is different from the original,
> then indeed I would agree with you.
>

[SPK]

It was the difference that i was trying to focus on... "Bisimulation"
is, after all, a form of identity if exact.

> In the case of human self-awareness, I thought it was implemented not
> by simulation as such, but by decoupling the actual inputs and outputs
> from the real world, and then feeding the scenario input into the
> actual brain circuit, and examine the output _without_ actually moving
> a muscle. It has something to do with the "mirror" neurons, and it
> really is quite a neat trick (at least Dennett thinks so, and I tend
> to agree).
>

[SPK]

Ok, but that is it that is "generating" and "examining" the inputs and
outputs? I am trying to frame this in software terms...

> Not being into supernatural explanations, I think a perfectly
> mechanical, or formal model should be able to capture this
> ability. But how to do it without running into infinite regress is the
> challenge. And if and when we have this formal model, we can then see
> whether
> this idea of solving incompleteness has any merit. I'm as sceptical as
> anyone, but I do believe the case is more subtle than to be destroyed
> by a couple of lines of handwaving argument :).
>

[SPK]

We avoid infinite regress by having only finite computational resourses.
X can only generate a simulation of itself with a resolution whose upper
bound is determined by the resourses that X has available within the span of
of the simulation of X. Remember, X is not a static entity...

Kindest regards,

Stephen

Mohsen Ravanbakhsh

unread,
May 26, 2007, 2:26:59 PM5/26/07
to everyth...@googlegroups.com
Hi everybody,
I need to clarify. When we build this new combined system, we would be immune to Godelian statements for one of them not for the whole system, whatever it might be. So Jesse's argument does not hold, and of course the new system does not contradict the Godel's theorem, it's (was!) just a way to avoid it.

Bruno:*"...Only if S2 is much more simple than S1, can S1 be complete on S2." *
And finally this ruins everything...

-

Mohsen Ravanbakhsh.

Jesse Mazer

unread,
May 26, 2007, 2:38:28 PM5/26/07
to everyth...@googlegroups.com
Mohsen Ravanbakhsh wrote:

>
>Hi everybody,
>I need to clarify. When we build this new combined system, we would be
>immune to Godelian statements for one of them not for the whole system,
>whatever it might be. So Jesse's argument does not hold, and of course the
>new system does not contradict the Godel's theorem, it's (was!) just a way
>to avoid it.

But didn't you claim the combination of the two of them would be "complete"?
"Complete" is supposed to mean a system will print out *every* true
statement about arithmetic, and a Godel statement for a theorem-proving
system is itself a true statement about arithmetic. So if the combined
system has a Godel statement that it will never print out, and the combined
system prints out every statement that the two of them can print out, then
the combination of the two of them does not allow you to escape
incompleteness.

Mohsen Ravanbakhsh

unread,
May 26, 2007, 3:01:03 PM5/26/07
to everyth...@googlegroups.com
On 5/26/07, Jesse Mazer <laser...@hotmail.com> wrote:

Mohsen Ravanbakhsh wrote:

>
>Hi everybody,
>I need to clarify. When we build this new combined system, we would be
>immune to Godelian statements for one of them not for the whole system,
>whatever it might be. So Jesse's argument does not hold, and of course the
>new system does not contradict the Godel's theorem, it's (was!) just a way
>to avoid it.

But didn't you claim the combination of the two of them would be "complete"?
"Complete" is supposed to mean a system will print out *every* true
statement about arithmetic, and a Godel statement for a theorem-proving
system is itself a true statement about arithmetic. So if the combined
system has a Godel statement that it will never print out, and the combined
system prints out every statement that the two of them can print out, then
the combination of the two of them does not allow you to escape
incompleteness.

Jesse

_________________________________________________________________
More photos, more messages, more storage--get 2GB with Windows Live Hotmail.
http://imagine-windowslive.com/hotmail/?locale=en-us&ocid=TXT_TAGHM_migration_HM_mini_2G_0507





Aha, there seems to be subtle point here:
 My claim was: the new system would be complete regarding the statements each one of two systems(one of them, because they're the same) can generate, BUT building such a system would require some additional staff(!) to glue those similar systems so they'd rely on each other, and follow the mentioned mechanism. Those extra THINGS still should(!) cause the incompleteness of the new system as a whole regarding the statements that consider this coupling.
Anyway... it didn't work.

--

Mohsen Ravanbakhsh.

Russell Standish

unread,
May 26, 2007, 4:32:29 PM5/26/07
to everyth...@googlegroups.com
On Fri, May 25, 2007 at 04:00:40PM +0200, Bruno Marchal wrote:
>
>
> Le 25-mai-07, à 04:12, Russell Standish a écrit :
>
> > I don't think anyone yet has managed a self aware formal system,
>
> I would say all my work is about that. You can interpret Godel's
> theorem, or more exactly the fact that machine can prove their own
> provability logic, and even guess correctly the true but non provable
> part, as a general statement of self-awareness. Sometimes,
> self-awareness is defined more precisely by the "4" modal formula: Bp
> -> BBp. This is a theorem for PA, but not for LRA.
> When LRA proves p, then soon or later, if not already, LRA will prove
> Bp (that is LRA will prove that she prove p). So Bp -> BBp is true for
> LRA, but only PA can prove such a proposition on itself.
>

Absolutely your work is all about that. And constructing a formal
self-aware system will just as likely come from an approach like
yours as anything else. My only statement was that it hasn't been done
already, and certainly I wasn't aware of you claiming self-awareness
already. Admittedly, I miss some of the conclusions you do make, such as the
Theatetic axioms encapsulating knowledge, and also Kripke frames being
the phenomena of time (I can see that they could be a model of time,
but that is another thing entirely).

>
>
> > although self-reproducing systems have been known since the 1950s, and
> > are popularly encountered in the form of computer viruses.
>
> The principle is really the same. That is what I show in my paper
> "amoeba, planaria, and dreaming machine". self-reproduction and
> self-awareness are a consequence of the closure of the set of partial
> computable function for the daigonalization procedure.
>

Incidently, I read your Elsevier paper the other day. It has inspired
me to take a look at combinators for artificial life applications. I
discovered that Pietro Speroni di Fenizio has spent a PhD looking at
this. He was a student of Walter Banzhaf and Peter Dittrich - I met
Walter shortly before I caught up with you in Brussels in 2003. Peter
I met in the Lausanne ECAL in 1999. Are you aware of any of this work?
AFACT, he never achieved self-reproduction, which is a precursor to
evolution, but alas he didn't publish any of his code.

>
> bruno
>
>
>
> http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
May 29, 2007, 9:05:52 AM5/29/07
to everyth...@googlegroups.com

Le 26-mai-07, à 22:32, Russell Standish a écrit :

>
> On Fri, May 25, 2007 at 04:00:40PM +0200, Bruno Marchal wrote:
>>
>>
>> Le 25-mai-07, à 04:12, Russell Standish a écrit :
>>
>>> I don't think anyone yet has managed a self aware formal system,
>>
>> I would say all my work is about that. You can interpret Godel's
>> theorem, or more exactly the fact that machine can prove their own
>> provability logic, and even guess correctly the true but non provable
>> part, as a general statement of self-awareness. Sometimes,
>> self-awareness is defined more precisely by the "4" modal formula: Bp
>> -> BBp. This is a theorem for PA, but not for LRA.
>> When LRA proves p, then soon or later, if not already, LRA will prove
>> Bp (that is LRA will prove that she prove p). So Bp -> BBp is true for
>> LRA, but only PA can prove such a proposition on itself.
>>
>
> Absolutely your work is all about that. And constructing a formal
> self-aware system will just as likely come from an approach like
> yours as anything else. My only statement was that it hasn't been done
> already, and certainly I wasn't aware of you claiming self-awareness
> already.


Of course many things depends on definitions, but I thought it was
clear that I consider that any theorem prover machine, for a theory
like ZF or PA, is already self-aware. And of course such theorem prover
already exist (not just in platonia).
The "interview" of the lobian machine has already begun, through the
papers of Godel, Lob, ... Then Solovay's results have captured, through
the modal logic G and G*, the whole (infinite) conversation we can
have with and about such machines (G gives (self-) science and G* gives
the (self) correct hope/bet/interrogation/theology).
The technical part of my thesis is really the *result* of the interview
of such self-aware (and even aware of being self-aware) machine. But
frankly I am glad with Smullyan (rather classical) definition of
self-awareness in his little book "FOREVER UNDECIDED":
A machine M is self-aware if, having proved p, she will prove soon or
later if not already, Bp.
A machine M is self-aware of its self-awareness if for any p she proves
Bp -> BBp (that is: she knows the preceding line for any proposition
p).
No need to put too much philosophy here, it could be prematured.
As a platonist (or just mathematician), no need to have the interview
with an "incarnate" machine, we can stay in Platonia, where such
machine and their interviews exist out of space and time. That is good
given that the goal (made necessary by the UDA) is to derived the
structure of space and time from the interview!
But yes, indeed I consider already PA or ZF as person, and this with or
without terrestrial local implementations.

> Admittedly, I miss some of the conclusions you do make, such as the

> Theatetic axioms encapsulating knowledge, ...

I remember. You said some times ago that the definition of knowledge of
P (Kp) by true justified opinion (Bp & p) is debatable, and you are
right; it is somehow debated since, well, a very long time, and in many
places: China, India, Greece up to Plato.
I postpone a bit a debate on that. Note that this idea is used
implicitly at step 6 of the Universal Dovetailer Argument. It is
related to the use of dream in metaphysics.
And such a definition has to be nuanced, at least if we trust Socrate's
critics of Theatetus, or Plotinus or ... the lobian machine itself!!!
Indeed, by incompleteness we have the new Bp & Dp nuances, and we have
the Bp & Dp & p nuances, etc.
Actually those who criticize the definition of knowledge by "Bp & p"
have to maintain that they are able to distinguish their dream and
awakeness states. But with comp, the thought experiments (or their
arithmetical godelian translation) show that we cannot, locally, or in
a bounded finite time, make that distinction.
This explains why a philosopher like Malcolm (ref in my thesis), who
reasons very rigorously, and who believes the theaetetical definition
of knowledge is wrong, is forced both to disbelieve in comp, and to
disbelieve in consciousness in dreams (like with lucidity, but not
necessarily).
Very interestingly, Barnes (ref in my theses) makes a similar move to
escape Maudlin's "~ comp or ~ materialism" argument. I am writing
(albeit slowly) a paper on that for the journal of philosophy as a
rejoinder to Maudlin and Barnes. We can come back on this later.

> ....and also Kripke frames being


> the phenomena of time (I can see that they could be a model of time,
> but that is another thing entirely).

I don't remember having said that "Kripke frames" *are* the phenomena
of times? What would that means? All I say is that assuming comp (well
the weak acomp) then the soul or the first person generates its
subjective time, and that this is confirmed by the fact that the
first-person hypostasis (the one given by the theaetetical definition
of knowability) has a temporal-intuitionistic logical semantics.
That this corresponds on the "real subjective time" for the
self-observing machine already follows from the UDA, where 1)
comp-practitioners does not modelize their brain with an artificial
one, but really accept it as a new brain + 2) the non distinction
between "real", virtual and arithmetical.

>>
>>> although self-reproducing systems have been known since the 1950s,
>>> and
>>> are popularly encountered in the form of computer viruses.
>>
>> The principle is really the same. That is what I show in my paper
>> "amoeba, planaria, and dreaming machine". self-reproduction and
>> self-awareness are a consequence of the closure of the set of partial
>> computable function for the daigonalization procedure.
>>
>
> Incidently, I read your Elsevier paper the other day. It has inspired
> me to take a look at combinators for artificial life applications. I
> discovered that Pietro Speroni di Fenizio has spent a PhD looking at
> this. He was a student of Walter Banzhaf and Peter Dittrich - I met
> Walter shortly before I caught up with you in Brussels in 2003. Peter
> I met in the Lausanne ECAL in 1999. Are you aware of any of this work?
> AFACT, he never achieved self-reproduction, which is a precursor to
> evolution, but alas he didn't publish any of his code.

Thanks for the information. I didn't know him. But I know the work by
Fontana and some of its followers in Artificial Chemistry, and which is
based on lambda calculus (the little cousin of combinatory logic).
Lambda expressions and combinators are very useful to analyze
computation with some fine graining (as my Elsevier paper illustrates a
bit). But theology needs more the notion of computability (which is
independent of the choice of formal systems) than the notion of
computation, which is far more involved. Of course there are some
links.


Bruno


http://iridia.ulb.ac.be/~marchal/

Russell Standish

unread,
May 29, 2007, 1:31:55 AM5/29/07
to everyth...@googlegroups.com
On Tue, May 29, 2007 at 03:05:52PM +0200, Bruno Marchal wrote:
>
>
> Of course many things depends on definitions, but I thought it was
> clear that I consider that any theorem prover machine, for a theory
> like ZF or PA, is already self-aware. And of course such theorem prover
> already exist (not just in platonia).

No it wasn't clear. Also theorem proving software exists already in
the real world, and has been used to validate elementary circuit
design. Unfortunately, more complicated circuits such as Pentium
processors are beyond these programs. But I've never heard of anyone
considering these programs self-aware before.

> But
> frankly I am glad with Smullyan (rather classical) definition of
> self-awareness in his little book "FOREVER UNDECIDED":
> A machine M is self-aware if, having proved p, she will prove soon or
> later if not already, Bp.
> A machine M is self-aware of its self-awareness if for any p she proves
> Bp -> BBp (that is: she knows the preceding line for any proposition
> p).

I will try to borrow a copy of this book to read his
justification. These definitions you give are not at all
obvious. Sydney Uni has a copy in its library - unfortunately there is
a bit of administrative hassle for me to organise borrowing priveleges
there.

> This explains why a philosopher like Malcolm (ref in my thesis), who
> reasons very rigorously, and who believes the theaetetical definition
> of knowledge is wrong, is forced both to disbelieve in comp, and to
> disbelieve in consciousness in dreams (like with lucidity, but not
> necessarily).

Well this is different from me. I don't believe the theatetical
definitions are wrong as such, I just don't really understand their
justification. If it were a question of terminology, I could happily
call it "Theatetical knowledge", and prove all sorts of wonderful
results about it, but still never know what it corresponds to in the
real world.

Maybe Smullyan has some justification for this?


> > ....and also Kripke frames being
> > the phenomena of time (I can see that they could be a model of
> > time,
> > but that is another thing entirely).
> >
>
>
>
> I don't remember having said that "Kripke frames" *are* the phenomena
> of times? What would that means? All I say is that assuming comp (well
> the weak acomp) then the soul or the first person generates its
> subjective time, and that this is confirmed by the fact that the

This is exactly what I meant. Time for me is subjective (unless one is
talking about coordinate time, which one does in physics, but is not
what we're talking about here).

> first-person hypostasis (the one given by the theaetetical definition
> of knowability) has a temporal-intuitionistic logical semantics.

This is the point I never understood. Something about it "quacks like
a duck", so must therefore be a duck?

> That this corresponds on the "real subjective time" for the
> self-observing machine already follows from the UDA, where 1)
> comp-practitioners does not modelize their brain with an artificial
> one, but really accept it as a new brain + 2) the non distinction
> between "real", virtual and arithmetical.
>
>
>
>

> Thanks for the information. I didn't know him. But I know the work by
> Fontana and some of its followers in Artificial Chemistry, and which is
> based on lambda calculus (the little cousin of combinatory logic).
> Lambda expressions and combinators are very useful to analyze
> computation with some fine graining (as my Elsevier paper illustrates a
> bit). But theology needs more the notion of computability (which is
> independent of the choice of formal systems) than the notion of
> computation, which is far more involved. Of course there are some
> links.
>

Fontana had a paper about this at the Brussels ECAL, which you attended
IIRC. I missed the Brussels ECAL due to having to return to Australia
before it (I was living in Germany in 1992), otherwise I would have
been there. I remember feeling very disappointed about it.

Speroni di Fenizio builds on Fontana's model in several important
respects, but is certainly in that tradition.

>
> Bruno

Bruno Marchal

unread,
May 30, 2007, 10:00:20 AM5/30/07
to everyth...@googlegroups.com

Le 29-mai-07, à 07:31, Russell Standish a écrit :

>
> On Tue, May 29, 2007 at 03:05:52PM +0200, Bruno Marchal wrote:
>>
>>
>> Of course many things depends on definitions, but I thought it was
>> clear that I consider that any theorem prover machine, for a theory
>> like ZF or PA, is already self-aware. And of course such theorem
>> prover
>> already exist (not just in platonia).
>
> No it wasn't clear. Also theorem proving software exists already in
> the real world, and has been used to validate elementary circuit
> design. Unfortunately, more complicated circuits such as Pentium
> processors are beyond these programs. But I've never heard of anyone
> considering these programs self-aware before.


This is due probably because they have not been conceived in that
spirit.


Actually Smullyan use the Theaetical definition without going in any
details. He just say that he will say that a machine believes p when
the machine asserts p, and that the machine knows p when both p is true
and the machine asserts p.
More involved explanations and critics occur in the original Theaetetus
by Plato, and all along the philosophical litterature.
Wittgenstein makes a case in his late writing for it when he realize
that knowledge and beliefs could corresponds to the same mental state
in different context.
What is really not obvious at all, is that the theaetical definition
works for the provability of correct machine. Why? Because obviously if
the machine is correct then Bp and Bp & p are obviously equivalent. The
key is that this equivalence cannot be proved or asserted by the
machine, because no correct (lobian) machine can ever know being
correct. She would know that Bp -> p, and in particular she would know
that Bf -> f, that is ~Bf, that is: she would know that she is
consistent, contradicting Godel II.
Same reasoning for all the hypostases: they are all equivalent, in term
of provability or assertability, but yet, from the machine point of
view, they obey drastically different logics.

>
>
>>> ....and also Kripke frames being
>>> the phenomena of time (I can see that they could be a model of
>>> time,
>>> but that is another thing entirely).
>>>
>>
>>
>>
>> I don't remember having said that "Kripke frames" *are* the
>> phenomena
>> of times? What would that means? All I say is that assuming comp
>> (well
>> the weak acomp) then the soul or the first person generates its
>> subjective time, and that this is confirmed by the fact that the
>
> This is exactly what I meant. Time for me is subjective (unless one is
> talking about coordinate time, which one does in physics, but is not
> what we're talking about here).
>
>> first-person hypostasis (the one given by the theaetetical definition
>> of knowability) has a temporal-intuitionistic logical semantics.
>
> This is the point I never understood. Something about it "quacks like
> a duck", so must therefore be a duck?

It is know that intuitionistic logic has a temporal flavor, and this is
nice from the point of view of informal intuitionistic philosophy which
is already a theory of "subjective time" à-la Bergson or Brouwer.

Now, Godel already suggested in 1933 (and btw Kolmogorov discovered
this in 1932) that the modal logic S4 provides an epistemic
interpretation of intuitionistic logic. Godel suggests to translate
intuitionistic logic in the (classical) modal logic S4 by using the
following recursive translation t_int:

t_int(p ) = Bp (atmic p)
t_int(~p) = ~B(t_int(p))
t_int(p -> q) = B(t_int(p) -> B(t_int(q)
t_int(p & q) = t_int(p) & t_int(q)
t_int(p V q) = B(t_int(p) V B(t_int(q)

McKinsey and Tarski will prove this works well later, and Grzegorczyk
will extend that result for S4Grz (that is S4 + the Grz formula). In
the Kripke frame, the Grz makes the accessibility relation
antisymmetrical, making this even more palatable for a theory of
subjective sense.

This is related to G, because as Boolos, Goldblatt (and also Kusnetzov
and Muravitsky in the ex-soviet union) will show that S4grz is sound
and complete for the provability-and-truth notion, that is for the
application of theatetus to provability. I have extend this with the
"weaker" provability-and-consistency, and
provability-and-consistency-and-truth, etc.


Gosh, I realize the translation above is more correct than the one I
put in my Plotinus (SIENA) paper. I always put typo errors ... Hmmm...

I will try a succinct summary of the arithmetical interpretations of
the Plotinian hypostases. Good preparation for me, indeed I have 20m in
Sienna to sum up years and years of work, not counting the work of the
many multi-disciplinay predecessors! But at least in Siena the
audience is suspected to know logic, and seems very open minded for
"less-logical things" ...
Hope I have not been too technical for the non technical people. I
really recommend the reading of Plato's Theatetus for the best
introduction to the arithmetical hypostases appearing when we interview
universal machines.

Bruno

http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
May 31, 2007, 10:23:07 AM5/31/07
to everyth...@googlegroups.com

Le 30-mai-07, à 16:00, Bruno Marchal a écrit :

>
>
> Le 29-mai-07, à 07:31, Russell Standish a écrit :
>
>>
>> On Tue, May 29, 2007 at 03:05:52PM +0200, Bruno Marchal wrote:
>>>
>>>
>>> Of course many things depends on definitions, but I thought it was
>>> clear that I consider that any theorem prover machine, for a theory
>>> like ZF or PA, is already self-aware. And of course such theorem
>>> prover
>>> already exist (not just in platonia).
>>
>> No it wasn't clear. Also theorem proving software exists already in
>> the real world, and has been used to validate elementary circuit
>> design. Unfortunately, more complicated circuits such as Pentium
>> processors are beyond these programs. But I've never heard of anyone
>> considering these programs self-aware before.
>
>
> This is due probably because they have not been conceived in that
> spirit.


Of course when I say that the machine or the theorem prover for PA is
self-aware, you have to keep in mind the whole background of the comp
assumption, including the main movie-graph consequence which forbids to
attach consciousness to any "spatio-temporal-physical activity". PA
consciousness is de-localized and distributed in *all* the LRA or UD
computations, and it is arguable that any of your own present
consciousness is already part of many PA's life (even without comp, you
*do* extend PA. ("you" = all of you, not just Russell Standish!).

Locally PA's awareness is just the fact that when PA proves p soon or
later PA proves Bew('p'), (which I write "modally" as Bp). Now even the
UD, or LRA (little robinson arithmetic) are aware in that sense.
PA's self-awareness is the fact that for any arithmetical p, PA proves
Bp -> BBp. LRA definitely lack that self-awareness. Both LRA and PA are
under the consequences of incompleteness, but only PA and its sound
logical descendants can overcome it.

Actually the mathematical capture of the Universal Dovetailer Argument
(UDA) or of Plotinus, can be appreciated without any ontological
assumption, and should please to positivist and formalist. On the
contrary the original UDA itself asks for some non mathematical
assumption like the "yes doctor", and can please only through some
(minimal) arithmetical-platonist and cognitive assumptions.

Of course I do believe all lobian machine are self-aware in that sense.
And thanks to their lack of (human? Aristotelian?) prejudices, they
appear to be very close to Plotinus when they describe what they
discover by looking inward.

In a sense PA, ZF, ... are more self-aware than us ... perhaps more in
this time where self-reflection seems to be a bit out of fashion ... I
think this is in part due to the mixing of theology religion and
politics, where the doubting attitude is discouraged ... (other but
related debate).


Bruno

http://iridia.ulb.ac.be/~marchal/

Reply all
Reply to author
Forward
0 new messages