MGA 3

34 views
Skip to first unread message

Bruno Marchal

unread,
Nov 26, 2008, 4:09:01 AM11/26/08
to everyth...@googlegroups.com
MGA 3

It is the last MGA !

I realize MGA is complete, as I thought it was, but I was doubting this recently. We don't need to refer to Maudlin, and MGA 4 is not necessary.
Maudlin 1989 is an independent argument of the 1988 Toulouse argument (which I present here). 
Note that Maudlin's very interesting "Olympization technic"  can be used to defeat a wrong form of MGA 3, that is, a wrong argument for the assertion that  the movie cannot be conscious. (the argument that the movie lacks the counterfactual). Below are hopefully correct (if not very simple) argument. ( I use Maudlin sometimes when people gives this non correct form of MGA 3, and this is probably what makes me think Maudlin has to be used, at some point).



MGA 1 shows that Lucky Alice is conscious, and MGA 2 shows that the "luckiness" feature of the MGA 1 experiment was a red herring. We can construct, from MEC+COMP, an home made lucky rays generator, and use it at will. If we accept both digital mechanism, in particular Dennet's principle that neurons have no intelligence, still less prescience, and this  *together with* the supervenience principle; we have to accept that Alice conscious dream experience supervenes on the projection of her brain activity movie.

Let us show now that Alice consciousness *cannot* supervene on that *physical* movie projection.



I propose two (deductive) arguments.

1)

Mechanism implies the following tautological functionalist principle: if, for some range of activity, a system does what it is supposed to do, and this before and after a change is made in its constitution, then the change does not change what the system is supposed to do, for that range of activity.
Example:
- A car is supposed to broken but only if the driver is faster than 90 miles/h. Pepe Pepito NEVER drives faster than 80 miles/h. Then the car is supposed to do what she is supposed to do, with respect of its range of activity defined by Pepe Pepito.
- Claude bought a 1000 thousand processors computer. One day he realized that he used only 990 processors, for his type of activity, so he decided to get rid of those 10 useless processors. And indeed the machine will satisfy Claude ever.

- Alice has (again) a math exam. Theoreticians have correctly predict that in this special circumstance, she will never use neurons X, Y and Z.  Now Alice go (again, again) to this exam in the same condition, but with the neurons X, Y, Z removed. Again, not only will she behaved like if she succeed her exam, but her consciousness, with both MEC *and* MAT still continue.
The idea is that if something is not useful, for an active process to go on, for some range of activity, then you can remove it, for that range of activity.

OK?

Now, consider the projection of the movie of the activity of Alice's brain, "the movie graph".
Is it necessary that someone look at that movie? Certainly not. No more than it is needed that someone is look at your reconstitution in Moscow for you to be conscious in Moscow after a teleportation. All right? (with MEC assumed of course).
Is it necessary to have a screen? Well, the range of activity here is just one dynamical description of one computation. Suppose we make a hole in the screen. What goes in and out of that hole is exactly the same, with the hole and without the hole. For that unique activity, the hole in the screen is functionally equivalent to the subgraph which the hole removed. Clearly we can make a hole as large as the screen, so no need for a screen.
But this reasoning goes through if we make the hole in the film itself. Reconsider the image on the screen: with a hole in the film itself, you get a "hole" in the movie, but everything which enters and go out of the hole remains the same, for that (unique) range of activity.  The "hole" has trivially the same functionality than the subgraph functionality whose special behavior was described by the film. And this is true for any subparts, so we can remove the entire film itself.

Does Alice's dream supervene (in real time and space) on the projection of the empty movie?

Remark. 
1° Of course, this argument can be sum up by saying that the movie lacks causality between its parts so that it cannot really be said that it computes any thing, at least physically. The movie is just an ordered record of computational states. This is neither a physical computation, nor an (immaterial) computation where the steps follows relatively to some universal machine. It is just a description of a computation, already existing in the Universal Deployment.
2° Note this: If we take into consideration the relative destiny of Alice, and supposing one day her brain broke down completely, she has more chance to survive through "holes in the screen" than to the "holes in the film". The film contains the relevant information to reconstitute Alice from her brain description, contained on this high resolution film. Keeping comp, and abandoning the physical supervenience thesis, means that we do no more associate consciousness, neither on the movie, NOR on the brain special activity in a computation, but to the computation itself directly. A brain, and even a film,  will "only" be a way to make bigger the probability
for a consciousness to manifest itself relatively to a "probable" universal computational history.
Strictly speaking, running the movie dimimish Alice chance to have her conscious experience (life) continue, at least relatively to you, because of the many scratches the projector makes on the pellicle, which remove relevant information for a safe reconstitution later (again relatively to you).


2)

I give now what is perhaps a simpler argument

A projection of a movie is a relative phenomenon. On the planet 247a, nearby in the galaxy, they don't have screen. The film pellicle is as big as a screen, and they make the film passing behind a stroboscope at the right frequency in front of the public. But on planet 247b, movies are only for travellers! They dress their film, as big as those on planet 247a, in their countries all along their train rails with a lamp besides each frames, which is nice because from the train, through its speed, you get the usual 24 frames per second. But we already accepted that such movie does not need to be observed, the train can be empty of people. Well the train does not play any role, and what remains is the static film with a lamp behind each frame. Are the lamps really necessaries? Of course not, all right? So now we are obliged to accept that the consciousness of Alice during the projection of the movie supervenes of something completely inert in time and space. This contradicts the *physical*  supervenience thesis.


Exercises.

a) Someone could propose an alternate argument that a movie does not compute (and so consciousness does supervene on it) by alluding to the lack of causality in the movie:  the movie does not handle the counterfactual existing implicitly in computations (physical or not).  Use Maudlin's Olympization technic to refute that argument.

b) Make fun by using a non dreaming Alice. Shows that the movie (film or screen) graph border is needed to get the accidental zombies (the puppet).

And then the "important" exercise (the original goal). 

c) Eliminate the hypothesis "there is a concrete deployment" in the seventh step of the UDA. Use UDA(1...7) to define properly the computationalist supervenience thesis. Hint: reread the remarks above.

Michael Rosefield

unread,
Nov 26, 2008, 8:39:47 AM11/26/08
to everyth...@googlegroups.com
There's a quote you might like, by Korzybski: "That which makes no difference _is_ no difference."

--------------------------
- Did you ever hear of "The Seattle Seven"?
- Mmm.
- That was me... and six other guys.


2008/11/26 Bruno Marchal <mar...@ulb.ac.be>

Abram Demski

unread,
Nov 27, 2008, 2:02:10 PM11/27/08
to everyth...@googlegroups.com
Bruno,

It seems to me that this runs head-on into the problem of the
definition of time...

Here is my argument; I am sure there will be disagreement with it.

Supposing that Alice's consciousness is spread out over the movie
billboards next to the train track, there is no longer a normal
temporal relationship between mental moments. There must merely be a
"time-like" relationship, which Alice experiences as time. But, then,
we are saying that wherever a logical relationship exists that is
time-like, there is subjective time for those inside the time-like
relationship.

Now, what might constitute a time-like relationship? I see several
alternatives, but none seem satisfactory.

At any given moment, all we can be directly aware of is that one
moment. If we remember the past, that is because at the present moment
our brain has those memories; we don't know if they "really" came from
the past. What would it mean to put moments in a series? It changes
nothing essential about the moment itself; we can remove the past,
because it adds nothing.

The connection between moments doesn't seem like a physical
connection; the notion is non-explanatory, since if there were such a
physical connection we could remove it without altering the individual
moments, therefore not altering our memories, and our subjective
experience of time. Similarly, can it be a logical relationship? Is it
the structure of a single moment that connects it to the next? How
would this be? Perhaps we require that there is some function (a
"physics") from one moment to the next? But, this does not exactly
allow for things like relativity in which there is no single universal
clock. Of course, relativity could be simulated, creating a universe
that was run be a universal clock but whose internal facts did not
depend on which universal clock, exactly, the simulation was run from.
My problem is, I suppose, that any particular definition of "timelike
relationship" seems too arbitrary. As another example, should any
probabilistic elements be allowed into physics? In this case, we don't
have a function any more, but a relation-- perhaps a relation of
weighted transitions. But how would this relation make any difference
from inside the universe?

--Abram

Brent Meeker

unread,
Nov 27, 2008, 2:38:03 PM11/27/08
to everyth...@googlegroups.com
Abram Demski wrote:
> Bruno,
>
> It seems to me that this runs head-on into the problem of the
> definition of time...
>
> Here is my argument; I am sure there will be disagreement with it.
>
> Supposing that Alice's consciousness is spread out over the movie
> billboards next to the train track, there is no longer a normal
> temporal relationship between mental moments. There must merely be a
> "time-like" relationship, which Alice experiences as time. But, then,
> we are saying that wherever a logical relationship exists that is
> time-like, there is subjective time for those inside the time-like
> relationship.
>
> Now, what might constitute a time-like relationship? I see several
> alternatives, but none seem satisfactory.
>
> At any given moment, all we can be directly aware of is that one
> moment. If we remember the past, that is because at the present moment
> our brain has those memories; we don't know if they "really" came from
> the past. What would it mean to put moments in a series? It changes
> nothing essential about the moment itself; we can remove the past,
> because it adds nothing.

You raise some good points. I think the crux of the problem comes from chopping
a process up into "moments" and assuming that these infinitesimal, frozen slices
preserve all that is necessary for time. It is essentially the same as assuming
there is a "subsitution level" below which we can ignore causality and just talk
about states. It seems like a obvious idea, but it is contrary to quantum
mechanics and unitary evolution under the Schrodinger equation which was the
basis for the whole idea of a multiverse and "everything happens".


>
> The connection between moments doesn't seem like a physical
> connection; the notion is non-explanatory, since if there were such a
> physical connection we could remove it without altering the individual
> moments, therefore not altering our memories, and our subjective
> experience of time.

How do we know that? Memories and brain processes are distributed and parallel,
which means there are spacelike separated parts of the process - and neural
signals are orders of magnitude slower than light.

Brent

Bruno Marchal

unread,
Nov 27, 2008, 3:37:56 PM11/27/08
to everyth...@googlegroups.com
Hi Abram,


On 27 Nov 2008, at 20:02, Abram Demski wrote:


Bruno,

It seems to me that this runs head-on into the problem of the
definition of time...


Here is my argument; I am sure there will be disagreement with it.

Supposing that Alice's consciousness is spread out over the movie
billboards next to the train track, there is no longer a normal
temporal relationship between mental moments. There must merely be a
"time-like" relationship, which Alice experiences as time.


Assuming MAT (and MEC). OK.



But, then,
we are saying that wherever a logical relationship exists that is
time-like, there is subjective time for those inside the time-like
relationship.


With MAT,  (in a block view or not). OK.





Now, what might constitute a time-like relationship? I see several
alternatives, but none seem satisfactory.

At any given moment, all we can be directly aware of is that one
moment. If we remember the past, that is because at the present moment
our brain has those memories; we don't know if they "really" came from
the past. What would it mean to put moments in a series? It changes
nothing essential about the moment itself; we can remove the past,
because it adds nothing.


With MAT! As Stathis said, once consciousness supervenes on mathematical computations (up to some equivalence class), you can no more remove a past than you can remove the number 13 from elementary arithmetic.





The connection between moments doesn't seem like a physical
connection; the notion is non-explanatory, since if there were such a
physical connection we could remove it without altering the individual
moments, therefore not altering our memories, and our subjective
experience of time.


All right.




Similarly, can it be a logical relationship? Is it
the structure of a single moment that connects it to the next? How
would this be? Perhaps we require that there is some function (a
"physics") from one moment to the next?


It is here that we have to take into account what I think to be a major discovery, if not THE first big(**)  discovery of humanity:  the discovery of the (mathematical) Universal Machine (of Turing, if you want, but with Church thesis it is considerably more general).

That concept makes it possible to define a "space" of all computations. It is defined unambiguously by the Universal Dovetailer, a program itself. Or equivalently by the set of sigma_1 sentences and their proofs (note that the false sigma_1 sentences inherit infinite proofs).

Now, consider some trace of that program. It executes all computations. But (see the UDA), from our personal point of view, even between just two distinguishable computational states, there will be a continuum of computations going through those states, if only due to the "dumbness" of the UD, who re-execute each step of any computation by dovetailing them on all "real" oracle. And MEC predict that this "mutiplication" of computations appearing below our substitution level is indirectly observable.




But, this does not exactly
allow for things like relativity in which there is no single universal
clock.


(Sure, but we were trying to suppose here only MEC + MAT (not MEC + MAT + relativity). And this points to a little weakness of the second argument I gave.
But I would find only funny to just conclude that the falsity of relativity would safe "MAT".  That would be a type of ad hoc move I was hoping to exclude in some dreamy "MGA 5" some posts ago.



Of course, relativity could be simulated, creating a universe
that was run be a universal clock but whose internal facts did not
depend on which universal clock, exactly, the simulation was run from.


You have to read the UDA here I'm afraid: no "emulable physical reality" can save the problem. With MEC, "physical reality" has to emerge from ALL emulable histories, and this, as viewed by any universal machine, is really an infinite sum. That sum is hardly emulable, a priori, but I agree that quantum mechanics seems to show that such a sum is perhaps emulable (albeit with exponential slow down) by the sharable indeterminate part making our probable sharable neighborhood. 
MEC has everything to justify it, or to refute it (in which case the success of quantum computing would confirm or refute MEC). Up to now, QM confirms MEC (I can say more precise technical things here if interested).




My problem is, I suppose, that any particular definition of "timelike
relationship" seems too arbitrary.


There is a big difference between first person non sharable time, and sharable local (clock measurable) time. The first you experience, the second you guess, and you guess it only from an implicit bet on your own 


consistency. It makes a big "modal" difference.




As another example, should any
probabilistic elements be allowed into physics? In this case, we don't
have a function any more, but a relation-- perhaps a relation of
weighted transitions. But how would this relation make any difference
from inside the universe?



We are supported by infinity(*) of computations.  We can only bet on our most probable "histories", above our level of constitution. Those historie which can multiplie themselves from below, and thus in front of pure probabilistic event (noise) can win the measure game (on the computations or the OMs). The question is: can we explain from MEC, as we have too, why, as we can see empirically, the "probabilities" can also be subtracted. To we get here too classical mechanics in the limit? Open problem of course. 

Now the crazy thing is that we can already (thanks to Gödel, Löb, Solovay ...) interview a (Lobian) universal machine on that subject, and she gives a shadow of reason (and guess!) why indeed subtraction occurs. And thanks to the Solovay split between G (the provable part of self-reference, and G*, the true but unprovable part of self-reference, some intensional variant of G and G* split temselves into the sharable physics (indeteminate quanta) and unsharable physics (the qualia? the perceptible field, what you can only be the one to confim: a bit like being the one in Moscow after a self-duplication experiment).


Bruno


(*)  Even infinitIES, from the "third person point of view on the first person points of view. Hmmm do you know the first person comp indeterminacy? (step 3 of UDA).
(**) The second BIG discovery being the quantum computer !  (don't hesitate to use grain salts if it helps to swallows what I say). of course "nature" made those discoveries before us. Well, with MEC we have to consider that elementary arithmetic did those "discoveries" even out of time and space.

Abram Demski

unread,
Nov 27, 2008, 5:06:22 PM11/27/08
to everyth...@googlegroups.com
Bruno,

So, basically, you are saying that I'm offering an alternative
argument against materialism, correct? Supposing that reality has a
purely mathematical basis eliminates the problem, because removing the
past is like removing the number 13.

You say that the argument by counterfactuals won't work, thanks to
Olympization... I don't know what Olympization is, so I'd like to hear
the argument.

As for the remainder of your comments, I suppose I'll just have to go read UDA.

--Abram

Abram Demski

unread,
Nov 27, 2008, 5:22:40 PM11/27/08
to everyth...@googlegroups.com
Brent,

>> The connection between moments doesn't seem like a physical
>> connection; the notion is non-explanatory, since if there were such a
>> physical connection we could remove it without altering the individual
>> moments, therefore not altering our memories, and our subjective
>> experience of time.
>
> How do we know that? Memories and brain processes are distributed and parallel,
> which means there are spacelike separated parts of the process - and neural
> signals are orders of magnitude slower than light.

You are right-- the truth is that subjective experience occurs over
time. My point was really that supposing a physical connection doesn't
actually answer the question... it seems like it is equivalent to
simply taking "time" as a notion that cannot be defined in terms of
more basic terms.

--Abram

Bruno Marchal

unread,
Nov 28, 2008, 2:24:39 PM11/28/08
to everyth...@googlegroups.com

Hi Abram,

>
> Bruno,
>
> So, basically, you are saying that I'm offering an alternative
> argument against materialism, correct?

It seems to me you were going in that direction, yes.


> Supposing that reality has a
> purely mathematical basis eliminates the problem, because removing the
> past is like removing the number 13.

In the third person point of view of the "everything", or the Big
Thing, the "whole", the utimate of what-need-to-be. We cannot know
what that is, but if we assume we are (hopefully sound) lobian machine
(like PA and ZF to give two examples) then, well, ontologically, we
don't need more than arithmetic. If you are a mathematical logician I
would say "the standard model of PA". A non constructive object, which
denotes nothing more than what is supposed to be teached in school.
But we are forever prevented to prove this, if true. We can say it,
because we assume (at a more basic level) that the comp hypothesis
belongs itself in the non provable part. (Of course this is related to
the fact that we can never prove we will remain alive even a
nanosecond even when not using teleportation).
So the ontological part can be well described by the "usual" relation
between number using "0", the successor relation, with addition, and
multiplication.
But now, you could equivalently take, for the ontological part, any
programming language, or more precisely, a first-order specification
of any universal system, would do for defining a minimal ontology. (I
tell you a secret, when I say number I often personally think: a
combinator, I have send some post on that).
For the theoretical purpose here the exact choice does not matter, but
depending how exactly physics arise from the numbers, the choice of
some better and better suited minimal ontology can be addressed.

Now, a first order presentation of a specification of a universal
machine + the induction axioms = a universal machine which knows she
is universal. I call them Löbian machines. And this makes her somehow
infinitely modest: the Löb formula is really a modesty axiom.

With the MEC hypothesis, a "believer" in comp "go to hell". (Where a
"believer"in , is someone who takes p for granted).
Comp, is like self-consistency, a self-observing machine can guess it,
hope it, (or fear it), but can never take it for granted. It *is*
theological. No machine can prove its theology, but Löbian machine can
study the complete theology of more simple Löbian machines, find the
invariant for the consistent extensions, and lift it to themselves,
keeping consistency by "consciously" being aware that this is has to
be taken as an interrogation, it is not for granted, so that saying
"yes" to the doctor needs an act of faith, and never can be imposed.
(Of course we can argue biology has already "bet" on it).

The big event is Solovay's solution of the problem of axiomatizing
(the propositional part) of the self-referentially correct logic of
consistency and provability: G for the prouvable part, and G* for the
true part. (and thus G* \minus G for the unprovable-but-true part). It
is bridge between UDA and AUDA. In UDA I ask you a question, and in
AUDA I ask a universal machine about the same question.


>
>
> You say that the argument by counterfactuals won't work, thanks to
> Olympization... I don't know what Olympization is, so I'd like to hear
> the argument.




Tim Maudlin has found independently a related argument that MEC and
MAT (the supervenience thesis) are incompatible.
Maudlin's paper begins with a quotation from Hoffmann's "The Sandman",
a novel in which Nathanael felt in love with Olympia, just for
discovering latter she was an automata (and he suicides himself after).
Maudlin shows that for a special computation, which supports in time
some consciousness (by using the (physical) supervenience thesis), you
can build a device doing the same computation with much less physical
activity, actually with almost no physical activity at all. The
natural reply is that such a machine has no more the right
counterfactual behavior. Then Maudlin shows that you can render the
counterfactual correctness to such machine by adding, what will be for
the special computation, just inert material.
But this give to inert material something which plays no role, or
would give prescience to elementary material in computations; from
which you can conclude that MEC and MAT does not works well together.
At the end of his paper, Maudlin still try to add something to MEC and
he does not think about abandoning MAT. He is aware nevertheless that
MEC cannot be recovered by adding things to it, I think.


>
>
> As for the remainder of your comments, I suppose I'll just have to
> go read UDA.

Yes. But I see Kim Jones want explanation. If Kim really asks I will
propose him to do it, little step by little step, perhaps. It could be
annoying for those who got already the point though.
A problem is that a part of me estimates that to give the real thing
we have to begin by step seven. To explain what is a universal
dovetailer. And Church's thesis.

Abram, are you aware that Godel's incompleteness follows "easily" (=
in few lines) from Church thesis? Not the second theorem, but the
first, even a stronger form of the first.

Anna?


Bruno
http://iridia.ulb.ac.be/~marchal/



Abram Demski

unread,
Nov 28, 2008, 5:20:12 PM11/28/08
to everyth...@googlegroups.com
Hi Bruno,

>> So, basically, you are saying that I'm offering an alternative
>> argument against materialism, correct?
>
> It seems to me you were going in that direction, yes.
>

Well, *I* was suggesting that we run up against the problem of time in
*either* direction (physical reality / mathematical reality); so the
real problem would be a naive view of time, rather than COMP + MAT.
But, you are probably right: the problem really only applies to MAT.
On the other hand, I might try to take up the argument again after
reading UDA. :)

> With the MEC hypothesis, a "believer" in comp "go to hell". (Where a
> "believer"in , is someone who takes p for granted).
> Comp, is like self-consistency, a self-observing machine can guess it,
> hope it, (or fear it), but can never take it for granted. It *is*
> theological. No machine can prove its theology, but Löbian machine can
> study the complete theology of more simple Löbian machines, find the
> invariant for the consistent extensions, and lift it to themselves,
> keeping consistency by "consciously" being aware that this is has to
> be taken as an interrogation, it is not for granted, so that saying
> "yes" to the doctor needs an act of faith, and never can be imposed.
> (Of course we can argue biology has already "bet" on it).

Yes, this is fundamentally interesting :).

> Maudlin shows that for a special computation, which supports in time
> some consciousness (by using the (physical) supervenience thesis), you
> can build a device doing the same computation with much less physical
> activity, actually with almost no physical activity at all. The
> natural reply is that such a machine has no more the right
> counterfactual behavior. Then Maudlin shows that you can render the
> counterfactual correctness to such machine by adding, what will be for
> the special computation, just inert material.
> But this give to inert material something which plays no role, or
> would give prescience to elementary material in computations; from
> which you can conclude that MEC and MAT does not works well together.

I am not sure this convinces me. If the "inert" material is useful to
the computation in the counterfactual situations, then it is useful,
cannot be removed.


> Abram, are you aware that Godel's incompleteness follows "easily" (=
> in few lines) from Church thesis? Not the second theorem, but the
> first, even a stronger form of the first.

No, I do not know that one.

>> --Abram

Russell Standish

unread,
Nov 28, 2008, 4:46:47 AM11/28/08
to everyth...@googlegroups.com
On Wed, Nov 26, 2008 at 10:09:01AM +0100, Bruno Marchal wrote:
> MGA 3

...

> But this reasoning goes through if we make the hole in the film
> itself. Reconsider the image on the screen: with a hole in the film
> itself, you get a "hole" in the movie, but everything which enters and
> go out of the hole remains the same, for that (unique) range of
> activity. The "hole" has trivially the same functionality than the
> subgraph functionality whose special behavior was described by the
> film. And this is true for any subparts, so we can remove the entire
> film itself.
>

I don't think this step follows at all. Consciousness may supervene on
the stationary unprojected film, but if you start making holes in it,
you will eventually get a film on a nonconscious entity. At some
point, the consciousness is no longer supervening on the film (but may
well be supervening on other films that haven't been so adulterated,
or on running machines or whatever...

> Does Alice's dream supervene (in real time and space) on the
> projection of the empty movie?
>

No.

>
> 2)
>
> I give now what is perhaps a simpler argument
>
> A projection of a movie is a relative phenomenon. On the planet 247a,
> nearby in the galaxy, they don't have screen. The film pellicle is as
> big as a screen, and they make the film passing behind a stroboscope
> at the right frequency in front of the public. But on planet 247b,
> movies are only for travellers! They dress their film, as big as those
> on planet 247a, in their countries all along their train rails with a
> lamp besides each frames, which is nice because from the train,
> through its speed, you get the usual 24 frames per second. But we
> already accepted that such movie does not need to be observed, the
> train can be empty of people. Well the train does not play any role,
> and what remains is the static film with a lamp behind each frame. Are
> the lamps really necessaries? Of course not, all right? So now we are
> obliged to accept that the consciousness of Alice during the
> projection of the movie supervenes of something completely inert in
> time and space. This contradicts the *physical* supervenience thesis.
>

But the physics that Alice experiences will be fully dynamic. She will
experience time, and non-inert processes that she is supervening on.

Why does the physical supervenience require that all instantiations of
a consciousness be dynamic? Surely, it suffices that some are?

>
> c) Eliminate the hypothesis "there is a concrete deployment" in the
> seventh step of the UDA. Use UDA(1...7) to define properly the
> computationalist supervenience thesis. Hint: reread the remarks above.

I have no problems with this conclusion. However, we cannot eliminate
supervenience on phenomenal physics, n'est-ce pas?


--

----------------------------------------------------------------------------
A/Prof Russell Standish Phone 0425 253119 (mobile)
Mathematics
UNSW SYDNEY 2052 hpc...@hpcoders.com.au
Australia http://www.hpcoders.com.au
----------------------------------------------------------------------------

Abram Demski

unread,
Nov 28, 2008, 10:49:09 PM11/28/08
to everyth...@googlegroups.com
Bruno,

I have done some thinking, and decided that I don't think this last
step of the argument works for me. You provided two arguments, and so
I provide two refutations.

1. (argument by removal of unnecessary parts): Suppose Alice lives in
a cave all her life, with bread and water tossed down keeping her
alive, but nobody ever checking to see that she eats it; to the
outside world, she is functionally unnecessary. But from Alice's point
of view, she is not functionally removable, nor are the other things
in the cave that the outside world knows nothing about. The point is,
we need to be careful about labeling things functionally removable; we
need to ask "from whose perspective?". A believer in MAT who accepted
the consciousness of the movie could claim that such an error is being
made.

2. (argument by spreading movie in space instead of time): Here I need
to go back further in the argument... I still think the objection
about hypotheticals (ie counterfactuals) works just fine. :)

--Abram

On Wed, Nov 26, 2008 at 4:09 AM, Bruno Marchal <mar...@ulb.ac.be> wrote:

Bruno Marchal

unread,
Nov 29, 2008, 3:45:44 AM11/29/08
to everyth...@googlegroups.com
Abram,

On 29 Nov 2008, at 04:49, Abram Demski wrote:

>
> Bruno,
>
> I have done some thinking, and decided that I don't think this last
> step of the argument works for me. You provided two arguments, and so
> I provide two refutations.
>
> 1. (argument by removal of unnecessary parts): Suppose Alice lives in
> a cave all her life, with bread and water tossed down keeping her
> alive, but nobody ever checking to see that she eats it; to the
> outside world, she is functionally unnecessary. But from Alice's point
> of view, she is not functionally removable, nor are the other things
> in the cave that the outside world knows nothing about. The point is,
> we need to be careful about labeling things functionally removable; we
> need to ask "from whose perspective?". A believer in MAT who accepted
> the consciousness of the movie could claim that such an error is being
> made.


The argument was more of the type : "removal of unnecessay and
unconscious or unintelligent parts. Those parts have just no
perspective. If they have some perpective playing arole in Alice's
consciousness, it would mean we have not well chosen the substitution
level. You are reintroducing some consciousness on the elementary
parts, here, I think.

>
>
> 2. (argument by spreading movie in space instead of time): Here I need
> to go back further in the argument... I still think the objection
> about hypotheticals (ie counterfactuals) works just fine. :)


Then you think that if someone is conscious with some brain, which for
some reason, does never use some neurons, could loose consciousness
when that never used neuron is removed?
If that were true, how could still be confident with an artificial
digital brain. You may be right, but the MEC hypothesis would be put
in doubt.

Bruno

http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Nov 29, 2008, 3:59:26 AM11/29/08
to everyth...@googlegroups.com


Yes but with MAT, the inert material has no use in the particular
instantiation we have chosen. If it play a role, it cannot be in
virtue of the MEC hypothesis *together* with the MAT hypothesis. If
not, it means you already make consciousness supervening on the
abstract computation the pieces of materials instantiate
"accidentally" here and now, not really on the physical process
implementing that computation.
Feel free to criticize ....

>
>
>
>> Abram, are you aware that Godel's incompleteness follows "easily" (=
>> in few lines) from Church thesis? Not the second theorem, but the
>> first, even a stronger form of the first.
>
> No, I do not know that one.


I will have the occasion to explain if I decide to make the UDA
beginning by step seven.


Bruno
http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Nov 29, 2008, 4:11:30 AM11/29/08
to everyth...@googlegroups.com

On 28 Nov 2008, at 10:46, Russell Standish wrote:

>
> On Wed, Nov 26, 2008 at 10:09:01AM +0100, Bruno Marchal wrote:
>> MGA 3
>
> ...
>
>> But this reasoning goes through if we make the hole in the film
>> itself. Reconsider the image on the screen: with a hole in the film
>> itself, you get a "hole" in the movie, but everything which enters
>> and
>> go out of the hole remains the same, for that (unique) range of
>> activity. The "hole" has trivially the same functionality than the
>> subgraph functionality whose special behavior was described by the
>> film. And this is true for any subparts, so we can remove the entire
>> film itself.
>>
>
> I don't think this step follows at all. Consciousness may supervene on
> the stationary unprojected film,

This, I don't understand. And, btw, if that is true, then the physical
supervenience thesis is already wrong. The
physical supervenience thesis asks that consciousness is associated in
real time and space with the activity of some machine (with MEC).


What do you mean by an instantiation of a dynamical process which is
not dynamic. Even a block universe describe a dynamical process, or a
variety of dynamical processes.


>
>
>>
>> c) Eliminate the hypothesis "there is a concrete deployment" in the
>> seventh step of the UDA. Use UDA(1...7) to define properly the
>> computationalist supervenience thesis. Hint: reread the remarks
>> above.
>
> I have no problems with this conclusion. However, we cannot eliminate
> supervenience on phenomenal physics, n'est-ce pas?

We cannot eliminate supervenience of consciousness on what we take as
other persons indeed. Of course phenomenal physics is a first person
subjective creation, and it helps to entangle our (abstract)
computational histories. That is the role of a "brain". It does not
create consciousness, it does only make higher the probability for
that consciousness to be able to manifest itself relatively to "other
consciousness". But consciousness can rely, with MEC, only to the
abstract computation.

Sorry for being a bit short, I have to go,

Bruno


http://iridia.ulb.ac.be/~marchal/

Abram Demski

unread,
Nov 29, 2008, 9:56:07 AM11/29/08
to everyth...@googlegroups.com
Bruno,

> The argument was more of the type : "removal of unnecessay and
> unconscious or unintelligent parts. Those parts have just no
> perspective. If they have some perpective playing arole in Alice's
> consciousness, it would mean we have not well chosen the substitution
> level. You are reintroducing some consciousness on the elementary
> parts, here, I think.
>

The problem would not be with removing individual elementary parts and
replacing them with functionally equivalent pieces; this obviously
preserves the whole. Rather with removing whole subgraphs and
replacing them with equivalent pieces. As Alice-in-the-cave is
supposed to show, this can remove consciousness, at least in the limit
when the entire movie is replaced...


>
> Then you think that if someone is conscious with some brain, which for
> some reason, does never use some neurons, could loose consciousness
> when that never used neuron is removed?
> If that were true, how could still be confident with an artificial
> digital brain. You may be right, but the MEC hypothesis would be put
> in doubt.
>

I am thinking of it as being the same as someone having knowledge
which they never actually use. Suppose that the situation is so
extreme that if we removed the neurons involved in that knowledge, we
will not alter the person's behavior; yet, we will have removed the
knowledge. Similarly, if the behavior of Alice in practice comes from
a recording, yet a dormant conscious portion is continually ready to
intervene if needed, then removing that dormant portion removes her
consciousness.

--Abram

Brent Meeker

unread,
Nov 29, 2008, 12:49:24 PM11/29/08
to everyth...@googlegroups.com
Bruno Marchal wrote:
>
> On 28 Nov 2008, at 10:46, Russell Standish wrote:
>
>> On Wed, Nov 26, 2008 at 10:09:01AM +0100, Bruno Marchal wrote:
>>> MGA 3
>> ...
>>
>>> But this reasoning goes through if we make the hole in the film
>>> itself. Reconsider the image on the screen: with a hole in the film
>>> itself, you get a "hole" in the movie, but everything which enters
>>> and
>>> go out of the hole remains the same, for that (unique) range of
>>> activity. The "hole" has trivially the same functionality than the
>>> subgraph functionality whose special behavior was described by the
>>> film. And this is true for any subparts, so we can remove the entire
>>> film itself.
>>>
>> I don't think this step follows at all. Consciousness may supervene on
>> the stationary unprojected film,
>
> This, I don't understand. And, btw, if that is true, then the physical
> supervenience thesis is already wrong. The
> physical supervenience thesis asks that consciousness is associated in
> real time and space with the activity of some machine (with MEC).

Then assuming MEC requires some definition of "activity" and consciousness may
cease when there is no activity of the required kind.

Brent

Bruno Marchal

unread,
Nov 29, 2008, 3:12:37 PM11/29/08
to everyth...@googlegroups.com

On 29 Nov 2008, at 15:56, Abram Demski wrote:

>
> Bruno,
>
>> The argument was more of the type : "removal of unnecessay and
>> unconscious or unintelligent parts. Those parts have just no
>> perspective. If they have some perpective playing arole in Alice's
>> consciousness, it would mean we have not well chosen the substitution
>> level. You are reintroducing some consciousness on the elementary
>> parts, here, I think.
>>
>
> The problem would not be with removing individual elementary parts and
> replacing them with functionally equivalent pieces; this obviously
> preserves the whole. Rather with removing whole subgraphs and
> replacing them with equivalent pieces. As Alice-in-the-cave is
> supposed to show, this can remove consciousness, at least in the limit
> when the entire movie is replaced...


The limit is not relevant. I agrre that if you remove Alice, you
remove any possibility for Alice to manifest herself in your most
probable histories. The problem is that in the range activity of the
projected movie, removing a part of the graph change nothing. It
changes only the probability of recoevering Alice from her history in,
again, your most probable history. IThere are no physical causal link
between the experience attributed to the physical computation and the
"causal history of projecting a movie". The incremental removing of
the graph hilighted the lack of causality in the movie. Perhaps not in
the best clearer way, apparently. Perhaps I should have done the case
of a non dream. I will come back on this.


>
>
>
>>
>> Then you think that if someone is conscious with some brain, which
>> for
>> some reason, does never use some neurons, could loose consciousness
>> when that never used neuron is removed?
>> If that were true, how could still be confident with an artificial
>> digital brain. You may be right, but the MEC hypothesis would be put
>> in doubt.
>>
>
> I am thinking of it as being the same as someone having knowledge
> which they never actually use. Suppose that the situation is so
> extreme that if we removed the neurons involved in that knowledge, we
> will not alter the person's behavior; yet, we will have removed the
> knowledge. Similarly, if the behavior of Alice in practice comes from
> a recording, yet a dormant conscious portion is continually ready to
> intervene if needed, then removing that dormant portion removes her
> consciousness.


You should definitely do the removing of the graph in the non-dream
situation. Let us do it.
Let us take a situation without complex inputs. Let us imagine Alice
is giving a conference in a big room, so, as input she is just blinded
by some projector, + some noise, and she makes a talk on Astronomy (to
fix the things). Now from 8h30 to 8H45 pm, she has just no brain, she
get the "motor" info from a projected recording of a previous *perfect
dream* of that conference, dream done the night before, or send from
Platonia (possible in principle). Then, by magic, to simplify, at 8h45
she get back the original brain, which by optical means inherits the
stage at the end of the conference in that perfect dream. I ask you,
would you say Alice was a zombie, during the conference?

Bruno

http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Nov 29, 2008, 3:18:26 PM11/29/08
to everyth...@googlegroups.com

On 29 Nov 2008, at 18:49, Brent Meeker wrote:

>> This, I don't understand. And, btw, if that is true, then the
>> physical
>> supervenience thesis is already wrong. The
>> physical supervenience thesis asks that consciousness is associated
>> in
>> real time and space with the activity of some machine (with MEC).
>
> Then assuming MEC requires some definition of "activity" and
> consciousness may
> cease when there is no activity of the required kind.


We require a notion of physical activity related to a computation for
having MEC *and* the supervenience thesis.
With MEC alone, we abandon MAT, the computational supervenience thesis
will have to make any notion of physical causality a statistically
emerging pattern from (hopefully sharable) first person (plural)
points of view.

Bruno
http://iridia.ulb.ac.be/~marchal/

Abram Demski

unread,
Nov 29, 2008, 5:17:52 PM11/29/08
to everyth...@googlegroups.com
Bruno,

My answer would have to be, no, she lacks the necessary counterfactual
behaviors during that time. And, moreover, if only part of the brain
were being run by a recording then she would lack only some
counterfactuals, and so she would count as partially conscious.

--Abram

Brent Meeker

unread,
Nov 29, 2008, 10:23:17 PM11/29/08
to everyth...@googlegroups.com
Bruno Marchal wrote:
>
> On 29 Nov 2008, at 15:56, Abram Demski wrote:
>
>> Bruno,
>>
>>> The argument was more of the type : "removal of unnecessay and
>>> unconscious or unintelligent parts. Those parts have just no
>>> perspective. If they have some perpective playing arole in Alice's
>>> consciousness, it would mean we have not well chosen the substitution
>>> level. You are reintroducing some consciousness on the elementary
>>> parts, here, I think.
>>>
>> The problem would not be with removing individual elementary parts and
>> replacing them with functionally equivalent pieces; this obviously
>> preserves the whole. Rather with removing whole subgraphs and
>> replacing them with equivalent pieces. As Alice-in-the-cave is
>> supposed to show, this can remove consciousness, at least in the limit
>> when the entire movie is replaced...
>
>
> The limit is not relevant. I agrre that if you remove Alice, you
> remove any possibility for Alice to manifest herself in your most
> probable histories. The problem is that in the range activity of the
> projected movie, removing a part of the graph change nothing. It
> changes only the probability of recoevering Alice from her history in,
> again, your most probable history.

Isn't this reliance on probable histories assuming some physical theory that is
no in evidence?

>IThere are no physical causal link
> between the experience attributed to the physical computation and the
> "causal history of projecting a movie".

But there is a causal history for the creation of the movie - it's a recording
of Alice's brain functions which were causally related to her physical world.

>The incremental removing of
> the graph hilighted the lack of causality in the movie.

It seems to me there is still a causal chain - it is indirected via creating the
movie.

Brent

Russell Standish

unread,
Nov 30, 2008, 5:57:36 AM11/30/08
to everyth...@googlegroups.com
On Sat, Nov 29, 2008 at 10:11:30AM +0100, Bruno Marchal wrote:
>
>
> On 28 Nov 2008, at 10:46, Russell Standish wrote:
>
> >
> > On Wed, Nov 26, 2008 at 10:09:01AM +0100, Bruno Marchal wrote:
> >> MGA 3
> >
> > ...
> >
> >> But this reasoning goes through if we make the hole in the film
> >> itself. Reconsider the image on the screen: with a hole in the film
> >> itself, you get a "hole" in the movie, but everything which enters
> >> and
> >> go out of the hole remains the same, for that (unique) range of
> >> activity. The "hole" has trivially the same functionality than the
> >> subgraph functionality whose special behavior was described by the
> >> film. And this is true for any subparts, so we can remove the entire
> >> film itself.
> >>
> >
> > I don't think this step follows at all. Consciousness may supervene on
> > the stationary unprojected film,
>
> This, I don't understand. And, btw, if that is true, then the physical
> supervenience thesis is already wrong. The
> physical supervenience thesis asks that consciousness is associated in
> real time and space with the activity of some machine (with MEC).

I am speaking as someone unconvinced that MGA2 implies an
absurdity. MGA2 implies that the consciousness is supervening on the
stationary film.

BTW - I don't think the film is conscious by virtue of the
counterfactuals issue, but that's a whole different story. And
"Olympization" doesn't work, unless we rule out the multiverse.

> >
> > Why does the physical supervenience require that all instantiations of
> > a consciousness be dynamic? Surely, it suffices that some are?
>
>
> What do you mean by an instantiation of a dynamical process which is
> not dynamic. Even a block universe describe a dynamical process, or a
> variety of dynamical processes.
>

A block universe is nondynamic by definition. But looked at another
way, (ie from the inside) it is dynamic. It neatly illustrates why
consciousness can supervene on a stationary film (because it is
stationary when viewed from the inside). The "film", however does need
to be sufficiently rich, and also needs to handle counterfactuals
(unlike the usual sort of movie we see which has only one plot).

>
> >
> >
> >>
> >> c) Eliminate the hypothesis "there is a concrete deployment" in the
> >> seventh step of the UDA. Use UDA(1...7) to define properly the
> >> computationalist supervenience thesis. Hint: reread the remarks
> >> above.
> >
> > I have no problems with this conclusion. However, we cannot eliminate
> > supervenience on phenomenal physics, n'est-ce pas?
>
> We cannot eliminate supervenience of consciousness on what we take as
> other persons indeed. Of course phenomenal physics is a first person
> subjective creation, and it helps to entangle our (abstract)
> computational histories. That is the role of a "brain". It does not
> create consciousness, it does only make higher the probability for
> that consciousness to be able to manifest itself relatively to "other
> consciousness". But consciousness can rely, with MEC, only to the
> abstract computation.
>

The problem is that eliminating the brain from phenomenal experience
makes that experience even more highly probable than without. This is
the Occam catastrophe I mention in my book. Obviously this contradicts
experience.

Therefore I conclude that supervenience on a phenomenal physical brain
is necessary for consciousness. I speculate a bit that this may be due
to self-awareness, but don't have a good argument for it. It is the
"elephant in the room" with respect to pure MEC theories.

> Sorry for being a bit short, I have to go,
>
> Bruno
>
>
>
>
> http://iridia.ulb.ac.be/~marchal/
>
>
>
>
>

Bruno Marchal

unread,
Nov 30, 2008, 12:33:23 PM11/30/08
to everyth...@googlegroups.com

Not at all. I have defined "history" by a computation as see from a
first person (plural or not).
Of course, well I guess I should insist on that perhaps, by
computation I always mean the mathematical object; It makes sense only
with respect to to some universal machine, and I have chosen
elementary arithmetic as the primitive one.

Although strictly speaking the notion of computable is an epistemic
notion, it happens that Church thesis makes it equivalent with purely
mathematical notion, and this is used for making the notion of
probable history a purely mathematical notion, (once we got a
mathematical notion of first person, but this is simple in the thought
experience (memory, diary ..., and a bit more subtle in the interview
(AUDA)).

A difficulty, in those post correspondences, is that I am reasoning
currently with MEC and MAT, just to get the contradiction, but in many
(most) posts I reason only with MEC (having abandon MAT).
After UDA, you can already understand that "physical" has to be
equivalent with "probable history" for those who followed the whole
UDA+MGA. "physical" has to refer the most probable (and hopefully)
sharable relative computational history.
This is already the case with just UDA, if you assume both the
existence of a "physical universe" and of a concrete UD running in
that concrete universe. MGA is designed to eliminate the assumption of
"a physical universe" and of the "concrete UD".


>
>
>> IThere are no physical causal link
>> between the experience attributed to the physical computation and the
>> "causal history of projecting a movie".
>
> But there is a causal history for the creation of the movie - it's a
> recording
> of Alice's brain functions which were causally related to her
> physical world.

Assuming MEC+MAT you are right indeed. But the causal history of the
creation of the movie, is not the same "computation" or causal chain
than the execution of Alice's mind and Alice's brain during her
"original dream". If you make abstraction of that difference, it means
you already don't accept the physical supervenience thesis, or, again,
you are introducing "magical knowledge" in the elementary part running
the computation.
You can only forget the difference of those two "computations" by
abstracting from the physical part of the story. This means you are
using exclusively the computational supervenience. MGA should make
clear (but OK, I warned MGA is subtle) that the consciousness has to
be related to the genuine causality or history. But it is that very
genuineness that physics can accidentally reproduced in a non genuine
way, like the brain movie projection, making the physical
supervenience absurd.

It seems to me quasi obvious that it is ridiculous to attribute
consciousness to the physical events of projecting the movie of a
brain. That movie gives a pretty detailed description of the
computations, but there is just no computation, nor even genuine
causal relation between the states. Even one frame is not a genuine
physical computational states. Only a relative description of it. In a
cartoon, if you see someone throwing a ball on a window, the
description of the broken glass are not caused by the description of
someone throwing a ball. And nothing changes, for the moment of the
projection of the movie, if the cartoon has been made from a real
similar filmed situation.
To attribute consciousness to the stationary (non projected)
contradict immediately the supervenience thesis of course.

All this is a bit complex because we have to take well into account
the distinction between

A computation in the "real" world,
A description of a computation in the "real" world,

And then most importantly:

A computation in Platonia
A description of a computation in Platonia.

I argue that consciousness supervenes on computation in Platonia. Even
in Platonia consciousness does not supervene on description of the
computation, even if those description are 100% precise and correct

I said a long time ago that a counting algorithm is not a universal
dovetailer, despite the fact that a counting algorithm can be said to
generate all the description of all computations (meaning also that
natural numbers and the successor function, but without addition and
multiplication law, are not enough). A UD does compute, in the
arithmetical platonia, it does not just enumerate the description of
the computations.

It is subtle, a bit like the difference between "A implies B", and
"the deduction of A from B" in Logic. In the interview of the Lobian
machine, those distinctions remains subtle but at least can be made
mathematically transparent. (But this is not needed to understand the
UDA-MGA or UDA(1...8).


>
>
>> The incremental removing of
>> the graph hilighted the lack of causality in the movie.
>
> It seems to me there is still a causal chain - it is indirected via
> creating the
> movie.

OK, but you cannot say that Alice is conscious from 3h30 to 3h45 by
supervening on an event which is "another story" than the computation
supposed to be done by the graph for that experience. The
"indirectness" you mention is enough to discredit the physical
supervenience thesis.

Bruno
http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Nov 30, 2008, 12:51:17 PM11/30/08
to everyth...@googlegroups.com
Abram,


My answer would have to be, no, she lacks the necessary counterfactual
behaviors during that time.

? The film of the graph lacks also the counterfactuals. 



And, moreover, if only part of the brain
were being run by a recording


... which lacks the counterfactual, ...


then she would lack only some
counterfactuals,


I don't understand. The recording lacks all the counterfactuals. You can recover them from inert material, true, but this is true for the empty graph too (both in dream and awake situations).



and so she would count as partially conscious.


Hmmm .... Can she be conscious that she is partially conscious? I mean is it like after we drink alcohol or something?


Bruno



Günther Greindl

unread,
Nov 30, 2008, 12:53:47 PM11/30/08
to everyth...@googlegroups.com
Hi all,

Bruno, do you still keep a notion of causality and the likes in
platonia? I have collected these snips from some recent posts:

Brent Meeker wrote:

>But is causality an implementation detail? There seems to be an
>implicit
>assumption that digitally represented states form a sequence just
>because there
>is a rule that defines that sequence, but in fact all digital (and
>other)
>sequences depend on causal chains.

Kory wrote:

> I have an intuition that causality
>(or its logical equivalent in Platonia) is somehow important for
>consciousness. You argue that the the slide from Fully-Functional
>Alice to Lucky Alice (or Fully-Functional Firefox to Lucky Firefox)
>indicates that there's something wrong with this idea. However, you
>have an intuition that order is somehow important for consciousness.

But we must realise that causality is a concept that is deeply related
(cognitively, in humans) to time and physical change.

But both time and space _emerge_ only from the inside view (1st person
or 1st person shareable) in the sum over all computations.

In Platonia (viewed, for the time being, ludicrously and impossibly,
from the outside) - there is no notion of time, space, sequentiality,
before and after.

The very notion of causation must be one that arises only in the inside
view, as a "succession" of consistent patterns.

In a sense, order (shareable histories) must arise from the Platonic
Eternal Mess (chaos) -> somehow along the lines of self-organization maybe:
http://en.wikipedia.org/wiki/Self-organization#Self-organization_in_mathematics_and_computer_science

In this sense, the computations would "assemble themselves" to
"consistent histories".

Bruno said:
>Even
>in Platonia consciousness does not supervene on description of the
>computation, even if those description are 100% precise and correct

Hmm, I understand the difference between description and computation in
maths and logic, and also in real world, but I do not know if this still
makes sense in Platonia -> viewed from the acausal perspective outlined
above. Well maybe in the sense that in some histories there will be
platonic descriptions that are not conscious.

But in other histories those descriptions will be computations and
conscious.

Cheers,
Günther

Bruno Marchal

unread,
Nov 30, 2008, 1:10:43 PM11/30/08
to everyth...@googlegroups.com
On 30 Nov 2008, at 11:57, Russell Standish wrote:


On Sat, Nov 29, 2008 at 10:11:30AM +0100, Bruno Marchal wrote:


On 28 Nov 2008, at 10:46, Russell Standish wrote:


On Wed, Nov 26, 2008 at 10:09:01AM +0100, Bruno Marchal wrote:
MGA 3

...

But this reasoning goes through if we make the hole in the film
itself. Reconsider the image on the screen: with a hole in the film
itself, you get a "hole" in the movie, but everything which enters  
and
go out of the hole remains the same, for that (unique) range of
activity.  The "hole" has trivially the same functionality than the
subgraph functionality whose special behavior was described by the
film. And this is true for any subparts, so we can remove the entire
film itself.


I don't think this step follows at all. Consciousness may supervene on
the stationary unprojected film,

This, I don't understand. And, btw, if that is true, then the physical  
supervenience thesis is already wrong. The
physical supervenience thesis asks that consciousness is associated in  
real time and space with the activity of some machine (with MEC).

I am speaking as someone unconvinced that MGA2 implies an
absurdity. MGA2 implies that the consciousness is supervening on the
stationary film.


?  I could agree, but is this not absurd enough, given MEC and the definition of the physical superveneience thesis;




BTW - I don't think the film is conscious by virtue of the
counterfactuals issue, but that's a whole different story. And
"Olympization" doesn't work, unless we rule out the multiverse.


Why does the physical supervenience require that all instantiations of
a consciousness be dynamic? Surely, it suffices that some are?


What do you mean by an instantiation of a dynamical process which is  
not dynamic. Even a block universe describe a dynamical process, or a  
variety of dynamical processes.


A block universe is nondynamic by definition. But looked at another
way, (ie from the inside) it is dynamic. It neatly illustrates why
consciousness can supervene on a stationary film (because it is
stationary when viewed from the inside).

OK, but then you clearly change the physical supervenience thesis.


The "film", however does need
to be sufficiently rich, and also needs to handle counterfactuals
(unlike the usual sort of movie we see which has only one plot).


OK. Such a film could be said to be a computation. Of course you are not talking about a stationary thing, which, be it physical or immaterial, cannot handle counterfactuals. 


The problem is that eliminating the brain from phenomenal experience
makes that experience even more highly probable than without. This is
the Occam catastrophe I mention in my book. Obviously this contradicts
experience.

Therefore I conclude that supervenience on a phenomenal physical brain
is necessary for consciousness.


It is vague enough so that I can interpret it favorably through MEC.


Bruno




I speculate a bit that this may be due
to self-awareness, but don't have a good argument for it. It is the
"elephant in the room" with respect to pure MEC theories.

Sorry for being a bit short, I have to go,

Bruno




http://iridia.ulb.ac.be/~marchal/





--

----------------------------------------------------------------------------
A/Prof Russell Standish                  Phone 0425 253119 (mobile)
Mathematics                         
UNSW SYDNEY 2052                  hpc...@hpcoders.com.au
Australia                                http://www.hpcoders.com.au
----------------------------------------------------------------------------



Günther Greindl

unread,
Nov 30, 2008, 1:14:31 PM11/30/08
to everyth...@googlegroups.com
Hello Bruno,

I must admit you have completely lost me with MGA 3.

With MGA 1 and 2, I would say that, with MEC+MAT, also the the
projection of the movie (and Lucky Alice in 1) are conscious - because
it supervenes on the physical activity.

MEC says: it's the computation that counts, not the substrate.

MAT says: we need some substrate to perform a computation. In MGA 1 and
2 we have substrates (neurons or optical boolean graph that performs the
computation).

Now in MGA 3 you say:

> Now, consider the projection of the movie of the activity of Alice's
> brain, "the movie graph".
> Is it necessary that someone look at that movie? Certainly not.

Agreed.

> Is it necessary to have a screen? Well, the range of activity here is
> just one dynamical description of one computation. Suppose we make a
> hole in the screen. What goes in and out of that hole is exactly the
> same, with the hole and without the hole. For that unique activity, the
> hole in the screen is functionally equivalent to the subgraph which the
> hole removed.

We can remove those optical boolean nodes which are not relevant for the
caterpillar dream

>Clearly we can make a hole as large as the screen, so no
> need for a screen.

but no! Then we wouldn't have a substrate anymore. You are dropping MAT
at this step, not leading MEC+MAT to a contradiction.

> But this reasoning goes through if we make the hole in the film itself.
> Reconsider the image on the screen: with a hole in the film itself, you
> get a "hole" in the movie, but everything which enters and go out of the
> hole remains the same, for that (unique) range of activity. The "hole"
> has trivially the same functionality than the subgraph functionality
> whose special behavior was described by the film. And this is true for
> any subparts, so we can remove the entire film itself.

We can talk about this part after I understand why you can drop our
optical boolean network *grin*


Cheers,
Günther

Abram Demski

unread,
Nov 30, 2008, 1:17:07 PM11/30/08
to everyth...@googlegroups.com
Bruno,

No, she cannot be conscious that she is partially conscious in this
case, because the scenario is set up such that she does everything as
if she were fully conscious-- only the counterfactuals change. But, if
someone tested those counterfactuals by doing something that the
recording didn't account for, then she may or may not become conscious
of the fact of her partial consciousness-- in that case it would be
very much like brain damage.

Anyway, yes, I am admitting that the film of the graph lacks
counterfactuals and is therefore not conscious. My earlier splitting
of the argument into an argument about (1) and a separate argument
against (2) was perhaps a bit silly, because the objection to (2) went
far enough back that it was also an objection to (1). I split the
argument like that just because I saw an independent flaw in the
reasoning of (1)... anyway...

Basically, I am claiming that there is a version of COMP+MAT that MGA
is not able to derive a contradiction from. The version goes something
like this:

"Yes, consciousness supervenes on computation, but that computation
needs to actually take place (meaning, physically). Otherwise, how
could consciousness supervene on it? Now, in order for a computation
to be physically instantiated, the physical instantiation needs to
satisfy a few properties. One of these properties is clearly some sort
of isomorphism between the computation and the physical instantiation:
the actual steps of the computation are represented in physical form.
A less obvious requirement is that the physical computation needs to
have the proper counterfactuals: if some external force were to modify
some step in the computation, the computation must progress according
to the new computational state (as translated by the isomorphism)."

--Abram

Günther Greindl

unread,
Nov 30, 2008, 1:24:18 PM11/30/08
to everyth...@googlegroups.com
Bruno,

I have reread MGA 2 and would like to add the following:

We have the

optical boolean graph: OBG -> this computes alice's dream.
we make a movie of this computation.


Now we run again, but in OBG some nodes do not make the computation
correctly, BUT the movie _triggers_ the nodes, so in the end, the
"computation" is performed.

So, with MEC+MAT and ALL NODES broken, I say this:

a) If the OBG nodes MALFUNCTION, but their function is subsituted with
the movie (on/off), it is conscious.

b) If the OBG is broken that in a way that all nodes are not active
anymore (no on/off, no signal passing), then no consciousness.

I think we can split the intuitions along these lines: if you assume
that consciousness depends on activity along the vertices, then Alice is
conscious neither in a nor in b, and then indeed I see why already MGA 2
leads to a problem with MEC+MAT.

But if I think that consciousness supervenes only on the correct
lighting up of the nodes (not the vertices!! -> I don't need causality
then, only the correct order), than a) would be conscious, b) not, and
MGA 3 does not work I you take away my OBG (with the node intuition)!

Cheers,
Günther

Brent Meeker

unread,
Nov 30, 2008, 2:10:46 PM11/30/08
to everyth...@googlegroups.com
I agree. But what is it about the patterns that creates a succession as viewed
from "the inside"? And how do we know that this does not obtain in the
projection of the MGA?

Brent

Abram Demski

unread,
Nov 30, 2008, 2:23:58 PM11/30/08
to everyth...@googlegroups.com
Günther,

This analysis is also needed for the counterfactual objection to MGA
to be fully fleshed-out... the counterfactuals on the physical level
need to correspond to counterfactuals on the platonic level,
presumably arising from a notion of causality on the platonic level.
Perhaps we could see causation in Platonia as some form of direct
logical implication. This works despite the timeless nature of
platonia... but it also introduces a dependance on the particular
axiomization used to represent a logical structure. One axiomization
of numbers might give rise to a somewhat different causal structure
than another.

--Abram

Kory Heath

unread,
Nov 30, 2008, 3:20:00 PM11/30/08
to everyth...@googlegroups.com

On Nov 30, 2008, at 9:53 AM, Günther Greindl wrote:
> Kory wrote:
>
>> I have an intuition that causality
>> (or its logical equivalent in Platonia) is somehow important for
>> consciousness. You argue that the the slide from Fully-Functional
>> Alice to Lucky Alice (or Fully-Functional Firefox to Lucky Firefox)
>> indicates that there's something wrong with this idea. However, you
>> have an intuition that order is somehow important for consciousness.
>
> But we must realise that causality is a concept that is deeply related
> (cognitively, in humans) to time and physical change.
>
> But both time and space _emerge_ only from the inside view (1st person
> or 1st person shareable) in the sum over all computations.
>
> In Platonia (viewed, for the time being, ludicrously and impossibly,
> from the outside) - there is no notion of time, space, sequentiality,
> before and after.
>
> The very notion of causation must be one that arises only in the
> inside
> view, as a "succession" of consistent patterns.

For what it's worth, I do think that that there's a *kind* of
causality in Platonia. Let me once again trot out the picture of a
platonic block universe in which the initial state is the binary
digits of PI, and the succeeding states are determined by the rules of
Conway's Life. This block universe exists unchangingly and eternally
in Platonia, but the states of the bits within it are related in a
kind of causal fashion. The state of each bit in the block is
determined (in a sense, "caused") by the pyramid of cells beneath it,
stretching back to the initial state, which is determined by the
algorithm for computing the binary digits of PI. In this sense,
causality is an essential aspect of the platonic notion of computation.

One might argue that this is really a misuse of the concept of
"causality" - that I should just talk about the necessary logical
relationships that are there "by definition" in my platonic object.
But my point is that these logical relationships fill the exact role
that "causality" is supposed to fill for the physicalist. When
patterns of bits within this platonic block universe "discuss" their
own physics, they might talk about how current configurations of
physical matter were "caused" by previous states. The logical
connections in Platonia are a good candidate for what they can
actually be talking about.

This platonic form of "causality" may not always be directly related
to the concept of time that patterns of bits in a block universe might
have. For instance, there's a cellular automaton rule (which deserves
to be much more widely known than it is) called Critters which is as
simple as Conway's Life, uses only bits (on or off), is known to be
computation universal, and is also fully reversible. This gets weird,
because the computational structures within a Critters block universe
will still seem to favor one direction in time - they'll store
memories about the "past" and try to anticipate the "future", etc. But
in fact, our own physics seems to be reversible, so we have these same
issues to work out regarding our own consciousness. The point is that,
within a Critters block universe in Platonia, the states will still be
logically related to each other in a way that precisely matches what
physicists in the block universe (the "critters within Critters"!)
would think of as causality.

-- Kory

Kory Heath

unread,
Nov 30, 2008, 3:43:16 PM11/30/08
to everyth...@googlegroups.com

On Nov 30, 2008, at 10:14 AM, Günther Greindl wrote:
> I must admit you have completely lost me with MGA 3.

I still find the whole thing easier to grasp when presented in terms
of cellular automata.

Let's say we have a computer program that starts with a large but
finite 2D grid of bits, and then iterates the rules to some CA
(Conway's Life, Critters, whatever) on that grid a large but finite
number of times, and stores all of the resulting computations in
memory, so that we have a 3D block universe in memory. And lets say
that the resulting block universe contains patterns that MECH-MAT
would say are conscious.

If we believe that consciousness supervenes on the physical act of
"playing back" the data in our block universe like a movie, then we
have a problem. Because before we play back the movie, we can fill any
portions of the block universe we want with zeros. So then our played
back movie can contain "conscious" creatures who are walking around
with (say) zeros where their visual cortexes should be, or their high-
level brain functions should be, etc. In other words, we have a fading
qualia problem (which we have also called a "partial zombie" problem
in these threads).

I find the argument compelling as far as it goes. But I'm not
convinced that all or most actual, real-world mechanist-materialists
believe that consciousness supervenes on the physical act of playing
back the stored computations. Bruno indicates that it must, by the
logical definitions of MECH and MAT. This just makes me feel like I
don't really understand the logical definitions of MECH and MAT.

-- Kory

Russell Standish

unread,
Nov 30, 2008, 9:25:14 PM11/30/08
to everyth...@googlegroups.com
On Sun, Nov 30, 2008 at 07:10:43PM +0100, Bruno Marchal wrote:
> >
> > I am speaking as someone unconvinced that MGA2 implies an
> > absurdity. MGA2 implies that the consciousness is supervening on the
> > stationary film.
>
>
> ? I could agree, but is this not absurd enough, given MEC and the
> definition of the physical superveneience thesis;

It is, prima facie, no more absurd than consciousness supervening on a
block universe.

> >
> > A block universe is nondynamic by definition. But looked at another
> > way, (ie from the inside) it is dynamic. It neatly illustrates why
> > consciousness can supervene on a stationary film (because it is
> > stationary when viewed from the inside).
>
> OK, but then you clearly change the physical supervenience thesis.
>

How so? The stationary film is a physical object, I would have thought.

>
> > The "film", however does need
> > to be sufficiently rich, and also needs to handle counterfactuals
> > (unlike the usual sort of movie we see which has only one plot).
>
>
> OK. Such a film could be said to be a computation. Of course you are
> not talking about a stationary thing, which, be it physical or
> immaterial, cannot handle counterfactuals.
>

If true, then a block universe could not represent the
Multiverse. Maybe so, but I think a lot of people might be surprised
at this one.

> >
> > The problem is that eliminating the brain from phenomenal experience
> > makes that experience even more highly probable than without. This is
> > the Occam catastrophe I mention in my book. Obviously this contradicts
> > experience.
> >
> > Therefore I conclude that supervenience on a phenomenal physical brain
> > is necessary for consciousness.
>
>
> It is vague enough so that I can interpret it favorably through MEC.
>

That is my point - physical supervenience (aka materialism) is not
only not contradicted by MEC (aka COMP), but in fact is necessary for
to even work. Only what I call naive physicalism,
(aka the need for a concrete instantiation of a computer running the
UD) is contradicted by MEC.

What _is_ interesting is that not all philosophers distinguish between
physicalism and materialism. David Chalmers does not, but Michael
Lockwood does, for instance. Much of this revolves around the
ontological status of emergence.

Bruno Marchal

unread,
Dec 1, 2008, 10:38:54 AM12/1/08
to everyth...@googlegroups.com
Hi Günther,


On 30 Nov 2008, at 18:53, Günther Greindl wrote:

>
> Hi all,
>
> Bruno, do you still keep a notion of causality and the likes in
> platonia? I have collected these snips from some recent posts:


OK, I will comment, and perhaps say more for the benefit of the
others. But in a nutshell, the simplest notion of "causality" in
Platonia is the implication. A "causes" B, if and only A is false or B
is true. I recall that "the Platonia" of Peano Arithmetic is just
arithmetical truth or the "standard model of Elementary Arithmetic,
like the Platonia of Zermelo Fraenkel set theory is (the more dubious)
Set Theoretical truth. In some context I can use deduction as a form
of Platonist causality, which, for first order Lobian Machine.
As expected it is a mathematical causality, and has a priori no
relation with physical causality ...

Then, you can consider key subset of the implication/deduction
causalities: the computational causality, for example A "causes" B if
all computations (executed by the UD) going through A are going
through B. Or things like that (they will be many variants). All
notions should be translatable in formal arithmetic (or combinators,
fortran programs, etc.) when we interview the machines in Platonia,
notably to retrieve the physical laws (or the believe in the physical
laws). When this is done we should have the comp physical notions
capable of explaining or intuitive notion of physical causality.

>
>
> Brent Meeker wrote:
>
>> But is causality an implementation detail? There seems to be an
>> implicit
>> assumption that digitally represented states form a sequence just
>> because there
>> is a rule that defines that sequence, but in fact all digital (and
>> other)
>> sequences depend on causal chains.
>
> Kory wrote:
>
>> I have an intuition that causality
>> (or its logical equivalent in Platonia) is somehow important for
>> consciousness. You argue that the the slide from Fully-Functional
>> Alice to Lucky Alice (or Fully-Functional Firefox to Lucky Firefox)
>> indicates that there's something wrong with this idea. However, you
>> have an intuition that order is somehow important for consciousness.
>
> But we must realise that causality is a concept that is deeply related
> (cognitively, in humans) to time and physical change.


I agree. Especially physical causality. But even the notion of
"responsibility" is deeply related to time (and causality).

>
>
> But both time and space _emerge_ only from the inside view (1st person
> or 1st person shareable) in the sum over all computations.


Assuming comp, and that we are correct, ok.


>
>
> In Platonia (viewed, for the time being, ludicrously and impossibly,
> from the outside)

A powerfull lobian machine like ZF can do this, looking at some
Platonia, in a precise way when reasoning on the "Platonia" of a
simpler sound Lobian Machine. (Even for the 1-Platonias, the first
person pov in Platonia (this gives the hypostases)).


> - there is no notion of time, space, sequentiality,
> before and after.


Right. But don't overlook that the number zero is before the number
one, which is itself before the number two, which is before the number
three, etc. (With "before" interpreted by minus one). The UD itself
has a first computation step, then a second, then a third, etc. But
like "a movie", you can look at all of them, well if you are
infinitely patient of course, and immortal. It is in that sense
(before making things more technical) that the UD computes in Platonia.


>
>
> The very notion of causation must be one that arises only in the
> inside
> view, as a "succession" of consistent patterns.

OK.


>
>
> In a sense, order (shareable histories) must arise from the Platonic
> Eternal Mess (chaos) -> somehow along the lines of self-organization
> maybe:
> http://en.wikipedia.org/wiki/Self-organization#Self-organization_in_mathematics_and_computer_science
>
> In this sense, the computations would "assemble themselves" to
> "consistent histories".
>
> Bruno said:
>> Even
>> in Platonia consciousness does not supervene on description of the
>> computation, even if those description are 100% precise and correct
>
> Hmm, I understand the difference between description and computation
> in
> maths and logic, and also in real world, but I do not know if this
> still
> makes sense in Platonia -> viewed from the acausal perspective
> outlined
> above. Well maybe in the sense that in some histories there will be
> platonic descriptions that are not conscious.
>
> But in other histories those descriptions will be computations and
> conscious.


A "movie" *in* Platonia, would be a description of a computation
encoded in some static way by some occasional program or entity. Even
in Platonia, such a description is not a computation, but only a
description (without any causality, even in the simple imlication/
deduction sense). It is the difference between the fact that three
added to two gives five, and the writting or the Gödel number of the
sentence "3+2=5".

It is really the difference between a reality (be it mathematical),
and a picture (be it dynamic) of that reality.

Consciousness will be related to "true mathematical facts", not on
description of those facts. The difficulty here is that consciousness
is also related to description in the memory of a machine.

Comp implies the "unification" in the sense of Bostrom. Conscious
experiences are unique, they cannot be replicated, and they supervene
on all the computations, and thus all implementations going through
their corresponding 3 comp states, or below.


Best,

Bruno

http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Dec 1, 2008, 11:03:28 AM12/1/08
to everyth...@googlegroups.com


Oh( That is not true! We still have the projector and the film. We can
project the movie in the air or directly in your eyes.
I agree for this for this when the film itself is made empty, but then
I can recover a conterfactually correct computation by adding inert
material!


> You are dropping MAT
> at this step,

No. Only when I got that Alice's consciousness supervene on the empty
film (with or without inert material).

> not leading MEC+MAT to a contradiction.
>
>> But this reasoning goes through if we make the hole in the film
>> itself.
>> Reconsider the image on the screen: with a hole in the film itself,
>> you
>> get a "hole" in the movie, but everything which enters and go out
>> of the
>> hole remains the same, for that (unique) range of activity. The
>> "hole"
>> has trivially the same functionality than the subgraph functionality
>> whose special behavior was described by the film. And this is true
>> for
>> any subparts, so we can remove the entire film itself.
>
> We can talk about this part after I understand why you can drop our
> optical boolean network *grin*


it is really something people have to meditate. I could have conclude
in the absurdity of MAT (with MEC) at MGA 2. It is hard for me to take
people seriously when they argue that the consciousness of Alice
supervenes on a movie of its brain activity. There is no causality,
nor computations, during the *projection* of the movie. Alice's
experience is related to ALL computations going through those states,
not to descriptions of those states which can been made and collected
in other histories. Locally it makes sense to ascribe *that*
consciousness when you have the mean to interpret (through some
universal machine) her computational states.

[Consciousness of (x,t)] is never [physical states] at (x,t)

it is:

[Consciousness of (x,t)] is always all computational states (in the UD
°) corresponding to that experience. (It is an indexical view of
reality).

And computational states can be defined by true platonic relation
between numbers. (The usual way is done with Kleene predicate).

Bruno

http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Dec 1, 2008, 11:20:34 AM12/1/08
to everyth...@googlegroups.com
Hi Abram,


On 30 Nov 2008, at 19:17, Abram Demski wrote:

>
> Bruno,
>
> No, she cannot be conscious that she is partially conscious in this
> case, because the scenario is set up such that she does everything as
> if she were fully conscious-- only the counterfactuals change. But, if
> someone tested those counterfactuals by doing something that the
> recording didn't account for, then she may or may not become conscious
> of the fact of her partial consciousness-- in that case it would be
> very much like brain damage.


A very serious brain damage!


>
>
> Anyway, yes, I am admitting that the film of the graph lacks
> counterfactuals and is therefore not conscious.


OK.

> My earlier splitting
> of the argument into an argument about (1) and a separate argument
> against (2) was perhaps a bit silly, because the objection to (2) went
> far enough back that it was also an objection to (1). I split the
> argument like that just because I saw an independent flaw in the
> reasoning of (1)... anyway...
>
> Basically, I am claiming that there is a version of COMP+MAT that MGA
> is not able to derive a contradiction from. The version goes something
> like this:
>
> "Yes, consciousness supervenes on computation, but that computation
> needs to actually take place (meaning, physically). Otherwise, how
> could consciousness supervene on it?


Yes but with UDA the contrary happens. Even if a material world, the
question becomes: how could consciousness remain attached on this
matter.
(It is simpler to understand this issue by supposing some concrete
universal deployment in the "real" universe, and this provides the
motivation for MGA. the concreteness of the UD is a red herring.

You seem to forget that the MAT mind-body problem is not solved. I
mean this is what all experts in the field agree on. To invoke matter
to have something on which consciousness can supervene on, seems to me
a "gap explanation". It introduces more mystery than needed.


> Now, in order for a computation
> to be physically instantiated, the physical instantiation needs to
> satisfy a few properties. One of these properties is clearly some sort
> of isomorphism between the computation and the physical instantiation:
> the actual steps of the computation are represented in physical form.
> A less obvious requirement is that the physical computation needs to
> have the proper counterfactuals: if some external force were to modify
> some step in the computation, the computation must progress according
> to the new computational state (as translated by the isomorphism)."

You will be led to difficulties, like giving a computational role to
inert material. It is ok, because it saves the counterfactual (and
thus MEC), but on the price of attributing a flow of conscious
experience (in real time) to inert material. I can't swallow that,
especially if the motivation is going back to the unsolved problems of
mind, matter and their relations.

By dropping MAT, we have an explanation of consciousness or of the
reason why numbers, due to their true relations with many other
numbers, can develop from inside stable (from their views) believes on
reality and realities including, evidences can be found, physical
realities. Numbers, or combinators, etc.

Bruno

http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Dec 1, 2008, 12:07:30 PM12/1/08
to everyth...@googlegroups.com

On 01 Dec 2008, at 03:25, Russell Standish wrote:

>
> On Sun, Nov 30, 2008 at 07:10:43PM +0100, Bruno Marchal wrote:
>>>
>>> I am speaking as someone unconvinced that MGA2 implies an
>>> absurdity. MGA2 implies that the consciousness is supervening on the
>>> stationary film.
>>
>>
>> ? I could agree, but is this not absurd enough, given MEC and the
>> definition of the physical superveneience thesis;
>
> It is, prima facie, no more absurd than consciousness supervening on a
> block universe.
>
>>>
>>> A block universe is nondynamic by definition. But looked at another
>>> way, (ie from the inside) it is dynamic. It neatly illustrates why
>>> consciousness can supervene on a stationary film (because it is
>>> stationary when viewed from the inside).
>>
>> OK, but then you clearly change the physical supervenience thesis.
>>
>
> How so? The stationary film is a physical object, I would have
> thought.


I don't understand this. The physical supervenience thesis associate
consciousness AT (x,t) to a computational state AT (x,t). The idea is
that consciousness can be "created" in real time by the physical
"running" of a computation (viewed of not in a block universe).

With the stationary film, this does not make sense. Alice experience
of a dream is finite and short, the film lasts as long as you want. I
think I see what you are doing: you take the stationary film as an
incarnation of a computation in Platonia. In that sense you can
associate the platonic experience of Alice to it, but this is a
different physical supervenience thesis. And I argue that even this
cannot work, because the movie does not capture a computation. The
universal interpreter is lacking. It could even correspond to another
experience, if the graph was a movie of another sort of computer, for
example with NAND substituted for the NOR.

>
>
>>
>>> The "film", however does need
>>> to be sufficiently rich, and also needs to handle counterfactuals
>>> (unlike the usual sort of movie we see which has only one plot).
>>
>>
>> OK. Such a film could be said to be a computation. Of course you are
>> not talking about a stationary thing, which, be it physical or
>> immaterial, cannot handle counterfactuals.
>>
>
> If true, then a block universe could not represent the
> Multiverse. Maybe so, but I think a lot of people might be surprised
> at this one.


I am not sure I can give sense to an expression like "the multiverse"
or the "block universe" can or cannot handle counterfactuals. They
have no inputs, nor outputs.


Bruno
http://iridia.ulb.ac.be/~marchal/

Günther Greindl

unread,
Dec 1, 2008, 4:53:06 PM12/1/08
to everyth...@googlegroups.com
Hi Bruno,

>> but no! Then we wouldn't have a substrate anymore.
> Oh( That is not true! We still have the projector and the film. We can
> project the movie in the air or directly in your eyes.

Ok I see now where our intuitions differ (always the problem with
thought experiment) - but maybe we can clear this up and see where it
leads...

> it is really something people have to meditate. I could have conclude
> in the absurdity of MAT (with MEC) at MGA 2. It is hard for me to take
> people seriously when they argue that the consciousness of Alice
> supervenes on a movie of its brain activity. There is no causality,
> nor computations, during the *projection* of the movie.

If that is how you see MAT (you require causality) - then I would also
agree -> MGA 2 shows absurdity.

>Alice's
> experience is related to ALL computations going through those states,
> not to descriptions of those states which can been made and collected
> in other histories. Locally it makes sense to ascribe *that*
> consciousness when you have the mean to interpret (through some
> universal machine) her computational states.

That is already part of your theory (UDA and all) (as I understand it),
but not included already in COMP or in MAT.

> [Consciousness of (x,t)] is never [physical states] at (x,t)

For me, the above expresses the essence of (naive) MAT -> let's call it
NMAT.

So, clearly:
NMAT: [Consciousness of (x,t)] supervenes on [physical states] at (x,t)

And on physical states only! Not on the causal relations of these states
(block universe view).

Your argument goes like this:


> it is:
> [Consciousness of (x,t)] is always all computational states (in the UD
> °) corresponding to that experience. (It is an indexical view of
> reality).

And I share it IF we can show that MAT+MEC is inconsistent. But I am not
convinced yet.

For me, the essence of MEC (COMP) is this:

COMP: there is a level at which a person can be substituted at a digital
level (we don't have to go down to "infinity"), and where this digital
description is enough to reconsitute this person elsewhere and elsewhen,
independent of substrate.


NMAT additionally requires that the substrate for COMP be some
"mysterious" substance, and not only a platonic relation.

My intuition tells me this can't be -> we have to drop either MEC or NMAT.

But MGA 3, when dropping the boolean gates, violates NMAT, because:
NMAT: [Consciousness of (x,t)] supervenes on [physical states] at (x,t)

And the physical states relevant where the _states of the boolean graph_
(the movie projector was just the lucky cosmic ray).

Do you have different definition for MAT? Do you require causal dynamics
for MAT?

The problem with NMAT as I define it raises the issue as in the Putnam
paper -> does every rock implement every finite state-automaton?

Chalmers makes the move to implementation, so introduces causal dynamics.

So, sophisticated MAT would probably be:
SMAT: [Consciousness of (x,t)] supervenes on [physical states] at (x,t)
over a timespan delta(t) _if_ sufficiently complex causal dynamics are
at work during this timespan relating the physical states.


Then I would say: MGA 2 (already) shows that SMAT+MEC are not
compatible. No need for MGA 3.

For NMAT+MEC (which is problematic for other reasons) MGA 3 is not
convincing.

Would you agree with this?

Cheers,
Günther

Abram Demski

unread,
Dec 1, 2008, 7:05:06 PM12/1/08
to everyth...@googlegroups.com
Bruno,

It sounds like what you are saying in this reply is that my version of
COMP+MAT is consistent, but counter to your intuition (because you
cannot see how consciousness could be attached to physical stuff). If
this is the case, then it sounds like MGA only works for specific
versions of MAT-- say, versions of MAT that claim consciousness hinges
only on the matter, not on the causal relationships. In other words,
what Günther called NMAT. So you need a different argument against--
let's call it CMAT, for causal MAT. The olympization argument only
works if COMP+CMAT can be shown to imply the removability of inert
matter... which I don't think it can, because that inert matter here
has a causal role to play in the counterfactuals, and is therefore
essential to the physical computation.

--Abram

Brent Meeker

unread,
Dec 1, 2008, 9:33:57 PM12/1/08
to everyth...@googlegroups.com
Bruno Marchal wrote:
>
> On 01 Dec 2008, at 03:25, Russell Standish wrote:
>
>> On Sun, Nov 30, 2008 at 07:10:43PM +0100, Bruno Marchal wrote:
>>>> I am speaking as someone unconvinced that MGA2 implies an
>>>> absurdity. MGA2 implies that the consciousness is supervening on the
>>>> stationary film.
>>>
>>> ? I could agree, but is this not absurd enough, given MEC and the
>>> definition of the physical superveneience thesis;
>> It is, prima facie, no more absurd than consciousness supervening on a
>> block universe.
>>
>>>> A block universe is nondynamic by definition. But looked at another
>>>> way, (ie from the inside) it is dynamic. It neatly illustrates why
>>>> consciousness can supervene on a stationary film (because it is
>>>> stationary when viewed from the inside).
>>> OK, but then you clearly change the physical supervenience thesis.
>>>
>> How so? The stationary film is a physical object, I would have
>> thought.
>
>
> I don't understand this. The physical supervenience thesis associate
> consciousness AT (x,t) to a computational state AT (x,t).

Stated this way seems to assume that the causal relations between the states are
irrelevant, only the states matter.

>The idea is
> that consciousness can be "created" in real time by the physical
> "running" of a computation (viewed of not in a block universe).

Well we're pretty sure that brains do this.

>
> With the stationary film, this does not make sense. Alice experience
> of a dream is finite and short, the film lasts as long as you want. I
> think I see what you are doing: you take the stationary film as an
> incarnation of a computation in Platonia. In that sense you can
> associate the platonic experience of Alice to it, but this is a
> different physical supervenience thesis. And I argue that even this
> cannot work, because the movie does not capture a computation.

I was thinking along the same lines. But then the question is what does capture
a computation. Where in the thought experiments, starting with natural Alice
and ending with a pictures of Alice's brain states, did we lose computation? Is
it important that the sequence be time rather than space or some other order?
Is it the loss causal relations or counterfactuality?

Brent

Bruno Marchal

unread,
Dec 2, 2008, 12:18:07 PM12/2/08
to everyth...@googlegroups.com

On 02 Dec 2008, at 01:05, Abram Demski wrote:

>
> Bruno,
>
> It sounds like what you are saying in this reply is that my version of
> COMP+MAT is consistent, but counter to your intuition (because you
> cannot see how consciousness could be attached to physical stuff).

I have no problem a priori in attaching consciousness to physical
stuff. I do have problem when MEC + MAT forces me to attach
consciousness to an empty machine (with no physical activity) together
with inert material.


> If
> this is the case, then it sounds like MGA only works for specific
> versions of MAT-- say, versions of MAT that claim consciousness hinges
> only on the matter, not on the causal relationships.

On the contrary. I want consciousness related to the causal
relationship. But with MEC the causal relationship are in the
computations. The thought experiment shows that the physical
implementation plays the role of making them able to manifest
relatively to us, but are not responsible for their existence.


> In other words,
> what Günther called NMAT. So you need a different argument against--
> let's call it CMAT, for causal MAT. The olympization argument only
> works if COMP+CMAT can be shown to imply the removability of inert
> matter... which I don't think it can, because that inert matter here
> has a causal role to play in the counterfactuals, and is therefore
> essential to the physical computation.

OK, so now you have to disagree with MGA 1. No problem. But would you
still say "yes" to the mechanist doctor? I don't see how, because now
you appeal to something rather magic like influence in real time of
inactive material. Or, you are weakening the physical supervenience
thesis by appeal to a notion of causality which seems to me a bit
magical, and contrary to the local functionalism of the
computationalist.

The real question I have to ask to you, Günther and others is this
one: does your new supervenience thesis forced the UD to be
physically executed in a "real" universe to get the UDA conclusion?
Does MGA, even just as a refutation of "naïve mat" eliminate the use
of the concrete UD in UDA?

It is true that by weakening MEC or MAT, the reasoning doesn't go
through, but it seems to me the conclusion goes with any primitive
stuff view of MAT or Matter activity to which we could attach
consciousness through "causal" links. Once you begin to define matter
through causal links, and this keeping comp, and linking the
experience to those causal relation, perhaps made in other time at
other occasion, you are not a long way from the comp supervenience.
But if you don't see this, I guess the conversation will continue.

Bruno

http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Dec 2, 2008, 12:58:16 PM12/2/08
to everyth...@googlegroups.com


Ah, please, add the delta again (see my previews post). I did wrote
(dx,dt), but Anna thought it was infinitesimal. It could be fuzzy
deltas or whatever you want. Unless you attach your consciousness,
from here and now, to the whole block multiverse, the reasoning will
go through, assuming of course that the part of the multiverse, on
which you attach your mind, is Turing emulable (MEC).

>
>
>> The idea is
>> that consciousness can be "created" in real time by the physical
>> "running" of a computation (viewed of not in a block universe).
>
> Well we're pretty sure that brains do this.

Well, my point is that for believing this, you have to abandon the MEC
hypothesis, perhaps in a manner like Searle or Penrose. Consciousness
would be the product of some non Turing emulable chemical reactions.
But if everything in the brain (or the genralized brain) is turing
emulable, then the reasoning (uda+mga) is supposed to explain why
consciousness (an immaterial thing) is related only to the computation
made by the brain, but not the brain itself nor to its physical
activity during the physical implementation. Your locally physical
brain just makes higher the probability that your consciousness
remains entangled with mine (and others).


>
>
>>
>> With the stationary film, this does not make sense. Alice experience
>> of a dream is finite and short, the film lasts as long as you want. I
>> think I see what you are doing: you take the stationary film as an
>> incarnation of a computation in Platonia. In that sense you can
>> associate the platonic experience of Alice to it, but this is a
>> different physical supervenience thesis. And I argue that even this
>> cannot work, because the movie does not capture a computation.
>
> I was thinking along the same lines. But then the question is what
> does capture
> a computation. Where in the thought experiments, starting with
> natural Alice
> and ending with a pictures of Alice's brain states, did we lose
> computation? Is
> it important that the sequence be time rather than space or some
> other order?
> Is it the loss causal relations or counterfactuality?


We "lose a computation" relatively to us when the computation is not
executed by a stable (relatively to us) universal machine nearby, be
it a cell, a brain, a natural or artificial universal computer.

In the case of the movie, it is no so bad. Consciousness does not
supervene on the movie or its projection, but the movie can be used as
a backup of Alice's state. We can re-project a frame, of that movie,
on a functionally well working Boolean optical graph, and Alice will
be back ... with us.

Of course the computations themselves, and their many possible
differentiations, are already in Platonia (= in the solution of the
universal Diophantine equation, in the processing of the UD, or
perhaps in the Mandelbrot set).

Alice's brain and body are "just" local stable artifacts belonging to
our (most probable) computational history, and making possible for
Alice consciousness to differentiate through interactions with us,
relatively to us.

Bruno


http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Dec 2, 2008, 1:48:19 PM12/2/08
to everyth...@googlegroups.com
Hi Günther,


On 01 Dec 2008, at 22:53, Günther Greindl wrote:

>
> Hi Bruno,
>
>>> but no! Then we wouldn't have a substrate anymore.
>> Oh( That is not true! We still have the projector and the film. We
>> can
>> project the movie in the air or directly in your eyes.
>
> Ok I see now where our intuitions differ (always the problem with
> thought experiment) - but maybe we can clear this up and see where it
> leads...


OK.

>
>
>> it is really something people have to meditate. I could have conclude
>> in the absurdity of MAT (with MEC) at MGA 2. It is hard for me to
>> take
>> people seriously when they argue that the consciousness of Alice
>> supervenes on a movie of its brain activity. There is no causality,
>> nor computations, during the *projection* of the movie.
>
> If that is how you see MAT (you require causality) - then I would also
> agree -> MGA 2 shows absurdity.

Well I require at least a minimum of physical causality to implement
physically the computational "causality" (which incarnates platonic
relation existing among numbers).

MAT presupposes anything primitively material and causal of course.
Remember that I am using Materialism and physicalism (and naturalsim)
as synonymous, because the argument is very general. The (naïve) idea
is that the brain *does* compute something when you dream, for
example, and that it is the physical causality which is responsible
for the implementation of the computation.

>
>
>> Alice's
>> experience is related to ALL computations going through those states,
>> not to descriptions of those states which can been made and collected
>> in other histories. Locally it makes sense to ascribe *that*
>> consciousness when you have the mean to interpret (through some
>> universal machine) her computational states.
>
> That is already part of your theory (UDA and all) (as I understand
> it),
> but not included already in COMP or in MAT.


Not at all. This could be confusing for those who don't know UDA. Once
MEC+MAT is shown to be incompatible, we then chose MEC and thus
abandon MAT.
(why? just for not going out of the range of my working hypothesis, ok)

With MEC, there is no more physical supervenience thesis, on the kind
compatible with MEC. But we keep MEC, so we have to continue to
related consciousness with the computation, right? We do no more have
a notion of physical computation, so we attach consciousness to the
computation itself, LIKE it has already been done in the UDA, except
that we don't need no more to run the UD.

>
>
>> [Consciousness of (x,t)] is never [physical states] at (x,t)
>
> For me, the above expresses the essence of (naive) MAT -> let's call
> it
> NMAT.
>
> So, clearly:
> NMAT: [Consciousness of (x,t)] supervenes on [physical states] at
> (x,t)
>
> And on physical states only! Not on the causal relations of these
> states
> (block universe view).


You are perhaps taking me too much literally here. It is just
difficult, lengthy and confusing to make a precise definition of the
physical supervenience which would work for the different views of the
universe.
The physical supervenience thesis just says that 1) there is a
physical universe, 2) it can compute, and consciousness requires some
special local computations made *in* that universe.

>
>
> Your argument goes like this:
>> it is:
>> [Consciousness of (x,t)] is always all computational states (in the
>> UD
>> °) corresponding to that experience. (It is an indexical view of
>> reality).
>
> And I share it IF we can show that MAT+MEC is inconsistent. But I am
> not
> convinced yet.
>
> For me, the essence of MEC (COMP) is this:
>
> COMP: there is a level at which a person can be substituted at a
> digital
> level (we don't have to go down to "infinity"), and where this digital
> description is enough to reconsitute this person elsewhere and
> elsewhen,
> independent of substrate.
>
>
> NMAT additionally requires that the substrate for COMP be some
> "mysterious" substance, and not only a platonic relation.


Not so mysterious. It just seems to require some particular
computations. The "physical" one. People are used to think about it in
term of waves or particles, or field, geometrical dynamical object.
They believe those are particulars (which become mysterious only with
comp, but a priori with Mat they are rather "natural");

>
>
> My intuition tells me this can't be -> we have to drop either MEC or
> NMAT.
>
> But MGA 3, when dropping the boolean gates, violates NMAT, because:
> NMAT: [Consciousness of (x,t)] supervenes on [physical states] at
> (x,t)
>
> And the physical states relevant where the _states of the boolean
> graph_
> (the movie projector was just the lucky cosmic ray).
>
> Do you have different definition for MAT? Do you require causal
> dynamics
> for MAT?


MAT is very general, but indeed it requires the minimum amount of
causality so that we can implement a computation in the physical
world, if not I don't see how we could talk on physical supervenience.

>
>
> The problem with NMAT as I define it raises the issue as in the Putnam
> paper -> does every rock implement every finite state-automaton?
>
> Chalmers makes the move to implementation, so introduces causal
> dynamics.
>
> So, sophisticated MAT would probably be:
> SMAT: [Consciousness of (x,t)] supervenes on [physical states] at
> (x,t)
> over a timespan delta(t) _if_ sufficiently complex causal dynamics are
> at work during this timespan relating the physical states.


No this is MAT. But with COMP, if it happens that a boolean gates or
neurons *personal history" has to be taken into account, by definition
of MEC, it means we don't have chose correctly the level of
substitution.


>
>
>
> Then I would say: MGA 2 (already) shows that SMAT+MEC are not
> compatible.

All right then!

> No need for MGA 3.

OK.


>
>
> For NMAT+MEC (which is problematic for other reasons) MGA 3 is not
> convincing.

I think I have miss something.

>
>
> Would you agree with this?


I am not sure I understand the NMAT.
What MGA is supposed to do is to eliminate the hypothesis of the
concrete UD in the step seven of UDA. Do you think it does, or not
(yet)?

Best,

Bruno

http://iridia.ulb.ac.be/~marchal/

Abram Demski

unread,
Dec 2, 2008, 2:33:27 PM12/2/08
to everyth...@googlegroups.com
Bruno,

I am a bit confused. To me, you said

>Or, you are weakening the physical supervenience
> thesis by appeal to a notion of causality which seems to me a bit
> magical, and contrary to the local functionalism of the
> computationalist.

This seems to say that the version of MAT that MGA is targeted at does
not include causal requirements.

To Günther, you said:

>> Do you have different definition for MAT? Do you require causal
>> dynamics
>> for MAT?
>
>
> MAT is very general, but indeed it requires the minimum amount of
> causality so that we can implement a computation in the physical
> world, if not I don't see how we could talk on physical supervenience.

Does the MAT you are talking about include causal requirements or not?

About your other questions--

> OK, so now you have to disagree with MGA 1. No problem. But would you
> still say "yes" to the mechanist doctor? I don't see how, because now
> you appeal to something rather magic like influence in real time of
> inactive material.

So long as that "inert" material preserves the correct
counterfactuals, everything is fine. The only reason things seem
strange with olympized Alice is because *normally* we do not know in
advance which path cause and effect will take for something as
intricate as a conscious entity. The air bags in a car are "inert" in
the same way-- many cars never get in a crash, so the air bags remain
unused. But since we don't know that ahead of time, we want the air
bags. Similarly, when talking to the mechanist doctor, I will not be
convinced that a recording will suffice...

> The real question I have to ask to you, Günther and others is this
> one: does your new supervenience thesis forced the UD to be
> physically executed in a "real" universe to get the UDA conclusion?

Yes.

> Does MGA, even just as a refutation of "naïve mat" eliminate the use
> of the concrete UD in UDA?

No.

(By the way, I have read UDA now, but have refrained from posting a
commentary since there has been a great deal of discussion about it on
this list and I could just be repeating the comments of others...)

Also: Günther mentioned "SMAT", which actually sounds like the "CMAT"
I proposed... so I'll refer to it as SMAT from now on.

--Abram

Abram Demski

unread,
Dec 2, 2008, 2:37:31 PM12/2/08
to everyth...@googlegroups.com
Günther,

Why does MGA 2 show that SMAT + MEC is inconsistent?

The way I see it, SMAT + MEC should say that a recording of Alice does
not count as conscious, because it lacks the proper causal structure
(or equivalently, the proper counterfactual behavior).

--Abram

Brent Meeker

unread,
Dec 2, 2008, 4:24:08 PM12/2/08
to everyth...@googlegroups.com

OK, that clarifies things and it corresponds with my intuition that
consciousness is relative to an environment. I can't seem to answer the
question is MG-Alice conscious "yes" or "no", but I can say she is conscious
within the movie environment, but not within our environment. This is similar
to Stathis asking about consciousness within a rock. We could say the thermal
motions of atoms within the rock may compute consciousness, but it is a
consciousness within the rock environment, not in ours.

Brent

Jason Resch

unread,
Dec 2, 2008, 11:58:05 PM12/2/08
to everyth...@googlegroups.com
On Sun, Nov 30, 2008 at 11:33 AM, Bruno Marchal <mar...@ulb.ac.be> wrote:


All this is a bit complex because we have to take well into account
the distinction between

A computation in the "real" world,
A description of a computation in the "real" world,

And then most importantly:

A computation in Platonia
A description of a computation in Platonia.

I argue that consciousness supervenes on computation in Platonia. Even
in Platonia consciousness does not supervene on description of the
computation, even if those description are 100% precise and correct

Bruno, this is interesting and I have had similar thoughts of late regarding along this vein.  The trouble is, I don't see how the "real" world can be differentiated from Platonia.  Just as the UD contains instances of itself, and hence computations within computations, can't mathematical objects contain mathematical objects?  If so then aren't our actions in this universe just as mathematcally or computationally fundamental as any other instantiation in platonia?  Platonia might be highly interconnected even fractal and so performing a computation in this universe in a sense hasn't created anything new, but created a link to other identical things which have always been there, and in the timelessness of platonia one can't say which came before, or which is the original or most real.

After wrestling with block time, the MGA, and computationalism I'm starting to wonder how computations are implemented in a 4 dimensional and static mathematical object.  The best I can come up with is that the mathematical structure is defined by some equation or equations, and that by virtue of this imposed order, defines relations between particles.  Computation depends on relations, be it electrons in silicon, Chinese with radios or a system of beer cans and ping-pong balls; from the outside there is little or no indication what is going on is forming consciousness, it is only relative from the inside, and since these relations carry state and information across one of the 4 dimensions of the universe we end up with DNA and brains which record and process information in sequence, or so it appears to us being trapped on within this equation defined in platonia.

In the case of a movie that is in this physical world, no mathematical equation defines progression between frames and there is no conveyance of information, the alteration of one frame does not affect any other which would not be the case nor possible with a timeless mathematical object.

Jason

Bruno Marchal

unread,
Dec 3, 2008, 4:08:53 AM12/3/08
to everyth...@googlegroups.com
Hi Abram,

On 02 Dec 2008, at 20:33, Abram Demski wrote:

>
> Bruno,
>
> I am a bit confused. To me, you said
>
>> Or, you are weakening the physical supervenience
>> thesis by appeal to a notion of causality which seems to me a bit
>> magical, and contrary to the local functionalism of the
>> computationalist.
>
> This seems to say that the version of MAT that MGA is targeted at does
> not include causal requirements.


MAT is the usual idea that there is a physical world described through
physical laws. Those capture physical causality, generally under the
form of differential equations. If there were no causality in physics,
the very notion of physical supervenience would not make sense. Nor MEC
+MAT, at the start. Sorry if I have been unclear, but I was
criticizing only the *magical* causality which is necessary for
holding both the physical supervenience thesis and the mechanist
hypothesis, like attribution of prescience to the neurons (in MGA 1),
or attributing a computational role in inert Material.

>
>
> To Günther, you said:
>
>>> Do you have different definition for MAT? Do you require causal
>>> dynamics
>>> for MAT?
>>
>>
>> MAT is very general, but indeed it requires the minimum amount of
>> causality so that we can implement a computation in the physical
>> world, if not I don't see how we could talk on physical
>> supervenience.
>
> Does the MAT you are talking about include causal requirements or not?


Of course.


>
>
> About your other questions--
>
>> OK, so now you have to disagree with MGA 1. No problem. But would you
>> still say "yes" to the mechanist doctor? I don't see how, because
>> now
>> you appeal to something rather magic like influence in real time of
>> inactive material.
>
> So long as that "inert" material preserves the correct
> counterfactuals, everything is fine. The only reason things seem
> strange with olympized Alice is because *normally* we do not know in
> advance which path cause and effect will take for something as
> intricate as a conscious entity. The air bags in a car are "inert" in
> the same way-- many cars never get in a crash, so the air bags remain
> unused. But since we don't know that ahead of time, we want the air
> bags. Similarly, when talking to the mechanist doctor, I will not be
> convinced that a recording will suffice...


Me too. But that remark is out of the context of the argument. If I
want an artificial brain (MEC) I expect it to handle the
counterfactuals, because indeed we don't know things in advance. But
in the context of the proof we were in a situation where we did know
the things in advance. Suppose that my doctor discovers in my brain
some hardware build for managing my behavior only in front of
dinosaurs, like old unused subroutine being only relic of the past,
then it seems to me that the doctor, in the spirit of mechanist
functionalism can decide of dropping those subroutine for building me
a cheaper artificial brain. And that is all we need for the argument
for going through. Consciousness relies on the computation which
always kept the right counterfactuals, and never on their relative
implementations which will only change their relative measures.

>
>
>> The real question I have to ask to you, Günther and others is this
>> one: does your new supervenience thesis forced the UD to be
>> physically executed in a "real" universe to get the UDA conclusion?
>
> Yes.

Then it seems to me you are relying on some magical causality attached
to a magical notion of matter. I don't understand how you can still
say yes to a doctor with such a notion of mechanism. See above. I
would no more even trust a Darwinian brain.


>
>
>> Does MGA, even just as a refutation of "naïve mat" eliminate the use
>> of the concrete UD in UDA?
>
> No.
>
> (By the way, I have read UDA now, but have refrained from posting a
> commentary since there has been a great deal of discussion about it on
> this list and I could just be repeating the comments of others...)


Then you can read the answers I have given to the others. It seems to
me UDA(1..7) does no more pose any problem, except for those who have
decided to not understand, or believes "religioulsy" in matter and
comp. In public forum group you always end up discussing with those
who like cutting the hairs.


>
>
> Also: Günther mentioned "SMAT", which actually sounds like the "CMAT"
> I proposed... so I'll refer to it as SMAT from now on.


I am sorry if I have been unclear, but MAT is taken in a very large
sense. MAT is the belief in a physical universe obeying physical laws,
be it quantum, classical, or whatever. Actually, for a
computationalist (especially after UDA+MGA), MAT seems to be just a
way to single out one special computations above the others.

Kim Jones has convinced me to explain UDA, and the general idea, a new
time. It could be an opportunity to let us known your commentaries. To
be sure, some mathematicians get more easily the point when I
introduce the arithmetical translation of the UDA. You can study it in
my papers or wait I explain this also on the list, or search the list
archive for the explanation of the AUDA that I have already given. Of
course mathematicians are a minority here and I try not to exaggerate
with the list patience. What is subtle in MGA is a bit subtler in
AUDA, too, but then it relies at least on non controversial facts in
Mathematical Logic, which are unfortunately not well known nor well
taught. Except the logicians, logic is not very well known even by
mathematicians.

Bruno
http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Dec 3, 2008, 4:18:16 AM12/3/08
to everyth...@googlegroups.com

On 02 Dec 2008, at 22:24, Brent Meeker wrote:



....
Alice's brain and body are "just" local stable artifacts belonging to  
our (most probable) computational history, and making possible for  
Alice consciousness to differentiate through interactions with us,  
relatively to us.

Bruno

OK, that clarifies things and it corresponds with my intuition that
consciousness is relative to an environment.  I can't seem to answer the
question is MG-Alice conscious "yes" or "no", but I can say she is conscious
within the movie environment, but not within our environment.  This is similar
to Stathis asking about consciousness within a rock.  We could say the thermal
motions of atoms within the rock may compute consciousness, but it is a
consciousness within the rock environment, not in ours.

Your consciousness is related to all computations going through your (current) brain states. I have not find any reason to think that a rock implement some consciousness, but if this is the case you have to take it into account for the general measure, given that in this case the UD will generate the rock computations too.
Now, I don't think there is any consciousness in the movie, even it is generated in the UD. There is just no computation or relevant "physical" causality linkable to a computation in a movie.
So consciousness can never be ascribed to anything physical, and thus we have reduce the mind body problem into the comp body problem; how does the appearance of matter emerge from the (immaterial) execution of the platonic deployment.

Bruno


Bruno Marchal

unread,
Dec 3, 2008, 10:53:11 AM12/3/08
to everyth...@googlegroups.com
On 03 Dec 2008, at 05:58, Jason Resch wrote:



On Sun, Nov 30, 2008 at 11:33 AM, Bruno Marchal <mar...@ulb.ac.be> wrote:


All this is a bit complex because we have to take well into account
the distinction between

A computation in the "real" world,
A description of a computation in the "real" world,

And then most importantly:

A computation in Platonia
A description of a computation in Platonia.

I argue that consciousness supervenes on computation in Platonia. Even
in Platonia consciousness does not supervene on description of the
computation, even if those description are 100% precise and correct

Bruno, this is interesting and I have had similar thoughts of late regarding along this vein.  The trouble is, I don't see how the "real" world can be differentiated from Platonia.  


It is hard to answer this. I think that after UDA+MGA the "real" (physical) world is the sum on all computations going through my state or our sharable comp state. With comp, Platonia, 3-Platonia (to be sure), can be represented by a tiny part of arithmetic, or just by the deployment of the UD. The physical world will be an inside construction made by inside machines/numbers. It will appear that such an inside view will be much bigger than 3-platonia. Like in Skolem paradox, platonia can be rather little from "outside", and *very* big from inside.




Just as the UD contains instances of itself, and hence computations within computations,


I guess you mean the deployment. UD is the finite program which generates the deployment. At some point we will have to be cautious not identifying those two things. But it is OK. And indeed, the deployment contains an infinity of deployment which themselves contain an infinity off deployment ...



can't mathematical objects contain mathematical objects?  


Some can. Exemples: well, the deployment :)
But many fractals, universal or not, etc. Actually they are more included in themselves than element of themselves. But this is a bit math tech.



If so then aren't our actions in this universe just as mathematcally or computationally fundamental as any other instantiation in platonia?  


No more after UDA+MGA, or UDA(1...8). Our consciousness is "attached' to all (relative) instantiation in Platonia. If you make the usual static picture of the deployment, a big 3-dimensional or 2-dimensional cone, each of our states appears infinitely in a quasi dense way on its "border". The notion of ""this" universe" does no even make a clear sense, we can talk only about our most probable histories. And, by doing measurement, we never select one history among an infinity, we always select an infinity of histories among an infinity of histories.




Platonia might be highly interconnected even fractal and so performing a computation in this universe in a sense hasn't created anything new, but created a link to other identical things which have always been there, and in the timelessness of platonia one can't say which came before, or which is the original or most real.


Yes. Moreover, we are never singular. I think this is the startling part which is nevertheless confirmed by QM (Everett).




After wrestling with block time, the MGA, and computationalism I'm starting to wonder how computations are implemented in a 4 dimensional and static mathematical object.  


Why do you want to do that? We have to do the contrary: extract the physics and the math-physics, from the much simpler (yet non trivial) notions of computation and of computation "as seen from inside". To be sure, it is not even obvious that a notion of block-physical-universe will remain possible (I have no idea on this). Our sharable dreams glue well locally, but it is an open question to know if the gluing can be made global and define an objective general physical reality.


The best I can come up with is that the mathematical structure is defined by some equation or equations,


I really don't know. I expect that the mathematical structure, as seen from inside, is so big that Platonia cannot have it neither as element nor as subpart. (Ah, well, I am aware that this is counter-intuitive, but here mathematical logic can help to see the consistency, and the quasi necessity with formal version of comp).



and that by virtue of this imposed order, defines relations between particles.  Computation depends on relations, be it electrons in silicon, Chinese with radios or a system of beer cans and ping-pong balls;


Here you are talking about instantiations of computations relatively to our most probable computations, which have a physical "look". But strictly speaking computations are only relation between numbers.




from the outside there is little or no indication what is going on is forming consciousness, it is only relative from the inside, and since these relations carry state and information across one of the 4 dimensions of the universe we end up with DNA and brains which record and process information in sequence, or so it appears to us being trapped on within this equation defined in platonia.


It is the contrary. We cannot be trapped in any equation. We are trapped only in the deployment. We hope our histories (computations) will lead us to coherent and sufficiently computable local structure so that we can remain in a state of partial control, but we cannot hope for much more than that, unless some program (generated by the UD) wins the measure battle, but then it has to be a sort of super-universal dovetailer. I doubt this exist, despite a feeling that a quantum universal dovetailing could play that role, if it could make sense. Now, even if such sub-program exists, we have to derive its existence from the measure (on 1-computations) problem.



In the case of a movie that is in this physical world, no mathematical equation defines progression between frames and there is no conveyance of information, the alteration of one frame does not affect any other which would not be the case nor possible with a timeless mathematical object.

We agree on this.

Bruno


Jason Resch

unread,
Dec 3, 2008, 11:20:12 AM12/3/08
to everyth...@googlegroups.com
On Wed, Dec 3, 2008 at 9:53 AM, Bruno Marchal <mar...@ulb.ac.be> wrote:

and that by virtue of this imposed order, defines relations between particles.  Computation depends on relations, be it electrons in silicon, Chinese with radios or a system of beer cans and ping-pong balls;


Here you are talking about instantiations of computations relatively to our most probable computations, which have a physical "look". But strictly speaking computations are only relation between numbers.


Bruno,

Thanks for your reply, I am curious what exactly you mean by the most probable computations going through our state if these computations cannot be part of a larger (shared universe) computation.  Where does the data provided to the senses come from if not from a computation which also includes that of the environment as well?  Also, why does the computation have to be between numbers specifically, could a program in the deployment that calculates the evolution of a universe perform the necessary computations to generate an observer?  If they can, then it stands other mathematical objects besides pure turing machines and besides the UD could implement computations capable of generating observers.  I noticed in a previous post of yours you mentioned 'Kleene predicates' as a way of deriving computations from true statements, do you know of any good sources where I could learn more about Kleene predicates?

Thanks,

Jason 

Abram Demski

unread,
Dec 3, 2008, 11:23:53 AM12/3/08
to everyth...@googlegroups.com
Bruno,

I think I see better now. Here is my explanation of why the causality
I am invoking is not "magical" (and why it may not be an objection to
your argument after all).

Generally, the way I was viewing your argument was in terms of a
hypothetical definition of consciousness, X, that is to be supplied by
the reader of the argument. In other words, the reader comes to the
argument having their personal definition, X. X is basically treated
as an arbitrary choice of computations (/ physical configurations)
that the reader has chosen to label "conscious".

The arguments (both UDA and MGA) then go on to reason about X using
some commonalities between most people's personal definitions of
consciousness, and taking into account some possible variations where
there is not enough commonality for the argument to be taken through.

So, when I am talking about an X that depends on hypotheticals, I am
NOT saying that inert matter magically causes consciousness-- I am
saying instead that the definition of consciousness depends on the
hypotheticals. Basically, I am saying "a definition of consciousness
that does not require hypotheticals is not interesting to me-- I don't
see why I would worry about an entity defines in that way". (So it is
irrelevant of the word "conscious", it is the logical form that is
interesting/uninteresting or useful/not).

You, on the other hand, are talking about first-person-experience.
That is a different matter...

Does that make sense?

--Abram

Brent Meeker

unread,
Dec 3, 2008, 1:12:48 PM12/3/08
to everyth...@googlegroups.com
Bruno Marchal wrote:
> Hi Abram,
>
> On 02 Dec 2008, at 20:33, Abram Demski wrote:
>
>> Bruno,
>>
>> I am a bit confused. To me, you said
>>
>>> Or, you are weakening the physical supervenience
>>> thesis by appeal to a notion of causality which seems to me a bit
>>> magical, and contrary to the local functionalism of the
>>> computationalist.
>> This seems to say that the version of MAT that MGA is targeted at does
>> not include causal requirements.
>
>
> MAT is the usual idea that there is a physical world described through
> physical laws. Those capture physical causality, generally under the
> form of differential equations. If there were no causality in physics,
> the very notion of physical supervenience would not make sense. Nor MEC
> +MAT, at the start. Sorry if I have been unclear, but I was
> criticizing only the *magical* causality which is necessary for
> holding both the physical supervenience thesis and the mechanist
> hypothesis, like attribution of prescience to the neurons (in MGA 1),
> or attributing a computational role in inert Material.

This seems to assume there is causality apart from physical causality, but there
is no causality in logic or mathematics (except in a metaphorical, I might say
"magical", sense). So I don't see that Gunther is relying on anything magical.

Brent

Brent Meeker

unread,
Dec 3, 2008, 1:22:43 PM12/3/08
to everyth...@googlegroups.com
But there is causality. The sequence of events in the movie are directly caused
by the projector, but they have a causal linkage back to Alice and the part of
her environment that is captured in the movie. I see no principled reason to
consider only the immediate cause and not refer back further in the chain of
causation.


> So consciousness can never be ascribed to anything physical,

Doesn't your argument imply the opposite? Consciousness can only be ascribed to
physical things because consciousness is computation and computation requires
causal links and causality if a physical relation.

Brent

Bruno Marchal

unread,
Dec 4, 2008, 6:19:33 AM12/4/08
to everyth...@googlegroups.com
Hi Jason,

Le 03-déc.-08, à 17:20, Jason Resch a écrit :

> On Wed, Dec 3, 2008 at 9:53 AM, Bruno Marchal <mar...@ulb.ac.be>
> wrote:
>>
>>> and that by virtue of this imposed order, defines relations between
>>> particles.  Computation depends on relations, be it electrons in
>>> silicon, Chinese with radios or a system of beer cans and ping-pong
>>> balls;
>>
>>
>> Here you are talking about instantiations of computations relatively
>> to our most probable computations, which have a physical "look". But
>> strictly speaking computations are only relation between numbers.
>>
>
> Bruno,
>
> Thanks for your reply, I am curious what exactly you mean by the most
> probable computations going through our state if these computations
> cannot be part of a larger (shared universe) computation.  


Hmmm... It means you have still a little problem with step seven. I
wish we share a computable environment, but we cannot decide this at
will. I agree we have empirical evidence that here is such (partially)
computable environment, and I am willing to say I trust nature for
this. Yet, the fact is that to predict my next first person experience
I have to take into account ALL computations which exist in the
arithmetical "platonia" or in the universal dovetailing.


> Where does the data provided to the senses come from if not from a
> computation which also includes that of the environment as well?  

You don't know that. The data and their statistics come from all
computational histories going through my state. The game is to take
completely seriously the comp hyp, and if it contradicts facts, we will
abandon it. But that day has not yet come .... Until then we have to
derive the partial computability of our observable enviroment from a
statistic on all computations made by the UD.


> Also, why does the computation have to be between numbers specifically,

They don't. Sometimes I use the combinators. They have to be finite
objects, and this comes from the *digital* aspect of the comp. hyp.

> could a program in the deployment that calculates the evolution of a
> universe

This is something you have to define. If you do it I bet you will find
a program equivalent to a universal dovetailer, a bit like Everett
universal quantum wave.

> perform the necessary computations to generate an observer?  

Sure. The problem is that there will be an infinity of program
generating the same observer, in the same state, and the observer
cannot know in which computations it belongs. Never? Measurement
particularizes, but never get singular.

> If they can, then it stands other mathematical objects besides pure
> turing machines and besides the UD could implement computations
> capable of generating observers.

Not really. Those objects are internam construction made by programs
relatively to trheir most probable history.

>  I noticed in a previous post of yours you mentioned 'Kleene
> predicates' as a way of deriving computations from true statements, do
> you know of any good sources where I could learn more about Kleene
> predicates?

A very good introduction is the book by N.J. Cutland. See the reference
in my thesis. There are other books. I will think to make a list with
some comments. Actually I really love Kleene's original "Introduction
to Metamathematics", but the notations used are a bit old fashioned.

Hope I am not too short. I am a bit busy today,

Best,

Bruno

http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Dec 4, 2008, 6:28:33 AM12/4/08
to everyth...@googlegroups.com
Brent,

I try to single out where you depart from the comp hyp, to focus on the
essential. I could add comments later on other paragraphs of your
posts.

Le 03-déc.-08, à 19:22, Brent Meeker a écrit :

> But there is causality. The sequence of events in the movie are
> directly caused
> by the projector, but they have a causal linkage back to Alice and the
> part of
> her environment that is captured in the movie. I see no principled
> reason to
> consider only the immediate cause and not refer back further in the
> chain of
> causation.

If this were true, I don't see why I could say yes to a doctor for an
artificial brain. I have to take account of the "traceability" of all
part of the artificial brain. You have a problem with the "qua
computatio" part of the MEC+MAT hypotheses, I think.
This is coherent with the fact that you have still some shyness with
the step six, if I remember well. They will be opportunity to come
back.

I have to go now.

Bruno

PS Abram. I think I will have to meditate a bit longer on your
(difficult) post. You may have a point (hopefully only pedagogical :)


http://iridia.ulb.ac.be/~marchal/

Abram Demski

unread,
Dec 4, 2008, 9:58:36 AM12/4/08
to everyth...@googlegroups.com
> PS Abram. I think I will have to meditate a bit longer on your
> (difficult) post. You may have a point (hopefully only pedagogical :)

A little bit more commentary may be in order then... I think my point
may be halfway between pedagogical and serious...

What I am saying is that people will come to the argument with some
vague idea of which computations (or which physical entities) they
pick out as "conscious". They will compare this to the various
hypotheses that come along during the argument-- MAT, MEC, MAT + MEC,
"Lucky Alice is conscious", "Lucky Alice is not conscious", et
cetera... These notions are necessarily 3rd-person in nature. It seems
like there is a problem there. Your argument is designed to talk about
1st-person phenomena.

If a 1st-person-perspective is a sort of structure (computational
and/or physical), what type of structure is it? If we define it in
terms of behavior only, then a recording is fine. If we define it in
terms of inner workings, then a recording is probably not fine, but we
introduce "magical" dependence on things that shouldn't matter to
us... ie, we should not care if we are interacting with a perfectly
orchestrated recording, so long as to us the result is the same.

It seems like this is independent of the differences between
pure-comp / comp+mat.

--Abram

Jason Resch

unread,
Dec 4, 2008, 9:50:18 PM12/4/08
to everyth...@googlegroups.com
On Thu, Dec 4, 2008 at 5:19 AM, Bruno Marchal <mar...@ulb.ac.be> wrote:


Hmmm... It means you have still a little problem with step seven. I
wish we share a computable environment, but we cannot decide this at
will.  I agree we have empirical evidence that here is such (partially)
computable environment, and I am willing to say I trust nature for
this. Yet, the fact is that to predict my next first person experience
I have to take into account ALL computations which exist in the
arithmetical "platonia" or in the universal dovetailing.


Bruno, I am with you that none of us can decide which of the infinite number of histories contain/compute us; when I talk about a universe I refer to just a single such history.  Perhaps you use history to refer only to the computational history that implements the observer's mind where I use it to mean an object which computes the mind of one or more observers in a consistent and fully definable way.

What I am not clear on with regards to your position is whether or not you believe most observers (if we could locate them in platonia from a 3rd person view) exist in environments larger than their brains, and likely containing numerous other observers or if you believe the mind is the only thing reified by computation and it is meaningless to discuss the environments they perceive because they don't exist. 

The way I see it, using the example of this physical universe only, it is far more probable for a mind to come about from the self-ordering properties of a universe such as this than for there to be a computation where the mind is an initial condition.  The program that implements the physics of this universe is likely to be far smaller than the program that implements our minds, or so my intuition leads me to believe.


>  I noticed in a previous post of yours you mentioned 'Kleene
> predicates' as a way of deriving computations from true statements, do
> you know of any good sources where I could learn more about Kleene
> predicates?

A very good introduction is the book by N.J. Cutland. See the reference
in my thesis. There are other books. I will think to make a list with
some comments. Actually I really love Kleene's original "Introduction
to Metamathematics", but the notations used  are a bit old fashioned.

Thanks Bruno, I will look into those.

Jason

Russell Standish

unread,
Dec 4, 2008, 9:56:50 PM12/4/08
to everyth...@googlegroups.com
On Wed, Dec 03, 2008 at 04:53:11PM +0100, Bruno Marchal wrote:
>
> I really don't know. I expect that the mathematical structure, as seen
> from inside, is so big that Platonia cannot have it neither as element
> nor as subpart. (Ah, well, I am aware that this is counter-intuitive,
> but here mathematical logic can help to see the consistency, and the
> quasi necessity with formal version of comp).
>

This point rather depends on what Platonia contains. If it contains
all sets of cardinality 2^{\aleph_0}, then the inside view of the
deployment will be conatained in it.

I do understand that your concept of Platonia (Arithmetic Realism I
believe you call it) is a Kronecker-like "God made the integers, all
the rest was made by man", and so what you say would be true of that.

Cheers

--

----------------------------------------------------------------------------
A/Prof Russell Standish Phone 0425 253119 (mobile)
Mathematics
UNSW SYDNEY 2052 hpc...@hpcoders.com.au
Australia http://www.hpcoders.com.au
----------------------------------------------------------------------------

Stathis Papaioannou

unread,
Dec 5, 2008, 3:05:55 AM12/5/08
to everyth...@googlegroups.com
2008/12/1 Abram Demski <abram...@gmail.com>:

> "Yes, consciousness supervenes on computation, but that computation
> needs to actually take place (meaning, physically). Otherwise, how

> could consciousness supervene on it? Now, in order for a computation


> to be physically instantiated, the physical instantiation needs to
> satisfy a few properties. One of these properties is clearly some sort
> of isomorphism between the computation and the physical instantiation:
> the actual steps of the computation are represented in physical form.
> A less obvious requirement is that the physical computation needs to

> have the proper counterfactuals: if some external force were to modify


> some step in the computation, the computation must progress according
> to the new computational state (as translated by the isomorphism)."

So if you destroy the counterfactual behaviour by removing components
that are not utilised, you end up with a recording-equivalent, which
isn't conscious. But what if you destroy the counterfactual behaviour
by another means? For example, if I wear a device that will instantly
kill me if I deviate from a particular behaviour, randomly determined
by the device from moment to moment, but survive, will my
consciousness be diminished as a result? You might say, no, because if
the device were not there I would have been able to handle the
counterfactuals. But then it might also be argued for the first
example that if the unused components had not been removed, the
recording-equivalent would also have been able to handle the
counterfactuals; and you can make this more concrete by having the
extra machinery waiting to be dropped into place in a counterfactual
universe.


--
Stathis Papaioannou

Bruno Marchal

unread,
Dec 5, 2008, 4:06:30 AM12/5/08
to everyth...@googlegroups.com
On 05 Dec 2008, at 03:50, Jason Resch wrote:



On Thu, Dec 4, 2008 at 5:19 AM, Bruno Marchal <mar...@ulb.ac.be> wrote:


Hmmm... It means you have still a little problem with step seven. I
wish we share a computable environment, but we cannot decide this at
will.  I agree we have empirical evidence that here is such (partially)
computable environment, and I am willing to say I trust nature for
this. Yet, the fact is that to predict my next first person experience
I have to take into account ALL computations which exist in the
arithmetical "platonia" or in the universal dovetailing.


Bruno, I am with you that none of us can decide which of the infinite number of histories contain/compute us; when I talk about a universe I refer to just a single such history.  


This is ambiguous. Even the unique "history-computation" of, let us say, the Everett Universal Wave, contains many (perhaps an infinity) of cosmic histories. There are there as many "Jason" than there are possible position of your electrons, even for equal energy level, and thus same molecular behavior, so that you cannot discern them, except in term or relative probabilities of (self) measurement outcomes. But then you have many others that we cannot eliminate, because even if you are right in assigning a bigger importance to little programs and their computations, the big programs occurring in the deployment have a role too, mainly due to the impossibility to be aware of delays made by the UD. Without the comp equivalent of "random phase annihilating the aberrantly long path", we have to take them into account, a priori.



Perhaps you use history to refer only to the computational history that implements the observer's mind where I use it to mean an object which computes the mind of one or more observers in a consistent and fully definable way.


It seems to me that we have to take them all into account, or justify why we can throw away the pure aberrant histories. If not, it looks like putting infinities and white rabbits under the rug, by decision. But then we are cheating with respect of taking the digital hypopthesis completely seriously. We could miss a refutation of comp, or important consequences. It seems to me.




What I am not clear on with regards to your position is whether or not you believe most observers (if we could locate them in platonia from a 3rd person view) exist in environments larger than their brains, and likely containing numerous other observers or if you believe the mind is the only thing reified by computation and it is meaningless to discuss the environments they perceive because they don't exist. 


Empirically I am rather sure environments plays a key role, yet, this remains to be proved. Strictly speaking I would say it is an open comp problem.




The way I see it, using the example of this physical universe only, it is far more probable for a mind to come about from the self-ordering properties of a universe such as this than for there to be a computation where the mind is an initial condition.  The program that implements the physics of this universe is likely to be far smaller than the program that implements our minds, or so my intuition leads me to believe.


Perhaps, but the whole point is that remains to be justify. It is *the* problem. If we assume comp, then we have to justify this. No doubt little programs play a key role, but the bigger one too, unless some destructive probability phenomenon occur. Now, interviewing the universal machine gives indeed a shadow of explanation of why such destructive phenomenon do occur indeed from the first person (plural) points of view of self-observing machine.
I mainly agree with what you "want", but we have to explain it.

Bruno




Bruno Marchal

unread,
Dec 5, 2008, 4:20:14 AM12/5/08
to everyth...@googlegroups.com

On 05 Dec 2008, at 03:56, Russell Standish wrote:

>
> On Wed, Dec 03, 2008 at 04:53:11PM +0100, Bruno Marchal wrote:
>>
>> I really don't know. I expect that the mathematical structure, as
>> seen
>> from inside, is so big that Platonia cannot have it neither as
>> element
>> nor as subpart. (Ah, well, I am aware that this is counter-intuitive,
>> but here mathematical logic can help to see the consistency, and the
>> quasi necessity with formal version of comp).
>>
>
> This point rather depends on what Platonia contains. If it contains
> all sets of cardinality 2^{\aleph_0}, then the inside view of the
> deployment will be conatained in it.

I am not sure. In my opinion, to have a platonia capable of describing
the first person views emerging from the UD entire work, even the
whole of Cantor Paradise will be too little. Even big cardinals (far
bigger than 2^(aleph_0)) will be like too constrained shoes. Actually
I believe that the first person views raised through the deployment
just escape the whole of human conceivable mathematics. It is big. But
it is also structured. It could even be structured as a person. I
don't know.


>
>
> I do understand that your concept of Platonia (Arithmetic Realism I
> believe you call it) is a Kronecker-like "God made the integers, all
> the rest was made by man", and so what you say would be true of that.


Yes the 3-Platonia can be very little, once we assume comp. But the
first view inside could be so big that eventually all notion of 1-
Platonia will happen to be inconsistent. It is for sure unameable (in
the best case). I discussed this a long time ago with George Levy: the
first person plenitude is big, very big, incredibly big. Nothing can
expressed or give an idea of that bigness.

At some point I will explain that the "divine intellect" of a lobian
machine as simple as Peano-Arithmetic is really far bigger than the
"God" of Peano-Arithmetic. I know it is bizarre (and a bit too
technical for being addressed right now I guess).

Have a good day,

Bruno


http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Dec 5, 2008, 4:36:01 AM12/5/08
to everyth...@googlegroups.com

On 04 Dec 2008, at 15:58, Abram Demski wrote:

>
>> PS Abram. I think I will have to meditate a bit longer on your
>> (difficult) post. You may have a point (hopefully only pedagogical :)
>
> A little bit more commentary may be in order then... I think my point
> may be halfway between pedagogical and serious...
>
> What I am saying is that people will come to the argument with some
> vague idea of which computations (or which physical entities) they
> pick out as "conscious". They will compare this to the various
> hypotheses that come along during the argument-- MAT, MEC, MAT + MEC,
> "Lucky Alice is conscious", "Lucky Alice is not conscious", et
> cetera... These notions are necessarily 3rd-person in nature. It seems
> like there is a problem there. Your argument is designed to talk about
> 1st-person phenomena.

The whole problem consists, assuming hypotheses, in relating 1-views
with 3-views.
In UDA, the 1-views are approximated by 1-discourses (personal diary
notes, memories in the brain, ...). But I do rely on the minimal
intuition needed to give sense to the willingness of saying "yes" to a
digitalist surgeon, and the believe in a comp survival, or a belief in
the unchanged feeling of "my" consciousness in such annihilation-
(re)creation experiences.


>
>
> If a 1st-person-perspective is a sort of structure (computational
> and/or physical), what type of structure is it?

The surprise will be: there are none. The 1-views of a machine will
appears to be already not expressible by the machine. The first and
third God have no name. Think about Tarski theorem in the comp
context. A sound machine cannot define the whole notion of "truth
about me".


> If we define it in
> terms of behavior only, then a recording is fine.

We certainly avoid the trap of behaviorism. You can see this as a
weakness, or as the full strong originality of comp, as I define it.
We give some sense, albeit undefined, to the word "consciousness"
apart from any behavior. But to reason we have to assume some relation
between consciousness and possible discourses (by machines).


> If we define it in
> terms of inner workings, then a recording is probably not fine, but we
> introduce "magical" dependence on things that shouldn't matter to
> us... ie, we should not care if we are interacting with a perfectly
> orchestrated recording, so long as to us the result is the same.
>
> It seems like this is independent of the differences between
> pure-comp / comp+mat.

This is not yet quite clear for me. Perhaps, if you are patient
enough, you will be able to clarify this along the UDA reasoning which
I will do slowly with Kim. The key point will be the understanding of
the ultimate conclusion: exactly like Everett can be said to justify
correctly the phenomenal collapse of the wave, if comp is assumed, we
have to justify in a similar way the wave itself. Assuming comp, we
put ourself in a position where we have to explain why numbers
develops stable and coherent belief in both mind and matter. We can
presuppose neither matter, nor mind eventually, except our own
consciousness, although even consciousness will eventually be reduced
into our "believe in numbers".

Bruno


http://iridia.ulb.ac.be/~marchal/

Abram Demski

unread,
Dec 5, 2008, 2:40:53 PM12/5/08
to everyth...@googlegroups.com
Stathis,

I think I can get around your objection by pointing out that the
structure of counterfactuals is quite different for a recording vs. a
full human who is wired to be killed if they deviate from a recording.
Someone could fairly easily disarm the killing device, whereas it
would be quite difficult to reconstruct the person from the recording
(in fact there is not enough information to do so).

A related way out would be to point out that all the computational
machinery is present in one case (merely disabled), whereas it is
totally absent in the other case.

--Abram

Abram Demski

unread,
Dec 5, 2008, 2:51:13 PM12/5/08
to everyth...@googlegroups.com
Bruno,

Are you asserting this based on published findings concerning
provability logic? If so, I would be very interested in references. If
not, then your results obviously seem publishable :). That is, if you
can show that huge amounts of set theory beyond ZFC emerge from
provability logic in some way...

Anyway, I'd definitely be interested in hearing those ideas.

--Abram

Abram Demski

unread,
Dec 5, 2008, 4:11:52 PM12/5/08
to everyth...@googlegroups.com
Bruno,

Perhaps all I am saying is that you need to state more explicitly the
assumptions about the connection between 1st and 3rd person, in both
MEC and MAT. Simply taking them to be the general ideas that you take
them to be does not obviously justify the argument.

Furthermore, stating the assumptions more clearly will make it more
clear where the contradiction is coming from, and thus which versions
of MEC and MAT the argument applies to.

--Abram

Stathis Papaioannou

unread,
Dec 5, 2008, 7:58:16 PM12/5/08
to everyth...@googlegroups.com
2008/12/6 Abram Demski <abram...@gmail.com>:

>
> Stathis,
>
> I think I can get around your objection by pointing out that the
> structure of counterfactuals is quite different for a recording vs. a
> full human who is wired to be killed if they deviate from a recording.
> Someone could fairly easily disarm the killing device, whereas it
> would be quite difficult to reconstruct the person from the recording
> (in fact there is not enough information to do so).

This seems to be getting away from the simple requirement that the
computer be able to handle counterfactuals. What if the device were
not easy to disarm, but almost impossible to disarm? What if it had
tentacles in every neurone, ready to destroy it if it fired at the
wrong time?

> A related way out would be to point out that all the computational
> machinery is present in one case (merely disabled), whereas it is
> totally absent in the other case.

So you agree that in the case where the extra machinery is waiting to
be dropped into place, consciousness results?


--
Stathis Papaioannou

Abram Demski

unread,
Dec 5, 2008, 9:00:52 PM12/5/08
to everyth...@googlegroups.com
Hi Stathis,

> This seems to be getting away from the simple requirement that the
> computer be able to handle counterfactuals. What if the device were
> not easy to disarm, but almost impossible to disarm? What if it had
> tentacles in every neurone, ready to destroy it if it fired at the
> wrong time?

I do think you have a point there. I began by equating counterfactual
structure with cause/effect structure, but now am drifting away from
that... So, can I make the point purely talking about causality?

I still think the answer may be yes...

The causal structure of a recording still looks far different from the
causal structure of a person that happens to follow a recording and
also happens to be wired to a machine that will kill them if they
deviate. Or, even, correct them if they deviate. (Let's go with that
so that I can't point out the simplistic difference "a recording will
not die if some external force causes it to deviate".)

1. Realistic malfunctions of a machine playing a recording are far
different from realistic malfunctions of the person-machine-combo. The
person inherits the possible malfunctions of the machine, *plus*
malfunctions in which the machine fails to modify the person's
behavior to match the recording. (A malfunction can be defined in
terms of cause-effect counterfactuals in two ways: first, if we think
that cause/effect is somewhat probabilistic, we will think that any
machine will occasionally malfunction; second, varying external
factors can cause malfunctions.)

2. Even during normal functioning, the cause/effect structure is very
different; the person-combo will have a lot of extra structure, since
it has a functioning brain and a corrective mechanism, neither needed
for the recording.

Also-- the level of the correction matters quite a bit I think. If
only muscle actions are being corrected, the person seems obviously
conscious-- lots of computations (& corresponding causal structure) is
still going on.. If each neuron is corrected, this is not so
intuitively obvious. (I suppose my intuition says that the person
would lose consciousness when the first correction occurred, though
that is silly upon reflection.)

How does that sound?

--Abram

Abram Demski

unread,
Dec 5, 2008, 9:01:47 PM12/5/08
to everyth...@googlegroups.com
Bruno,

Could you possibly link to the conversation with George Levy you refer
to? I did not find it looking on my own.

--Abram

Stathis Papaioannou

unread,
Dec 6, 2008, 4:05:25 AM12/6/08
to everyth...@googlegroups.com
2008/12/6 Abram Demski <abram...@gmail.com>:

Yes, there are these differences, but why should the differences be
relevant to the question of whether consciousness occurs or not? And
what about the case where the extra machinery that would allow the
right sort of causal structure but isn't actually used in a particular
situation is temporarily disengaged?

It seems to me that everyone contributing to these threads has an
intuition about consciousness, then works backwards from this:
"obviously, recordings aren't conscious; now what are the qualities
that recordings have which distinguish them from entities that are
conscious?". There's nothing intrinsically wrong with this method, but
it is possible to reach an impasse when the different parties have
different intuitions.

--
Stathis Papaioannou

Bruno Marchal

unread,
Dec 6, 2008, 9:07:51 AM12/6/08
to everyth...@googlegroups.com

Le 05-déc.-08, à 20:51, Abram Demski a écrit :

>
> Bruno,
>
> Are you asserting this based on published findings concerning
> provability logic? If so, I would be very interested in references. If
> not, then your results obviously seem publishable :).

I have published this in french a long time ago, but then I have
discovered that it has been publishe before by Montague and Kaplan (see
also Thomason). It is related to the fact that Knowledge, like truth
(cf Tarski), is not definable through an arithmetical predicate. In
"conscience and mécanisme" I illustrate a similar fact by using
(informally) the Lowenheim Skolem theorems.
Then I think the provability logic put a immense light on this, in a
transparently clear (arithmetical) frame, and that is a big part of my
thesis (the AUDA part).



> That is, if you
> can show that huge amounts of set theory beyond ZFC emerge from
> provability logic in some way...

I guess I have been unclear, because I am not saying that. I am saying
the more obvious (once we are familiar with incompleteness,
indefinissability, uncomputability etc) fact that a machine can infer
true but unprovable (by her) things about herself. It is just that a
provability machine, having furthermore inductive inference abilities
will generate more truth about itself than those which are provable by
the machine.


>
> Anyway, I'd definitely be interested in hearing those ideas.

Those ideas constitute the AUDA part. It is an abstract translation of
UDA in the language of "the" universal machine. It is needed to extract
constructively physics from computer science. I only get the
propositional physics (which is a billionth of "real" physics, yet I
got both the communicable physical logic and the uncommunicable
physical logic, that is both the quanta and the qualia. In that sense
it is already more than "usual physics", which (methodologically or
not) put the qualia and its subject under the rug.

Bruno
http://iridia.ulb.ac.be/~marchal/

Bruno Marchal

unread,
Dec 6, 2008, 9:32:53 AM12/6/08
to everyth...@googlegroups.com

Le 05-déc.-08, à 22:11, Abram Demski a écrit :

>
> Bruno,
>
> Perhaps all I am saying is that you need to state more explicitly the
> assumptions about the connection between 1st and 3rd person, in both
> MEC and MAT. Simply taking them to be the general ideas that you take
> them to be does not obviously justify the argument.


I don't see why nor how. The first person notions are defined in the
three first steps of the UDA. Wait I come back on this in the
discussion with Kim perhaps. In AUDA I define the first person by the
"knower", and I use the classical definition proposed by Theaetetus in
the Theaetetus of Plato. Keep in mind that you arrived when I was
explaining the real last step of an already long argument.
Of course you may be right, and I would really appreciate any
improvements. But making things more precise could also be a red
herring sometimes, or be very confusing pedagogically, like with the
easy 1004 fallacy which can obviously crop here.
When I defended the thesis in France, it was already a work resulting
from 30 years of discussions with open minded physicists, engineers,
philosophers and mathematicians, and I have learned that what seems
obvious for one of them is not for the others.
I don't think there is anything controversial in my work. I got
academical problems in Brussels for not having find an original result
(but then I think they did not read the work). Pedagogical difficulties
stem from the intrinsical difficulty of the mind body problem, and from
the technical abyss between logicians and physicists to cite only them.
It is more easy to collide two protons at the speed of light (minus
epsilon) than to arrange an appointment between mathematical logicians
and mathematical physicists (except perhaps nowadays on quantum
computing issues thankfully).


>
> Furthermore, stating the assumptions more clearly will make it more
> clear where the contradiction is coming from, and thus which versions
> of MEC and MAT the argument applies to.

I would be pleased if you can give me a version of MAT or MEC to which
the argument does not apply. For example, the argument applies to most
transfinite variant of MEC. It does not apply when some "magic" is
introduced in MAT, and MAT is hard to define in a way to exclude that
magic. If you can help, I thank you in advance.

Bruno
http://iridia.ulb.ac.be/~marchal/

Brent Meeker

unread,
Dec 6, 2008, 10:43:39 AM12/6/08
to everyth...@googlegroups.com

Exactly so. Consciousness is probably not the unified thing that we
intuitively assume anyway. There was an article in the newspaper today
that Henry Molaison died. He had lived some 50yrs with profound amnesia
after an operation on his brain to cure severe seizures. He apparently
could not form new memories. But that only applied to verbal, i.e.
"conscious" memories. He could learn new tasks in the sense that he
improved with practice even though if asked he would say he'd never done
the task before.

Brent Meeker

Abram Demski

unread,
Dec 6, 2008, 11:57:41 PM12/6/08
to everyth...@googlegroups.com
Stathis,

Yes, you are right. My main point is to show that such a point of view
is possible, not to actually argue for it... but I am largely just
asserting my intuitions nonetheless.

--Abram

Abram Demski

unread,
Dec 7, 2008, 12:19:25 AM12/7/08
to everyth...@googlegroups.com
Bruno,

Yes, I think there is a big difference between making an argument more
detailed and making it more understandable. They can go together or be
opposed. So a version of the argument targeted at my complaint might
not be good at all pedagogically...

> I would be pleased if you can give me a version of MAT or MEC to which
> the argument does not apply. For example, the argument applies to most
> transfinite variant of MEC. It does not apply when some "magic" is
> introduced in MAT, and MAT is hard to define in a way to exclude that
> magic. If you can help, I thank you in advance.

My particular brand of "magic" appears to be a requirement of
counterfactual/causal structure that reflects the
counterfactual/causal structure of (abstract) computation. Stathis has
pointed out some possible ways to show such ideas incoherent (which I
am not completely skeptical of, despite my arguments). Since this type
of theory is the type that matches my personal intuition, MGA will
feel empty to me until such alternatives are explicitly dealt a
killing blow (after which the rest is obvious, since I intuitively
feel the contradiction in versions of COMP+MAT that don't require
counterfactuals).

Of course, as you say, you'd be in a hard spot if you were required to
deal with every various intuition that anybody had... but, for what
it's worth, that is mine.

--Abram

Abram Demski

unread,
Dec 7, 2008, 12:23:10 AM12/7/08
to everyth...@googlegroups.com
Bruno,

Thanks, I will look up those names. If you have the time to reference
specific papers, I would be grateful.

--Abram

Brent Meeker

unread,
Dec 7, 2008, 12:43:16 AM12/7/08
to everyth...@googlegroups.com
Abram Demski wrote:
> Bruno,
>
> Yes, I think there is a big difference between making an argument more
> detailed and making it more understandable. They can go together or be
> opposed. So a version of the argument targeted at my complaint might
> not be good at all pedagogically...
>
>> I would be pleased if you can give me a version of MAT or MEC to which
>> the argument does not apply. For example, the argument applies to most
>> transfinite variant of MEC. It does not apply when some "magic" is
>> introduced in MAT, and MAT is hard to define in a way to exclude that
>> magic. If you can help, I thank you in advance.
>
> My particular brand of "magic" appears to be a requirement of
> counterfactual/causal structure that reflects the
> counterfactual/causal structure of (abstract) computation. Stathis has
> pointed out some possible ways to show such ideas incoherent (which I
> am not completely skeptical of, despite my arguments). Since this type
> of theory is the type that matches my personal intuition, MGA will
> feel empty to me until such alternatives are explicitly dealt a
> killing blow (after which the rest is obvious, since I intuitively
> feel the contradiction in versions of COMP+MAT that don't require
> counterfactuals).

My intuition is similar, except I think it is causality that is necessary,
rather than counterfactuals. I find persuasive the argument that the brains
potential for dealing with a counterfactual that never occurs cannot have any
bearing on consciousnes. After all there are infinitely many counterfactuals
that never occur (that's why they're counterfactual) and my brain is no doubt
unprepared to deal with most of them.

Causality, ISTM, is a physical relation and it is not captured by mathematical
or logical relations. That's probably why it has almost disappeared from
physics theories, which are highly mathematical. Bruno may think this is
invoking "magic", like Peter's insistence that "existence" is contingent. So be it.

Brent

Bruno Marchal

unread,
Dec 7, 2008, 1:12:26 PM12/7/08
to everyth...@googlegroups.com
On 07 Dec 2008, at 06:19, Abram Demski wrote:


Bruno,

Yes, I think there is a big difference between making an argument more
detailed and making it more understandable. They can go together or be
opposed. So a version of the argument targeted at my complaint might
not be good at all pedagogically...

I would be pleased if you can give me a version of MAT or MEC to which
the argument does not apply. For example, the argument applies to most
transfinite variant of MEC. It does not apply when some "magic" is
introduced in MAT, and MAT is hard to define in a way to exclude that
magic. If you can help, I thank you in advance.

My particular brand of "magic" appears to be a requirement of
counterfactual/causal structure that reflects the
counterfactual/causal structure of (abstract) computation.


Sometimes I think I should first explain what a "computation" is. I take it in the sense of theoretical computer science, a computation is always define relatively to a universal computation from outside, and an infinity of universal computations from inside. This asks for a bit of computer science. But there is not really "abstract computation", there are always relative computation (both with comp and Everett QM). They are always concrete relatively to the universal machine which execute them. The starting point in no important (for our fundamental concerns), you can take number with addition and multiplication, or lambda terms with abstraction and application. 




Stathis has
pointed out some possible ways to show such ideas incoherent (which I
am not completely skeptical of, despite my arguments).


I appreciate.



Since this type
of theory is the type that matches my personal intuition, MGA will
feel empty to me until such alternatives are explicitly dealt a
killing blow (after which the rest is obvious, since I intuitively
feel the contradiction in versions of COMP+MAT that don't require
counterfactuals).


Understanding UD(1...7) could perhaps help you to figure out what happens when we abandon the physical supervenience thesis, and embrace what remains, if keeping comp, that is the comp supervenience. It will explain how the physical laws have to emerge and why we believe (quasi-correctly) in brains.






Of course, as you say, you'd be in a hard spot if you were required to
deal with every various intuition that anybody had... but, for what
it's worth, that is mine.



I respect your intuition and appreciate the kind attitude. My feeling is that if front of very hard problems we have to be open to the fact that we could be surprised and that truth could be counterintuitive. The incompleteness phenomena, from Godel and Lob, are surprising and counterintuitive, and in the empirical world the SWE, whatever interpretation we find more plausible, is always rather counterintuitive too.

I interpret the "self-referentially correct scientist M" by the logic of Godel's provability predicates beweisbar_M. But the intuitive knower, the first person, is modelled (or defined) by the Theatetus trick: the machine M knows p in case "beweisbar_M('p') and p". Although extensionally equivalent, their are intensionally different. They prove the same arithmetical propositions, but they obey different logics. This is enough for showing that the first person associated with the self-referentially correct scientist will already disbelieve the comp hypothesis or find it very doubtful. We are near a paradox: the correct machine cannot know or believe their are machine. No doubt comp will appear counterintuitive for them. I know it is a sort of trap/ the solution consists in admitting that comp needs a strong act of faith, and I try to put light on the consequences for a machine, when she makes the bet.


The best reference on the self-reference logics are

Boolos, G. (1979). The unprovability of consistency. Cambridge University Press, London.
Boolos, G. (1993). The Logic of Provability. Cambridge University Press, Cambridge.
Smoryński, P. (1985). Self-Reference and Modal Logic. Springer Verlag, New York.
Smullyan, R. (1987). Forever Undecided. Knopf, New York.


The last one is a recreative book, not so simple, and rather quick in the "heart of the matter" chapter. Smullyan wrote many lovely  books, recreative and technical on that theme. 

The bible, imo, is Martin Davis book "The undecidable" which contains some of the original papers by Gödel, Church, Kleene, Post and indeed the most key starting points of the parts of theoretical computer science we are confonted to. It has been reedited by Dover.

Bruno

Other references here:

Russell Standish

unread,
Dec 7, 2008, 6:27:27 PM12/7/08
to everyth...@googlegroups.com
On Fri, Dec 05, 2008 at 10:06:30AM +0100, Bruno Marchal wrote:
>
>
> Perhaps, but the whole point is that remains to be justify. It is
> *the* problem. If we assume comp, then we have to justify this. No
> doubt little programs play a key role, but the bigger one too, unless
> some destructive probability phenomenon occur. Now, interviewing the
> universal machine gives indeed a shadow of explanation of why such
> destructive phenomenon do occur indeed from the first person (plural)
> points of view of self-observing machine.
> I mainly agree with what you "want", but we have to explain it.
>
> Bruno
>

Destructive phenomena do occur. To see this, realise that an infinite
set of histories will correspond to a given logical statement. Two
inconsistent statements can be combined disjunctively (A or B),
and their conjunction is false. Such a disjunction corresponds to the
union of the two sets of histories consistent with each statement. The
intersection of these sets of histories is, of course, empty.

So the measure of the histories consistent with A or B is now just
given by the sum of the measures of the two individual
statements. Since the information is given by the negative logarithm of these
measures, we see that the information of A or B is less than that of
either A or B taken separately. Information has been destroyed by
taking the inconsistent statements together.

It is this "triangle inequality" nature of information that gives rise
to the vector space structure in quantum mechanics.

Russell Standish

unread,
Dec 7, 2008, 6:59:56 PM12/7/08
to everyth...@googlegroups.com
On Sat, Dec 06, 2008 at 03:32:53PM +0100, Bruno Marchal wrote:
>
> I would be pleased if you can give me a version of MAT or MEC to which
> the argument does not apply. For example, the argument applies to most
> transfinite variant of MEC. It does not apply when some "magic" is
> introduced in MAT, and MAT is hard to define in a way to exclude that
> magic. If you can help, I thank you in advance.
>
> Bruno
>

Michael Lockwood distinguishes between materialism (consciousness
supervenes on the physical world) and physicalism (the physical world
suffices to explain everything). The difference between the two is
that in physicalism, consciousness (indeed any emergent phenomenon) is
mere epiphenomena, a computational convenience, but not necessary for
explanation, whereas in non-physicalist materialism, there are emergent
phenomena that are not explainable in terms of the underlying physics,
even though supervenience holds. This has been argued in the famous
paper by Philip Anderson. One very obvious distinction between
the two positions is that strong emergence is possible in materialism,
but strictly forbidden by physicalism. An example I give of strong
emergence in my book is the strong anthropic principle.

So - I'm convinced your argument works to show the contradiction
between COMP and physicalism, but not so the more general
materialism. I think you have confirmed this in some of your previous
responses to me in this thread.

Which is just as well. AFAICT, supervenience is the only thing
preventing the Occam catastrophe. We don't live in a magical world,
because such a world (assuming COMP) would have so many contradictory
statements that we'd disappear in a puff of destructive logic!
(reference to my previous posting about destructive phenomena).

Bruno Marchal

unread,
Dec 8, 2008, 3:43:47 AM12/8/08
to everyth...@googlegroups.com

On 08 Dec 2008, at 00:59, Russell Standish wrote:

>
> On Sat, Dec 06, 2008 at 03:32:53PM +0100, Bruno Marchal wrote:
>>
>> I would be pleased if you can give me a version of MAT or MEC to
>> which
>> the argument does not apply. For example, the argument applies to
>> most
>> transfinite variant of MEC. It does not apply when some "magic" is
>> introduced in MAT, and MAT is hard to define in a way to exclude that
>> magic. If you can help, I thank you in advance.
>>
>> Bruno
>>
>
> Michael Lockwood distinguishes between materialism (consciousness
> supervenes on the physical world) and physicalism (the physical world
> suffices to explain everything). The difference between the two is
> that in physicalism, consciousness (indeed any emergent phenomenon) is
> mere epiphenomena, a computational convenience, but not necessary for
> explanation, whereas in non-physicalist materialism, there are
> emergent
> phenomena that are not explainable in terms of the underlying physics,
> even though supervenience holds.

In what sense are they emergent? They emerge from what?


> This has been argued in the famous
> paper by Philip Anderson. One very obvious distinction between
> the two positions is that strong emergence is possible in materialism,
> but strictly forbidden by physicalism. An example I give of strong
> emergence in my book is the strong anthropic principle.
>
> So - I'm convinced your argument works to show the contradiction
> between COMP and physicalism, but not so the more general
> materialism.

I don't see why. When I state the supervenience thesis, I explain that
the type of supervenience does not play any role, be it a causal
relation or an epiphenomenon.


> I think you have confirmed this in some of your previous
> responses to me in this thread.
>
> Which is just as well. AFAICT, supervenience is the only thing
> preventing the Occam catastrophe. We don't live in a magical world,
> because such a world (assuming COMP) would have so many contradictory
> statements that we'd disappear in a puff of destructive logic!
> (reference to my previous posting about destructive phenomena).


I don' really understand. If such argument is correct, how could
classical logic not be quantum like. The problem of the white rabbits
is that they are consistent. Your explanation would make the world
quantum or not independently of the "degree of independence" of the
computational histories. Observation would not make a logic classical,
as it is the case in QM.

Bruno

http://iridia.ulb.ac.be/~marchal/

Abram Demski

unread,
Dec 9, 2008, 10:21:05 PM12/9/08
to everyth...@googlegroups.com
Bruno,

Thanks for the references.

--Abram

ps- it is final exam crunch time, so I haven't been checking email so
much as usual... I may get around to more detailed replies et cetera
this weekend or next week.

> Press, London.Boolos, G. (1993). The Logic of Provability. Cambridge
> University Press, Cambridge.Smoryński, P. (1985). Self-Reference and Modal
> Logic. Springer Verlag, New York.Smullyan, R. (1987). Forever Undecided.

Michael Rosefield

unread,
Dec 10, 2008, 5:39:34 AM12/10/08
to everyth...@googlegroups.com
This distinction between physicalism and materialism, with materialism allowing for features to emerge, it sounds to me like a join-the-dots puzzle - the physical substrate provides the dots, but the supervening system also contains lines - abstract structures implied by but not contained within the system implementing it. But does that not mean that this also implies further possible layers to the underlying reality? That no matter how many turtles you go down, there's always more turtles to come?

--------------------------
- Did you ever hear of "The Seattle Seven"?
- Mmm.
- That was me... and six other guys.


2008/12/7 Russell Standish <li...@hpcoders.com.au>

Bruno Marchal

unread,
Dec 10, 2008, 9:17:19 AM12/10/08
to everyth...@googlegroups.com
Abram Demski wrote:

Bruno,

Thanks for the references.

You are welcome.


ps- it is final exam crunch time, so I haven't been checking email so
much as usual... I may get around to more detailed replies et cetera
this weekend or next week.

With pleasure.

Best,

Bruno

Russell Standish

unread,
Dec 11, 2008, 1:57:20 AM12/11/08
to everyth...@googlegroups.com
On Wed, Dec 10, 2008 at 10:39:34AM +0000, Michael Rosefield wrote:
> This distinction between physicalism and materialism, with materialism
> allowing for features to emerge, it sounds to me like a join-the-dots puzzle
> - the physical substrate provides the dots, but the supervening system also
> contains lines - abstract structures implied by but not contained within the
> system implementing it. But does that not mean that this also implies
> further possible layers to the underlying reality? That no matter how many
> turtles you go down, there's always more turtles to come?
>

I don't think it implies it, but it is certainly possible. Emergence
is possible with just two incommensurate levels.

Cheers

Russell Standish

unread,
Dec 11, 2008, 4:24:51 PM12/11/08
to everyth...@googlegroups.com
On Mon, Dec 08, 2008 at 09:43:47AM +0100, Bruno Marchal wrote:
> >
> > Michael Lockwood distinguishes between materialism (consciousness
> > supervenes on the physical world) and physicalism (the physical world
> > suffices to explain everything). The difference between the two is
> > that in physicalism, consciousness (indeed any emergent phenomenon) is
> > mere epiphenomena, a computational convenience, but not necessary for
> > explanation, whereas in non-physicalist materialism, there are
> > emergent
> > phenomena that are not explainable in terms of the underlying physics,
> > even though supervenience holds.
>
> In what sense are they emergent? They emerge from what?

They emerge from the underlying physics (or chemistry, or whatever the
syntactic layer is). Supervenience is AFAICT nothing other than the
concept of emergence applied to consciousness. In many respects it
could be considered to be synonymous.

>
>
> > This has been argued in the famous
> > paper by Philip Anderson. One very obvious distinction between
> > the two positions is that strong emergence is possible in materialism,
> > but strictly forbidden by physicalism. An example I give of strong
> > emergence in my book is the strong anthropic principle.
> >
> > So - I'm convinced your argument works to show the contradiction
> > between COMP and physicalism, but not so the more general
> > materialism.
>
> I don't see why. When I state the supervenience thesis, I explain that
> the type of supervenience does not play any role, be it a causal
> relation or an epiphenomenon.
>

In your Lille thesis (sorry I still haven't read your Brussels thesis)
you say at the end of section 4.4.1 that SUP-PHYS supposes at minimum
a concrete physical world. I don't see how this follows at all from
the concept of supervenience, but I accept that it is necessary for
(naive) physicalism.

>
> > I think you have confirmed this in some of your previous
> > responses to me in this thread.
> >
> > Which is just as well. AFAICT, supervenience is the only thing
> > preventing the Occam catastrophe. We don't live in a magical world,
> > because such a world (assuming COMP) would have so many contradictory
> > statements that we'd disappear in a puff of destructive logic!
> > (reference to my previous posting about destructive phenomena).
>
>
> I don' really understand. If such argument is correct, how could
> classical logic not be quantum like. The problem of the white rabbits
> is that they are consistent.

Sorry, to be clear - the white rabbits themselves are consistent, and
also also quite rare (ie improbable). However they also tend to come
in "equal and opposite" (ie contradictory) forms so when combined
contribute to the measure of a non-magical world. That is
information destructve phenomena.

As for logic, each individual observer sees a world according to
classical logic. Only by quantifying over multiple observers does
quantum logic come into play. This is a key point I make on page 219
of my book. I'm sorry I haven't found the best way to express the
argument yet - it really is quite subtle. I know Youness had
difficulties with this aspect as well.

I apologise - I have been speaking in coded sentences which require a
deal of unpacking if you are unfamiliar with the concepts. But I'm in
good company here...

Reply all
Reply to author
Forward
0 new messages