--
You received this message because you are subscribed to the Google Groups "Stanford AI Class" group.
To post to this group, send email to stanford...@googlegroups.com.
To unsubscribe from this group, send email to stanford-ai-cl...@googlegroups.com.
For more options, visit this group at http://groups.google.com/group/stanford-ai-class?hl=en.
On the basis of this, I got this wrong for two questions:
For the coin toss, each throw is independent from the next => no.
For the maze, each move requires knowledge about previous moves => yes.
Fully observable means you can know the whole context. Stochastic means you
don't have a pattern, it's more like "a chance", even if you have or don't a
context.
You get more info from each coin throw.
> For the maze, each move requires knowledge about previous moves => yes.
It doesn't:
"Your controls are move forward to the next wall or intersection, turn
left 90 degrees, and turn right 90 degrees." -- so there is no
backtracking possible. The worst someone could do would be to turn
e.g. right after you had just turned left. But after the first move
there's no harm done. No?
-Jodi
1) definition of Fully and Partial Observability:
"...An environment is fully observable if the sensors can always see the entire state of the environment. It's partially observable if the sensors can only see a fraction of the state, ..." - http://www.youtube.com/watch?feature=player_embedded&v=5lcLmhsmBnQ#t=94s
2) "Every Stochastic problem is defined as having a Partially Observable environment, because the randomness of actions or percepts is considered the result of a hidden element in that environment."This is not correct.
> ...
>
> mais »
> ...
>
> mais »
That would be wonderful, except for the fact that the course definitions are so imprecise that they don't tell us what constitutes state and what doesn't. How can we assign a meaning to "if the sensors can always see the entire state of the environment" if we don't know what is allowable as "state"?
The percepts that reach us indicate the presence of only two discrete states in the environment, and we certainly have not been told that the process that results in loading of the coin operates through hidden state. It's quite a stretch to suggest that loading implies hidden state when we can represent everything that we need to represent about a loaded die as an outcome probability on the throw, ie. the action. No hidden state needed. It's as silly as defining an "ether" in centuries past, not needed at all.
<CUT>
@StanfordAIClass It would be nicer if you would link to the explicit thread instead of to the home page of this google group
Being silly and blind to the obvious, my scenarios above seem to make sense to me.
> ...
>
> read more »
There is no way to determine Partial Observability without knowing the problem/goal of the Agent, because it is about the ability of the model to allow the goal to be satisfied, and it is always in the context of what is practical or workable.
"You're task [is] to understand from coin flips whether a coin is
loaded and if so at what probability..."
In order to work out if a coin is loaded or not requires the coin
flipper to take note of what the flips came out as, hence the task is
partially observable. It cannot be determined by any single flip.
Stuart
Excellent, Raf!
Your diagram and explanation are indeed helpful (extremely helpful), but not in the way you expect. They are helpful because they illustrate perfectly that there are no additional hidden states in the environment other than the two states T and C which are directly visible through percepts. This is true in both the case of a fair coin and that of a loaded coin. I am sure you agree with this, since you have not added any additional environment states.
Instead, what you have done is to create a search state space of your own, separate from the environment, in which each state of the search space has a name or attribute indicating the number of times that a particular percept has been seen. This search state space of yours (which is an great suggestion by the way) is an infinite state space, because the problem specification has not requested that we return the probability with any particular value of precision or error bound, so the tree has no lower boundary. An actual implementation could of course allow our calculation of probability to be read out at any time or at multiple times, but nevertheless the search state space is infinite, since we do not have a termination condition. And it's an infinite space in principle as well.