Dear All,
I am new to POMDPs and I am trying to use one for making biopsy decisions in prostate cancer screening schedules. From what I understood so far from the examples given on GitHub, is that transition probability matrices need to be defined on the basis of current state and action pair. However, in my case I require knowing the time at which this action has to be made, I also require knowing the history of actions and observations. That is my transition probability function is not just p(s_{t+1} | s_t, a_t), but is rather p(s_{t+1} | s_t, a_0, a_1, ..., a_{t-1}, a_t, y_0, y_1, ... y_t), where s_{t+1} is the state at next decision epoch t+1, and s_t is the state at the current decision epoch, a_0, a_1, .... a_{t-1} constitute action history up to time t, and a_t is potential new action, and y_0, y_1, .... y_t constitute observation history up to time t.
Is it possible to get access to these histories while defining my function for transition probability matrix or matrix for observation probabilities? Currently as given in the TigerPOMDP example (below), it seems only current action and state pair are available.
function POMDPs.transition(pomdp::TigerPOMDP, s::Bool, a::Symbol)
if a == :open_left || a == :open_right
# problem resets
return BoolDistribution(0.5)
elseif s
# tiger on the left stays on the left
return BoolDistribution(1.0)
else
return BoolDistribution(0.0)
end
endThank you,
Ani