optimal_action {pomdp} | R Documentation |
Optimal action for a belief
Description
Determines the optimal action for a policy (solved POMDP) for a given belief at a given epoch.
Usage
optimal_action(model, belief = NULL, epoch = 1)
Arguments
model |
a solved POMDP. |
belief |
The belief (probability distribution over the states) as a
vector or a matrix with multiple belief states as rows. If |
epoch |
what epoch of the policy should be used. Use 1 for converged policies. |
Value
The name of the optimal action.
Author(s)
Michael Hahsler
See Also
Other policy:
estimate_belief_for_nodes()
,
plot_belief_space()
,
plot_policy_graph()
,
policy()
,
policy_graph()
,
projection()
,
reward()
,
solve_POMDP()
,
solve_SARSOP()
,
value_function()
Examples
data("Tiger")
Tiger
sol <- solve_POMDP(model = Tiger)
# these are the states
sol$states
# belief that tiger is to the left
optimal_action(sol, c(1, 0))
optimal_action(sol, "tiger-left")
# belief that tiger is to the right
optimal_action(sol, c(0, 1))
optimal_action(sol, "tiger-right")
# belief is 50/50
optimal_action(sol, c(.5, .5))
optimal_action(sol, "uniform")
# the POMDP is converged, so all epoch give the same result.
optimal_action(sol, "tiger-right", epoch = 10)
[Package pomdp version 1.2.3 Index]