Hi,
Does Burlap support offline matched RL learning?
LSPI supports to learn from saved episodes.
How about other learning algorithms, td(lambda) ?
When I read it, other learning algorithms interact with env to get state, execute action and get rewards and next state.
Do they support offline training?
How to modify them to get offline training and use vfa as well?
Thanks in advance!