package scalarl
- Alphabetic
- By Inheritance
- scalarl
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Type Members
-
trait
ActionValueFn[Obs, A, T] extends AnyRef
Along with StateValueFn, this is the main trait in tabular reinforcement learning for tracking the value of an (observation, action) pair.
Along with StateValueFn, this is the main trait in tabular reinforcement learning for tracking the value of an (observation, action) pair.
- Obs
Observation returned by the State instances tracked by ActionValueFn.
- A
Action type on the State instances tracked by ActionValueFn.
- T
type of values tracked by ActionValueFn.
- trait Agent[Obs, A, R, T, M[_]] extends AnyRef
- type Cat[+T] = Categorical[T]
-
trait
Policy[Obs, A, R, M[_], S[_]] extends AnyRef
This is how agents actually choose what comes next.
This is how agents actually choose what comes next. This is a stochastic policy. We have to to be able to match this up with a state that has the same monadic return type, but for now it's hardcoded.
A - Action Obs - the observation offered by this state. R - reward M - the monadic type offered by the policy. S - the monad for the state.
-
final
case class
SARS[Obs, A, R, S[_]](state: State[Obs, A, R, S], action: A, reward: R, nextState: State[Obs, A, R, S]) extends Product with Serializable
Chunk that you get back for playing an episode.
-
trait
State[Obs, A, R, M[_]] extends AnyRef
A world should probably have a generator of states and actions...
A world should probably have a generator of states and actions... and then you can use that to get to the next thing. The state here is going to be useful in the Markov model; for the bandit we only have a single state, not that useful.
-
trait
StateValueFn[Obs, T] extends AnyRef
Along with ActionValueFn, this is the main trait in tabular reinforcement learning for tracking the value of a state as evidenced by the observation it returns.
Along with ActionValueFn, this is the main trait in tabular reinforcement learning for tracking the value of a state as evidenced by the observation it returns.
We need some way for this to learn, or see new observations, that's part of the trait.
- Obs
Observation returned by the State instances tracked by StateValueFn.
- T
type of values tracked by StateValueFn.
- final case class Time(value: Long) extends AnyVal with Product with Serializable
Value Members
-
object
ActionValueFn
Constructors and classes associated with ActionValueFn.
- object Agent
-
object
Evaluator
Contains traits and instances for the two evaluation methods.
- object Policy
- object State
-
object
StateValueFn
Constructors and classes associated with StateValueFn.
- object Time extends Serializable
- object Util
edit this text on github
ScalaRL
This is the API documentation for the ScalaRL functional reinforcement learning library.
Further documentation for ScalaRL can be found at the documentation site.
Check out the ScalaRL package list for all the goods.