- e1 - Variable in class burlap.domain.stochasticgames.gridgame.state.GGWall
-
- e2 - Variable in class burlap.domain.stochasticgames.gridgame.state.GGWall
-
- ECorrelatedQJointPolicy - Class in burlap.behavior.stochasticgames.madynamicprogramming.policies
-
A joint policy that computes the correlated equilibrium using the Q-values of the agents as input and then either
follows that policy or returns a random action with probability epsilon.
- ECorrelatedQJointPolicy(double) - Constructor for class burlap.behavior.stochasticgames.madynamicprogramming.policies.ECorrelatedQJointPolicy
-
Initializes with the epislon probability of a random joint action.
- ECorrelatedQJointPolicy(CorrelatedEquilibriumSolver.CorrelatedEquilibriumObjective, double) - Constructor for class burlap.behavior.stochasticgames.madynamicprogramming.policies.ECorrelatedQJointPolicy
-
Initializes with the correlated equilibrium objective and the epsilon probability of a random joint action.
- EGreedyJointPolicy - Class in burlap.behavior.stochasticgames.madynamicprogramming.policies
-
An epsilon greedy joint policy, in which the joint action with the highest Q-value for a given target agent is returned a 1-epsilon fraction
of the time, and a random joint action an epsilon fraction of the time.
- EGreedyJointPolicy(double) - Constructor for class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyJointPolicy
-
Initializes for a given epsilon value.
- EGreedyJointPolicy(MultiAgentQLearning, double, int) - Constructor for class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyJointPolicy
-
Initializes for a multi-agent Q-learning object and epsilon value.
- EGreedyMaxWellfare - Class in burlap.behavior.stochasticgames.madynamicprogramming.policies
-
An epsilon greedy joint policy, in which the joint aciton with the highest aggregate Q-values for each agent is returned a 1-epsilon fraction of the time and a random
joint action an epsilon fraction of the time.
- EGreedyMaxWellfare(double) - Constructor for class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyMaxWellfare
-
Initializes for a given epsilon value.
- EGreedyMaxWellfare(double, boolean) - Constructor for class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyMaxWellfare
-
Initializes for a given epsilon value and whether to break ties randomly.
- EGreedyMaxWellfare(MultiAgentQLearning, double) - Constructor for class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyMaxWellfare
-
Initializes for a multi-agent Q-learning object and epsilon value.
- EGreedyMaxWellfare(MultiAgentQLearning, double, boolean) - Constructor for class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyMaxWellfare
-
Initializes for a multi-agent Q-learning object and epsilon value.
- eligibility - Variable in class burlap.behavior.singleagent.learning.actorcritic.critics.TDLambda.StateEligibilityTrace
-
The eligibility value
- eligibility - Variable in class burlap.behavior.singleagent.learning.tdmethods.SarsaLam.EligibilityTrace
-
The eligibility value
- EligibilityTrace(HashableState, QValue, double) - Constructor for class burlap.behavior.singleagent.learning.tdmethods.SarsaLam.EligibilityTrace
-
Creates a new eligibility trace to track for an episode.
- EligibilityTraceVector(int, double, double) - Constructor for class burlap.behavior.singleagent.learning.tdmethods.vfa.GradientDescentSarsaLam.EligibilityTraceVector
-
Creates a trace for the given weight with the given eligibility value
- eligibilityValue - Variable in class burlap.behavior.singleagent.learning.tdmethods.vfa.GradientDescentSarsaLam.EligibilityTraceVector
-
The eligibility value
- EMinMaxPolicy - Class in burlap.behavior.stochasticgames.madynamicprogramming.policies
-
Class for following a minmax joint policy.
- EMinMaxPolicy(double) - Constructor for class burlap.behavior.stochasticgames.madynamicprogramming.policies.EMinMaxPolicy
-
Initializes for a given epsilon value; the fraction of the time a random joint action is selected
- EMinMaxPolicy(MultiAgentQLearning, double, int) - Constructor for class burlap.behavior.stochasticgames.madynamicprogramming.policies.EMinMaxPolicy
-
Initializes for a given Q-learning agent and epsilon value.
- encodePlanIntoPolicy(SearchNode) - Method in class burlap.behavior.singleagent.planning.deterministic.DeterministicPlanner
-
Encodes a solution path found by the valueFunction into this class's internal policy structure.
- endAllAgents() - Method in class burlap.behavior.singleagent.auxiliary.performance.PerformancePlotter
-
Informs the plotter that all data for all agents has been collected.
- endAllTrials() - Method in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentPerformancePlotter
-
Specifies that all trials are complete and that the average trial results and error bars should be plotted.
- endAllTrialsForAgent(String) - Method in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentPerformancePlotter
-
Ends all the trials, plotting the average trial data for the agent with the given name
- endAllTrialsHelper() - Method in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentPerformancePlotter
-
The end all trial methods helper called at the end of a swing update.
- endEpisode() - Method in class burlap.behavior.singleagent.auxiliary.performance.PerformancePlotter
-
Informs the plotter that all data for the last episode has been collected.
- endEpisode() - Method in interface burlap.behavior.singleagent.learning.actorcritic.Critic
-
This method is called whenever a learning episode terminates
- endEpisode() - Method in class burlap.behavior.singleagent.learning.actorcritic.critics.TDLambda
-
- endEpisode() - Method in class burlap.behavior.singleagent.learning.actorcritic.critics.TimeIndexedTDLambda
-
- endTrial() - Method in class burlap.behavior.singleagent.auxiliary.performance.PerformancePlotter
-
Informs the plotter that all data for the current trial as been collected.
- endTrial() - Method in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentPerformancePlotter
-
Ends the current trial data and updates the plots accordingly.
- endTrialsForCurrentAgent() - Method in class burlap.behavior.singleagent.auxiliary.performance.PerformancePlotter
-
Informs the plotter that all trials for the current agent have been collected and causes the average plots to be set and displayed.
- entrySet() - Method in class burlap.datastructures.HashedAggregator
-
The entry set for stored keys and values.
- enumerable() - Method in class burlap.behavior.singleagent.options.SubgoalOption
-
Returns true if the initiation states and termination states of this option are iterable; false if either of them are not.
- EnumerableBeliefState - Interface in burlap.mdp.singleagent.pomdp.beliefstate
-
An interface to be used by
BeliefState
implementations that also can enumerate
the set of states that have probability mass.
- EnumerableBeliefState.StateBelief - Class in burlap.mdp.singleagent.pomdp.beliefstate
-
A class for specifying the probability mass of an MDP state in a
BeliefState
.
- EnumerablePolicy - Interface in burlap.behavior.policy
-
An interface extension to
Policy
for policies that can enumerate their probability distribution over all actions.
- enumeration - Variable in class burlap.behavior.singleagent.auxiliary.StateEnumerator
-
The forward state enumeration map
- env - Variable in class burlap.shell.EnvironmentShell
-
- env - Variable in class burlap.shell.visual.VisualExplorer
-
- env_cleanup() - Method in class burlap.domain.singleagent.rlglue.RLGlueEnvironment
-
- env_init() - Method in class burlap.domain.singleagent.rlglue.RLGlueEnvironment
-
- env_message(String) - Method in class burlap.domain.singleagent.rlglue.RLGlueEnvironment
-
- env_start() - Method in class burlap.domain.singleagent.rlglue.RLGlueEnvironment
-
- env_step(Action) - Method in class burlap.domain.singleagent.rlglue.RLGlueEnvironment
-
- Environment - Interface in burlap.mdp.singleagent.environment
-
Environments define a current observation represetned with a
State
and manage state and reward transitions when an action is executed in the environment through
the
Environment.executeAction(Action)
method.
- environment - Variable in class burlap.mdp.singleagent.pomdp.BeliefAgent
-
The POMDP environment.
- EnvironmentDelegation - Interface in burlap.mdp.singleagent.environment.extensions
-
- EnvironmentDelegation.Helper - Class in burlap.mdp.singleagent.environment.extensions
-
A class that provides tools for working with Environment delegates
- EnvironmentObserver - Interface in burlap.mdp.singleagent.environment.extensions
-
A class that is told of interactions in an environment.
- EnvironmentOptionOutcome - Class in burlap.behavior.singleagent.options
-
- EnvironmentOptionOutcome(State, Action, State, double, boolean, double, Episode) - Constructor for class burlap.behavior.singleagent.options.EnvironmentOptionOutcome
-
Initializes.
- EnvironmentOutcome - Class in burlap.mdp.singleagent.environment
-
A tuple for specifying the outcome of executing an action in an
Environment
.
- EnvironmentOutcome(State, Action, State, double, boolean) - Constructor for class burlap.mdp.singleagent.environment.EnvironmentOutcome
-
Initializes.
- EnvironmentServer - Class in burlap.mdp.singleagent.environment.extensions
-
- EnvironmentServer(Environment, EnvironmentObserver...) - Constructor for class burlap.mdp.singleagent.environment.extensions.EnvironmentServer
-
- EnvironmentServer.StateSettableEnvironmentServer - Class in burlap.mdp.singleagent.environment.extensions
-
- EnvironmentServerInterface - Interface in burlap.mdp.singleagent.environment.extensions
-
- environmentSever - Variable in class burlap.behavior.singleagent.auxiliary.performance.LearningAlgorithmExperimenter
-
- EnvironmentShell - Class in burlap.shell
-
- EnvironmentShell(Domain, Environment, InputStream, PrintStream) - Constructor for class burlap.shell.EnvironmentShell
-
- EnvironmentShell(Domain, Environment) - Constructor for class burlap.shell.EnvironmentShell
-
Creates shell for std in and std out.
- EnvironmentShell(SADomain, State) - Constructor for class burlap.shell.EnvironmentShell
-
- eo - Variable in class burlap.mdp.singleagent.model.TransitionProb
-
The transition
- Episode - Class in burlap.behavior.singleagent
-
This class is used to keep track of all events that occur in an episode.
- Episode() - Constructor for class burlap.behavior.singleagent.Episode
-
Creates a new EpisodeAnalysis object.
- Episode(State) - Constructor for class burlap.behavior.singleagent.Episode
-
Initializes a new EpisodeAnalysis object with the initial state in which the episode started.
- episode - Variable in class burlap.behavior.singleagent.options.EnvironmentOptionOutcome
-
The executed episode from this execution
- episode - Variable in class burlap.behavior.singleagent.options.model.BFSNonMarkovOptionModel.NonMarkovOptionScanNode
-
- EpisodeBrowserCommand() - Constructor for class burlap.shell.command.env.EpisodeRecordingCommands.EpisodeBrowserCommand
-
- episodeFiles - Variable in class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
-
- episodeFiles - Variable in class burlap.behavior.stochasticgames.auxiliary.GameSequenceVisualizer
-
- episodeHistory - Variable in class burlap.behavior.singleagent.learning.actorcritic.ActorCritic
-
The saved and most recent learning episodes this agent has performed.
- episodeHistory - Variable in class burlap.behavior.singleagent.learning.lspi.LSPI
-
the saved previous learning episodes
- episodeHistory - Variable in class burlap.behavior.singleagent.learning.modellearning.artdp.ARTDP
-
the saved previous learning episodes
- episodeHistory - Variable in class burlap.behavior.singleagent.learning.modellearning.rmax.PotentialShapedRMax
-
the saved previous learning episodes
- episodeList - Variable in class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
-
- episodeList - Variable in class burlap.behavior.stochasticgames.auxiliary.GameSequenceVisualizer
-
- EpisodeRecordingCommands - Class in burlap.shell.command.env
-
Two
ShellCommand
s, rec and episode, for recording and browsing episodes of behavior that take place in the
Environment
.
- EpisodeRecordingCommands() - Constructor for class burlap.shell.command.env.EpisodeRecordingCommands
-
- EpisodeRecordingCommands.EpisodeBrowserCommand - Class in burlap.shell.command.env
-
- EpisodeRecordingCommands.RecordCommand - Class in burlap.shell.command.env
-
- episodes - Variable in class burlap.shell.command.env.EpisodeRecordingCommands
-
- episodeScroller - Variable in class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
-
- episodeScroller - Variable in class burlap.behavior.stochasticgames.auxiliary.GameSequenceVisualizer
-
- EpisodeSequenceVisualizer - Class in burlap.behavior.singleagent.auxiliary
-
This class is used to visualize a set of episodes that have been saved to files in a common directory or which are
provided to the object as a list of
Episode
objects.
- EpisodeSequenceVisualizer(Visualizer, Domain, String) - Constructor for class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
-
Initializes the EpisodeSequenceVisualizer.
- EpisodeSequenceVisualizer(Visualizer, Domain, String, int, int) - Constructor for class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
-
Initializes the EpisodeSequenceVisualizer.
- EpisodeSequenceVisualizer(Visualizer, Domain, List<Episode>) - Constructor for class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
-
Initializes the EpisodeSequenceVisualizer with a programatically supplied list of
Episode
objects to view.
- EpisodeSequenceVisualizer(Visualizer, Domain, List<Episode>, int, int) - Constructor for class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
-
Initializes the EpisodeSequenceVisualizer with a programatically supplied list of
Episode
objects to view.
- episodesListModel - Variable in class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
-
- episodesListModel - Variable in class burlap.behavior.stochasticgames.auxiliary.GameSequenceVisualizer
-
- episodeWeights - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.MLIRLRequest
-
The weight assigned to each episode.
- epsilon - Variable in class burlap.behavior.functionapproximation.dense.rbf.functions.GaussianRBF
-
The bandwidth parameter.
- epsilon - Variable in class burlap.behavior.policy.EpsilonGreedy
-
- epsilon - Variable in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearningRequest
-
The maximum feature score to cause termination of Apprenticeship learning
- epsilon - Variable in class burlap.behavior.singleagent.planning.deterministic.informed.astar.DynamicWeightedAStar
-
parameter > 1 indicating the maximum amount of greediness; the larger the more greedy.
- epsilon - Variable in class burlap.behavior.stochasticgames.agents.naiveq.history.SGQWActionHistoryFactory
-
The epislon value for epislon greedy policy.
- epsilon - Variable in class burlap.behavior.stochasticgames.madynamicprogramming.policies.ECorrelatedQJointPolicy
-
The epsilon parameter specifying how often random joint actions are returned
- epsilon - Variable in class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyJointPolicy
-
The epsilon parameter specifying how often random joint actions are returned
- epsilon - Variable in class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyMaxWellfare
-
The epsilon parameter specifying how often random joint actions are returned
- epsilon - Variable in class burlap.behavior.stochasticgames.madynamicprogramming.policies.EMinMaxPolicy
-
The epsilon parameter specifying how often random joint actions are returned
- EpsilonGreedy - Class in burlap.behavior.policy
-
This class defines a an epsilon-greedy policy over Q-values and requires a QComputable valueFunction to be specified.
- EpsilonGreedy(double) - Constructor for class burlap.behavior.policy.EpsilonGreedy
-
Initializes with the value of epsilon, where epsilon is the probability of taking a random action.
- EpsilonGreedy(QProvider, double) - Constructor for class burlap.behavior.policy.EpsilonGreedy
-
Initializes with the QComputablePlanner to use and the value of epsilon to use, where epsilon is the probability of taking a random action.
- epsilonP1 - Variable in class burlap.behavior.singleagent.planning.deterministic.informed.astar.StaticWeightedAStar
-
The > 1 epsilon parameter.
- epsilonWeight(int) - Method in class burlap.behavior.singleagent.planning.deterministic.informed.astar.DynamicWeightedAStar
-
Returns the weighted epsilon value at the given search depth
- equals(Object) - Method in class burlap.behavior.functionapproximation.FunctionGradient.PartialDerivative
-
- equals(Object) - Method in class burlap.behavior.functionapproximation.sparse.tilecoding.Tiling.FVTile
-
- equals(Object) - Method in class burlap.behavior.policy.support.AnnotatedAction
-
- equals(Object) - Method in class burlap.behavior.singleagent.interfaces.rlglue.RLGlueDomain.RLGlueActionType.RLGLueAction
-
- equals(Object) - Method in class burlap.behavior.singleagent.interfaces.rlglue.RLGlueState.RLGlueVarKey
-
- equals(Object) - Method in class burlap.behavior.singleagent.options.MacroAction
-
- equals(Object) - Method in class burlap.behavior.singleagent.options.SubgoalOption
-
- equals(Object) - Method in class burlap.behavior.singleagent.planning.deterministic.informed.PrioritizedSearchNode
-
- equals(Object) - Method in class burlap.behavior.singleagent.planning.deterministic.SearchNode
-
- equals(Object) - Method in class burlap.behavior.singleagent.planning.stochastic.montecarlo.uct.UCTStateNode
-
- equals(Object) - Method in class burlap.behavior.singleagent.planning.stochastic.sparsesampling.SparseSampling.HashedHeightState
-
- equals(Object) - Method in class burlap.behavior.singleagent.planning.stochastic.valueiteration.PrioritizedSweeping.BPTRNode
-
- equals(Object) - Method in class burlap.domain.singleagent.graphdefined.GraphDefinedDomain.GraphActionType.GraphAction
-
- equals(Object) - Method in class burlap.domain.singleagent.graphdefined.GraphDefinedDomain.NodeTransitionProbability
-
- equals(Object) - Method in class burlap.domain.singleagent.gridworld.GridWorldTerminalFunction.IntPair
-
- equals(Object) - Method in class burlap.domain.singleagent.lunarlander.LunarLanderDomain.ThrustType.ThrustAction
-
- equals(Object) - Method in class burlap.domain.singleagent.pomdp.tiger.TigerState
-
- equals(Object) - Method in class burlap.domain.stochasticgames.normalform.SingleStageNormalFormGame.MatrixAction
-
- equals(Object) - Method in class burlap.domain.stochasticgames.normalform.SingleStageNormalFormGame.StrategyProfile
-
- equals(Object) - Method in class burlap.mdp.core.action.SimpleAction
-
- equals(Object) - Method in class burlap.mdp.core.oo.propositional.GroundedProp
-
- equals(Object) - Method in class burlap.mdp.core.oo.propositional.PropositionalFunction
-
- equals(Object) - Method in class burlap.mdp.core.oo.state.OOVariableKey
-
- equals(Object) - Method in class burlap.mdp.core.state.NullState
-
- equals(Object) - Method in class burlap.mdp.singleagent.oo.ObjectParameterizedActionType.SAObjectParameterizedAction
-
- equals(Object) - Method in class burlap.mdp.singleagent.pomdp.beliefstate.TabularBeliefState
-
- equals(Object) - Method in class burlap.mdp.stochasticgames.agent.SGAgentType
-
- equals(Object) - Method in class burlap.mdp.stochasticgames.JointAction
-
- equals(Object) - Method in class burlap.statehashing.simple.IDSimpleHashableState
-
- equals(Object) - Method in class burlap.statehashing.simple.IISimpleHashableState
-
- equals(Object) - Method in class burlap.statehashing.WrappedHashableState
-
- EquilibriumPlayingSGAgent - Class in burlap.behavior.stochasticgames.agents.twoplayer.singlestage.equilibriumplayer
-
This agent plays an equilibrium solution for two player games based on the immediate joint rewards received for the given state, as if
it is a single stage game.
- EquilibriumPlayingSGAgent() - Constructor for class burlap.behavior.stochasticgames.agents.twoplayer.singlestage.equilibriumplayer.EquilibriumPlayingSGAgent
-
Initializes with the
MaxMax
solution concept.
- EquilibriumPlayingSGAgent(BimatrixEquilibriumSolver) - Constructor for class burlap.behavior.stochasticgames.agents.twoplayer.singlestage.equilibriumplayer.EquilibriumPlayingSGAgent
-
Initializes with strategies formed usign the solution concept generated by the given solver.
- EquilibriumPlayingSGAgent.BimatrixTuple - Class in burlap.behavior.stochasticgames.agents.twoplayer.singlestage.equilibriumplayer
-
A Bimatrix tuple.
- eStepCounter - Variable in class burlap.behavior.singleagent.learning.tdmethods.QLearning
-
A counter for counting the number of steps in an episode that have been taken thus far
- eStepCounter - Variable in class burlap.behavior.singleagent.learning.tdmethods.vfa.GradientDescentSarsaLam
-
A counter for counting the number of steps in an episode that have been taken thus far
- estimateFeatureExpectation(Episode, DenseStateFeatures, Double) - Static method in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearning
-
Calculates the Feature Expectations given one demonstration, a feature mapping and a discount factor gamma
- estimateFeatureExpectation(List<Episode>, DenseStateFeatures, Double) - Static method in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearning
-
Calculates the Feature Expectations given a list of demonstrations, a feature mapping and a
discount factor gamma
- estimateQs() - Method in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableSparseSampling.DiffStateNode
-
- estimateQs() - Method in class burlap.behavior.singleagent.planning.stochastic.sparsesampling.SparseSampling.StateNode
-
Estimates and returns the Q-values for this node.
- estimateV() - Method in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableSparseSampling.DiffStateNode
-
- estimateV() - Method in class burlap.behavior.singleagent.planning.stochastic.sparsesampling.SparseSampling.StateNode
-
Returns the estimated Q-value if this node is closed, or estimates it and closes it otherwise.
- EuclideanDistance - Class in burlap.behavior.functionapproximation.dense.rbf.metrics
-
A distance metric; returns sqrt( sum_i (x_i - y_i)^2 )
- EuclideanDistance() - Constructor for class burlap.behavior.functionapproximation.dense.rbf.metrics.EuclideanDistance
-
- evaluate(State, Action) - Method in class burlap.behavior.functionapproximation.dense.DenseLinearVFA
-
- evaluate(State) - Method in class burlap.behavior.functionapproximation.dense.DenseLinearVFA
-
- evaluate(State, Action) - Method in class burlap.behavior.functionapproximation.dense.DenseStateActionLinearVFA
-
- evaluate(State, Action) - Method in interface burlap.behavior.functionapproximation.ParametricFunction.ParametricStateActionFunction
-
Sets the input of this function to the given
State
and
Action
and returns the value of it.
- evaluate(State) - Method in interface burlap.behavior.functionapproximation.ParametricFunction.ParametricStateFunction
-
Sets the input of this function to the given
State
and returns
the value of it.
- evaluate(State, Action) - Method in class burlap.behavior.functionapproximation.sparse.LinearVFA
-
- evaluate(State) - Method in class burlap.behavior.functionapproximation.sparse.LinearVFA
-
- evaluateEpisode(Episode) - Method in class burlap.testing.TestPlanning
-
- evaluateEpisode(Episode, Boolean) - Method in class burlap.testing.TestPlanning
-
- evaluatePolicy(EnumerablePolicy, State) - Method in class burlap.behavior.singleagent.planning.stochastic.policyiteration.PolicyEvaluation
-
Computes the value function for the given policy after finding all reachable states from seed state s
- evaluatePolicy(EnumerablePolicy) - Method in class burlap.behavior.singleagent.planning.stochastic.policyiteration.PolicyEvaluation
-
Computes the value function for the given policy over the states that have been discovered
- evaluatePolicy() - Method in class burlap.behavior.singleagent.planning.stochastic.policyiteration.PolicyIteration
-
Computes the value function under following the current evaluative policy.
- evaluativePolicy - Variable in class burlap.behavior.singleagent.planning.stochastic.policyiteration.PolicyIteration
-
The current policy to be evaluated
- exactQEstimate(Action, DifferentiableSparseSampling.QAndQGradient) - Method in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableSparseSampling.DiffStateNode
-
- exactQValue(Action) - Method in class burlap.behavior.singleagent.planning.stochastic.sparsesampling.SparseSampling.StateNode
-
Computes the exact Q-value using full Bellman update with the actual transition dynamics.
- executeAction(Action) - Method in class burlap.behavior.singleagent.interfaces.rlglue.RLGlueAgent
-
- executeAction(Action) - Method in class burlap.behavior.stochasticgames.agents.interfacing.singleagent.LearningAgentToSGAgentInterface
-
- executeAction(Action) - Method in interface burlap.mdp.singleagent.environment.Environment
-
Executes the specified action in this environment
- executeAction(Action) - Method in class burlap.mdp.singleagent.environment.extensions.EnvironmentServer
-
- executeAction(Action) - Method in class burlap.mdp.singleagent.environment.SimulatedEnvironment
-
- executeAction(Action) - Method in class burlap.mdp.singleagent.pomdp.SimulatedPOEnvironment
-
- executeAction(Action) - Method in class burlap.shell.visual.VisualExplorer
-
Executes the provided
Action
in the explorer's environment and records
the result if episodes are being recorded.
- ExecuteActionCommand - Class in burlap.shell.command.env
-
- ExecuteActionCommand(Domain) - Constructor for class burlap.shell.command.env.ExecuteActionCommand
-
- executeCommand(String) - Method in class burlap.shell.BurlapShell
-
- executeJointAction(JointAction) - Method in class burlap.mdp.stochasticgames.world.World
-
Manually attempts to execute a joint action in the current world state, if a game is currently not running.
- exit(int, int) - Static method in class burlap.domain.singleagent.blockdude.state.BlockDudeCell
-
- exit - Variable in class burlap.domain.singleagent.blockdude.state.BlockDudeState
-
- ExitPainter(int, int) - Constructor for class burlap.domain.singleagent.blockdude.BlockDudeVisualizer.ExitPainter
-
Initializes.
- expandStateActionWeights(int) - Method in class burlap.behavior.functionapproximation.dense.DenseLinearVFA
-
Expands the state-action function weight vector by a fixed sized and initializes their value
to the default weight value set for this object.
- expectedDepth - Variable in class burlap.behavior.singleagent.planning.deterministic.informed.astar.DynamicWeightedAStar
-
The expected depth required for a plan
- expectedGap - Variable in class burlap.behavior.singleagent.planning.stochastic.rtdp.BoundedRTDP.StateSelectionAndExpectedGap
-
The expected margin/gap of the value function from the source transition
- expectedPayoffs(double[][], double[][], double[], double[]) - Static method in class burlap.behavior.stochasticgames.solvers.GeneralBimatrixSolverTools
-
Computes the expected payoff for each player in a bimatrix game according to their strategies.
- expectedPayoffs(double[][], double[][], double[][]) - Static method in class burlap.behavior.stochasticgames.solvers.GeneralBimatrixSolverTools
-
- ExperimentalEnvironment - Interface in burlap.behavior.singleagent.auxiliary.performance
-
An interface to be used in conjunction with
Environment
implementations
that can accept a message informing the environment that a new experiment for a
LearningAgent
has started.
- expertEpisodes - Variable in class burlap.behavior.singleagent.learnfromdemo.IRLRequest
-
The input trajectories/episodes that are to be modeled.
- explorationBias - Variable in class burlap.behavior.singleagent.planning.stochastic.montecarlo.uct.UCT
-
- explorationQBoost(int, int) - Method in class burlap.behavior.singleagent.planning.stochastic.montecarlo.uct.UCT
-
Returns the extra value added to the average sample Q-value that is sued to produce the upper confidence Q-value.
- ExponentialDecayLR - Class in burlap.behavior.learningrate
-
This class provides a learning rate that decays exponentially with time according to r^t, where r is in [0,1] and t is the time step, from an initial
learning rate.
- ExponentialDecayLR(double, double) - Constructor for class burlap.behavior.learningrate.ExponentialDecayLR
-
Initializes with an initial learning rate and decay rate for a state independent learning rate.
- ExponentialDecayLR(double, double, double) - Constructor for class burlap.behavior.learningrate.ExponentialDecayLR
-
Initializes with an initial learning rate and decay rate for a state independent learning rate that will decay to a value no smaller than minimumLearningRate
- ExponentialDecayLR(double, double, HashableStateFactory, boolean) - Constructor for class burlap.behavior.learningrate.ExponentialDecayLR
-
Initializes with an initial learning rate and decay rate for a state or state-action (or state feature-action) dependent learning rate.
- ExponentialDecayLR(double, double, double, HashableStateFactory, boolean) - Constructor for class burlap.behavior.learningrate.ExponentialDecayLR
-
Initializes with an initial learning rate and decay rate for a state or state-action (or state feature-action) dependent learning rate that will decay to a value no smaller than minimumLearningRate
If this learning rate function is to be used for state state features, rather than states,
then the hashing factory can be null;
- ExponentialDecayLR.MutableDouble - Class in burlap.behavior.learningrate
-
A class for storing a mutable double value object
- ExponentialDecayLR.StateWiseLearningRate - Class in burlap.behavior.learningrate
-
A class for storing a learning rate for a state, or a learning rate for each action for a given state