Skip navigation links
A B C D E F G H I J K L M N O P Q R S T U V W X Y Z 

D

dataset - Variable in class burlap.behavior.singleagent.learning.lspi.LSPI
The SARS dataset on which LSPI is performed
dataset - Variable in class burlap.behavior.singleagent.learning.lspi.SARSData
The underlying list of SARSData.SARS tuples.
datasets - Variable in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentPerformancePlotter.DatasetsAndTrials
The series datasets for this agent
DatasetsAndTrials(String) - Constructor for class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentPerformancePlotter.DatasetsAndTrials
Initializes for an agent with the given name.
DDPlannerPolicy - Class in burlap.behavior.singleagent.planning.deterministic
This is a dynamic deterministic valueFunction policy, which means if the source deterministic valueFunction has not already computed and cached the plan for a query state, then this policy will first compute a plan using the valueFunction and then return the answer
DDPlannerPolicy() - Constructor for class burlap.behavior.singleagent.planning.deterministic.DDPlannerPolicy
 
DDPlannerPolicy(DeterministicPlanner) - Constructor for class burlap.behavior.singleagent.planning.deterministic.DDPlannerPolicy
Initializes with the deterministic valueFunction
DEBUG_CODE_RF_WEIGHTS - Static variable in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearning
 
DEBUG_CODE_SCORE - Static variable in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearning
 
debugCode - Variable in class burlap.behavior.singleagent.auxiliary.performance.LearningAlgorithmExperimenter
The debug code used for debug printing.
debugCode - Variable in class burlap.behavior.singleagent.interfaces.rlglue.RLGlueAgent
Debug code used for printing debug information.
debugCode - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.MLIRL
The debug code used for printing information to the terminal.
debugCode - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.MultipleIntentionsMLIRL
The debug code used for printing information to the terminal.
debugCode - Variable in class burlap.behavior.singleagent.MDPSolver
The debug code use for calls to DPrint
debugCode - Variable in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentExperimenter
The debug code used for debug printing.
debugCode - Variable in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentPerformancePlotter
The debug code used for debug printing.
debugCode - Variable in class burlap.behavior.stochasticgames.madynamicprogramming.dpplanners.MAValueIteration
The debug code used for printing VI progress.
DebugFlags - Class in burlap.debugtools
A data structure for specifying debug flags that can be accessed and modified from any class
debugID - Static variable in class burlap.behavior.singleagent.auxiliary.StateReachability
The debugID used for making calls to DPrint.
debugId - Variable in class burlap.mdp.stochasticgames.tournament.Tournament
 
debugId - Variable in class burlap.mdp.stochasticgames.world.World
 
decayConstantShift - Variable in class burlap.behavior.learningrate.SoftTimeInverseDecayLR
The division scale offset
decayRate - Variable in class burlap.behavior.learningrate.ExponentialDecayLR
The exponential base by which the learning rate is decayed
deepCopyActionNameMapArray(SingleStageNormalFormGame.ActionNameMap[]) - Static method in class burlap.domain.stochasticgames.normalform.SingleStageNormalFormGame.ActionNameMap
 
DeepCopyState - Annotation Type in burlap.mdp.core.state.annotations
A marker for State implementations that indicates that their copy operation is deep.
DeepOOState - Class in burlap.mdp.core.oo.state.generic
An alternative implementation of GenericOOState in which the DeepOOState.copy() operations performs a deep copy (DeepCopyState) of all ObjectInstance objects, thereby allows safe modification of any of its ObjectInstance objects without using the GenericOOState.touch(String) method.
DeepOOState() - Constructor for class burlap.mdp.core.oo.state.generic.DeepOOState
 
DeepOOState(OOState) - Constructor for class burlap.mdp.core.oo.state.generic.DeepOOState
 
DeepOOState(ObjectInstance...) - Constructor for class burlap.mdp.core.oo.state.generic.DeepOOState
 
deepTouchLocations() - Method in class burlap.domain.singleagent.gridworld.state.GridWorldState
 
deepTouchObstacles() - Method in class burlap.domain.singleagent.lunarlander.state.LLState
 
deepTouchPlatforms() - Method in class burlap.domain.singleagent.frostbite.state.FrostbiteState
 
DEFAULT_EPSILON - Static variable in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearningRequest
 
DEFAULT_MAXITERATIONS - Static variable in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearningRequest
 
DEFAULT_POLICYCOUNT - Static variable in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearningRequest
 
DEFAULT_USEMAXMARGIN - Static variable in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearningRequest
 
defaultCost(int, JointAction) - Method in class burlap.domain.stochasticgames.gridgame.GridGame.GGJointRewardFunction
Returns a default cost for an agent assuming the agent didn't transition to a goal state.
defaultMultiple - Variable in class burlap.statehashing.discretized.DiscConfig
The default multiple to use for any continuous attributes that have not been specifically set.
defaultMultiple - Variable in class burlap.statehashing.maskeddiscretized.DiscMaskedConfig
The default multiple to use for any continuous attributes that have not been specifically set.
defaultQ - Variable in class burlap.behavior.stochasticgames.agents.naiveq.SGNaiveQFactory
The default Q-value to which Q-values will be initialized
defaultReward - Variable in class burlap.domain.singleagent.frostbite.FrostbiteRF
 
defaultReward - Variable in class burlap.domain.singleagent.lunarlander.LunarLanderRF
The default reward received for moving through the air
defaultReward - Variable in class burlap.mdp.singleagent.common.GoalBasedRF
 
defaultToLowerValueAfterPlanning - Variable in class burlap.behavior.singleagent.planning.stochastic.rtdp.BoundedRTDP
Sets what the DynamicProgramming valueFunction reference points to (the lower bound or upperbound) once a planning rollout is complete.
defaultWeight - Variable in class burlap.behavior.functionapproximation.dense.DenseLinearVFA
A default weight value for the functions weights.
defaultWeight - Variable in class burlap.behavior.functionapproximation.dense.DenseStateActionLinearVFA
A default weight value for the functions weights.
defaultWeight - Variable in class burlap.behavior.functionapproximation.sparse.LinearVFA
A default weight for the functions
definedFor(State) - Method in class burlap.behavior.policy.BoltzmannQPolicy
 
definedFor(State) - Method in class burlap.behavior.policy.CachedPolicy
 
definedFor(State) - Method in class burlap.behavior.policy.EpsilonGreedy
 
definedFor(State) - Method in class burlap.behavior.policy.GreedyDeterministicQPolicy
 
definedFor(State) - Method in class burlap.behavior.policy.GreedyQPolicy
 
definedFor(State) - Method in interface burlap.behavior.policy.Policy
Specifies whether this policy is defined for the input state.
definedFor(State) - Method in class burlap.behavior.policy.RandomPolicy
 
definedFor(State) - Method in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearning.StationaryRandomDistributionPolicy
 
definedFor(State) - Method in class burlap.behavior.singleagent.learning.actorcritic.actor.BoltzmannActor
 
definedFor(State) - Method in class burlap.behavior.singleagent.learning.modellearning.rmax.UnmodeledFavoredPolicy
 
definedFor(State) - Method in class burlap.behavior.singleagent.planning.deterministic.DDPlannerPolicy
 
definedFor(State) - Method in class burlap.behavior.singleagent.planning.deterministic.SDPlannerPolicy
 
definedFor(State) - Method in class burlap.behavior.singleagent.planning.stochastic.montecarlo.uct.UCTTreeWalkPolicy
 
definedFor(State) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.policies.ECorrelatedQJointPolicy
 
definedFor(State) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyJointPolicy
 
definedFor(State) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyMaxWellfare
 
definedFor(State) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.policies.EMinMaxPolicy
 
definedFor(State) - Method in class burlap.behavior.stochasticgames.PolicyFromJointPolicy
 
delay - Variable in class burlap.behavior.singleagent.auxiliary.performance.PerformancePlotter
the delay in milliseconds between which the charts are updated automatically
delay - Variable in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentPerformancePlotter
the delay in milliseconds between which the charts are updated automatically
delegate - Variable in class burlap.mdp.singleagent.environment.extensions.EnvironmentServer
the Environment delegate that handles all primary Environment functionality.
DelegatedModel - Class in burlap.mdp.singleagent.model
An implementation of FullModel that will delegate transition estimates for different actions to different SampleModel or FullModel implementations.
DelegatedModel(SampleModel) - Constructor for class burlap.mdp.singleagent.model.DelegatedModel
 
delta - Static variable in class burlap.testing.TestPlanning
 
DenseBeliefVector - Interface in burlap.mdp.singleagent.pomdp.beliefstate
An interface to be used in conjunction with BeliefState instances for belief states that can generate a dense belief vector representation.
DenseCrossProductFeatures - Class in burlap.behavior.functionapproximation.dense
Class that generates state-action features as cross product of underlying state-features with the action set.
DenseCrossProductFeatures(DenseStateFeatures, int) - Constructor for class burlap.behavior.functionapproximation.dense.DenseCrossProductFeatures
 
DenseCrossProductFeatures(DenseStateFeatures, int, Map<Action, Integer>) - Constructor for class burlap.behavior.functionapproximation.dense.DenseCrossProductFeatures
 
DenseLinearVFA - Class in burlap.behavior.functionapproximation.dense
This class can be used to perform linear value function approximation, either for a states or state-actions (Q-values).
DenseLinearVFA(DenseStateFeatures, double) - Constructor for class burlap.behavior.functionapproximation.dense.DenseLinearVFA
Initializes.
DenseStateActionFeatures - Interface in burlap.behavior.functionapproximation.dense
 
DenseStateActionLinearVFA - Class in burlap.behavior.functionapproximation.dense
 
DenseStateActionLinearVFA(DenseStateActionFeatures, double) - Constructor for class burlap.behavior.functionapproximation.dense.DenseStateActionLinearVFA
 
DenseStateActionLinearVFA(DenseStateActionFeatures, double[], double) - Constructor for class burlap.behavior.functionapproximation.dense.DenseStateActionLinearVFA
 
DenseStateFeatures - Interface in burlap.behavior.functionapproximation.dense
Many functions approximation techniques require a fixed feature vector to work and in many cases, using abstract features from the state attributes is useful.
depth - Variable in class burlap.behavior.singleagent.planning.stochastic.montecarlo.uct.UCTStateNode
The depth the UCT tree
depthMap - Variable in class burlap.behavior.singleagent.planning.deterministic.informed.astar.DynamicWeightedAStar
Data structure for storing the depth of explored states
DeterministicPlanner - Class in burlap.behavior.singleagent.planning.deterministic
This class extends the OOMDPlanner to provide the interface and common mechanisms for classic deterministic forward search planners.
DeterministicPlanner() - Constructor for class burlap.behavior.singleagent.planning.deterministic.DeterministicPlanner
 
DeterministicPlanner.PlanningFailedException - Exception in burlap.behavior.singleagent.planning.deterministic
Exception class for indicating that a solution failed to be found by the planning algorithm.
deterministicPlannerInit(SADomain, StateConditionTest, HashableStateFactory) - Method in class burlap.behavior.singleagent.planning.deterministic.DeterministicPlanner
Initializes the valueFunction.
deterministicPolicyDistribution(Policy, State) - Static method in class burlap.behavior.policy.PolicyUtils
A helper method for defining deterministic policies.
deterministicTransition(SampleModel, State, Action) - Static method in class burlap.mdp.singleagent.model.FullModel.Helper
Method to easily implement the FullModel.transitions(State, Action) method for deterministic domains.
deterministicTransition(SampleStateModel, State, Action) - Static method in class burlap.mdp.singleagent.model.statemodel.FullStateModel.Helper
Method to easily implement the FullStateModel.stateTransitions(State, Action) method for deterministic domains.
deterministicTransition(JointModel, State, JointAction) - Static method in class burlap.mdp.stochasticgames.model.FullJointModel.Helper
A helper method for deterministic transition dynamics.
DFS - Class in burlap.behavior.singleagent.planning.deterministic.uninformed.dfs
Implements depth-first search.
DFS(SADomain, StateConditionTest, HashableStateFactory) - Constructor for class burlap.behavior.singleagent.planning.deterministic.uninformed.dfs.DFS
Basic constructor for standard DFS without a depth limit
DFS(SADomain, StateConditionTest, HashableStateFactory, int) - Constructor for class burlap.behavior.singleagent.planning.deterministic.uninformed.dfs.DFS
Basic constructor for standard DFS with a depth limit
DFS(SADomain, StateConditionTest, HashableStateFactory, int, boolean) - Constructor for class burlap.behavior.singleagent.planning.deterministic.uninformed.dfs.DFS
Constructor of DFS with specification of depth limit and whether to maintain a closed list that affects exploration.
DFS(SADomain, StateConditionTest, HashableStateFactory, int, boolean, boolean) - Constructor for class burlap.behavior.singleagent.planning.deterministic.uninformed.dfs.DFS
Constructor of DFS with specification of depth limit, whether to maintain a closed list that affects exploration, and whether paths generated by options should be explored first.
dfs(SearchNode, int, Set<HashableState>) - Method in class burlap.behavior.singleagent.planning.deterministic.uninformed.dfs.DFS
Runs DFS from a given search node, keeping track of its current depth.
dfs(SearchNode, int, Set<HashableState>) - Method in class burlap.behavior.singleagent.planning.deterministic.uninformed.dfs.LimitedMemoryDFS
Runs DFS from a given search node, keeping track of its current depth.
DFSInit(SADomain, StateConditionTest, HashableStateFactory, int, boolean, boolean) - Method in class burlap.behavior.singleagent.planning.deterministic.uninformed.dfs.DFS
Constructor of DFS with specification of depth limit, whether to maintain a closed list that affects exploration, and whether paths generated by options should be explored first.
dheight - Variable in class burlap.domain.singleagent.gridworld.GridWorldVisualizer.CellPainter
 
dheight - Variable in class burlap.domain.singleagent.gridworld.GridWorldVisualizer.LocationPainter
 
dheight - Variable in class burlap.domain.singleagent.gridworld.GridWorldVisualizer.MapPainter
 
DifferentiableDP - Class in burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners
A class for performing dynamic programming with a differentiable value backup operator.
DifferentiableDP() - Constructor for class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableDP
 
DifferentiableDPOperator - Interface in burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.dpoperator
A DPOperator that is differentiable.
DifferentiableQFunction - Interface in burlap.behavior.singleagent.learnfromdemo.mlirl.support
An interface for a valueFunction that can produce Q-value gradients.
DifferentiableRF - Interface in burlap.behavior.singleagent.learnfromdemo.mlirl.support
An interface for defining differentiable reward functions.
DifferentiableSoftmaxOperator - Class in burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.dpoperator
Provides the gradient for the SoftmaxOperator
DifferentiableSoftmaxOperator() - Constructor for class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.dpoperator.DifferentiableSoftmaxOperator
 
DifferentiableSoftmaxOperator(double) - Constructor for class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.dpoperator.DifferentiableSoftmaxOperator
 
DifferentiableSparseSampling - Class in burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners
A Differentiable finite horizon valueFunction that can also use sparse sampling over the transition dynamics when the transition function is very large or infinite.
DifferentiableSparseSampling(SADomain, DifferentiableRF, double, HashableStateFactory, int, int, double) - Constructor for class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableSparseSampling
Initializes.
DifferentiableSparseSampling.DiffStateNode - Class in burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners
A class for value differentiable state nodes.
DifferentiableSparseSampling.QAndQGradient - Class in burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners
A tuple for storing Q-values and their gradients.
DifferentiableSparseSampling.VAndVGradient - Class in burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners
A tuple for storing a state value and its gradient.
DifferentiableStateActionValue - Interface in burlap.behavior.functionapproximation
An extension of ParametricFunction.ParametricStateActionFunction that is differentiable.
DifferentiableStateValue - Interface in burlap.behavior.functionapproximation
An extension of ParametricFunction.ParametricStateFunction that that is differentiable.
DifferentiableValueFunction - Interface in burlap.behavior.singleagent.learnfromdemo.mlirl.support
 
DifferentiableVI - Class in burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners
Performs Differentiable Value Iteration using the Boltzmann backup operator and a DifferentiableRF.
DifferentiableVI(SADomain, DifferentiableRF, double, double, HashableStateFactory, double, int) - Constructor for class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableVI
Initializes the valueFunction.
DifferentiableVIFactory(HashableStateFactory) - Constructor for class burlap.behavior.singleagent.learnfromdemo.mlirl.support.QGradientPlannerFactory.DifferentiableVIFactory
Initializes the factory with the given HashableStateFactory.
DifferentiableVIFactory(HashableStateFactory, TerminalFunction, double, int) - Constructor for class burlap.behavior.singleagent.learnfromdemo.mlirl.support.QGradientPlannerFactory.DifferentiableVIFactory
Initializes.
DifferentiableVInit - Interface in burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.diffvinit
An interface for value function initialization that is differentiable with respect to some parameters.
DiffStateNode(HashableState, int) - Constructor for class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableSparseSampling.DiffStateNode
 
DiffVFRF - Class in burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.diffvinit
A differentiable reward function wrapper for use with MLIRL when the reward function is known, but the value function initialization for leaf nodes is to be learned.
DiffVFRF(RewardFunction, DifferentiableVInit) - Constructor for class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.diffvinit.DiffVFRF
 
diffVInit - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.diffvinit.DiffVFRF
 
dim - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.commonrfs.LinearStateActionDifferentiableRF
The dimension of this reward function
dim - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.commonrfs.LinearStateDifferentiableRF
The dimension of this reward function
dim - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.diffvinit.DiffVFRF
 
dim - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.diffvinit.LinearDiffRFVInit
 
dim - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.diffvinit.LinearStateDiffVF
 
dimensionMask - Variable in class burlap.behavior.functionapproximation.sparse.tilecoding.Tiling
The dimensions on which this tiling are dependent
dir - Variable in class burlap.domain.singleagent.blockdude.state.BlockDudeAgent
 
dir(String) - Method in class burlap.domain.singleagent.mountaincar.MountainCar.MCModel
 
directEpisodes - Variable in class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
 
directGameEpisodes - Variable in class burlap.behavior.stochasticgames.auxiliary.GameSequenceVisualizer
 
direction - Variable in class burlap.behavior.singleagent.auxiliary.valuefunctionvis.common.ArrowActionGlyph
The direction of the arrow.0: north; 1: south; 2: east; 3:west
DiscConfig - Class in burlap.statehashing.discretized
 
DiscConfig() - Constructor for class burlap.statehashing.discretized.DiscConfig
 
DiscConfig(double) - Constructor for class burlap.statehashing.discretized.DiscConfig
 
DiscConfig(Map<Object, Double>, double) - Constructor for class burlap.statehashing.discretized.DiscConfig
 
DiscMaskedConfig - Class in burlap.statehashing.maskeddiscretized
 
DiscMaskedConfig() - Constructor for class burlap.statehashing.maskeddiscretized.DiscMaskedConfig
 
DiscMaskedConfig(double) - Constructor for class burlap.statehashing.maskeddiscretized.DiscMaskedConfig
 
DiscMaskedConfig(Set<Object>, Set<String>, Map<Object, Double>, double) - Constructor for class burlap.statehashing.maskeddiscretized.DiscMaskedConfig
 
discount - Variable in class burlap.behavior.singleagent.interfaces.rlglue.RLGlueAgent
The RLGlue specified discount factor
discount - Variable in class burlap.behavior.singleagent.options.EnvironmentOptionOutcome
The discount factor to apply to the value of time steps immediately following the application of an Option.
discount - Variable in class burlap.behavior.singleagent.options.model.BFSMarkovOptionModel
 
discount - Variable in class burlap.behavior.singleagent.shaping.potential.PotentialShapedRF
The discount factor the MDP (required for this to shaping to preserve policy optimality)
discount - Variable in class burlap.behavior.stochasticgames.agents.madp.MADPPlannerFactory.MAVIPlannerFactory
The discount factor in [0, 1]
discount - Variable in class burlap.behavior.stochasticgames.agents.maql.MAQLFactory
 
discount - Variable in class burlap.behavior.stochasticgames.agents.maql.MultiAgentQLearning
The discount factor
discount - Variable in class burlap.behavior.stochasticgames.agents.naiveq.history.SGQWActionHistoryFactory
The discount rate the Q-learning algorithm will use
discount - Variable in class burlap.behavior.stochasticgames.agents.naiveq.SGNaiveQFactory
The discount rate the Q-learning algorithm will use
discount - Variable in class burlap.behavior.stochasticgames.agents.naiveq.SGNaiveQLAgent
The discount factor
discount - Variable in class burlap.behavior.stochasticgames.madynamicprogramming.MADynamicProgramming
The discount factor in [0, 1]
discount - Variable in class burlap.domain.singleagent.rlglue.RLGlueEnvironment
The discount factor of the task
discountedReturn(double) - Method in class burlap.behavior.singleagent.Episode
Will return the discounted return received from the first state in the episode to the last state in the episode.
DiscreteObservationFunction - Interface in burlap.mdp.singleagent.pomdp.observations
Defines additional methods for an ObservationFunction for the case when the set of observations are discrete and able to be enumerated.
DiscretizingHashableStateFactory - Class in burlap.statehashing.discretized
A factory for producing HashableState objects that computes hash codes and test for state equality after discretizing any real values (Float or Double).
DiscretizingHashableStateFactory(double) - Constructor for class burlap.statehashing.discretized.DiscretizingHashableStateFactory
Initializes with object identifier independence and no hash code caching.
DiscretizingHashableStateFactory(boolean, double) - Constructor for class burlap.statehashing.discretized.DiscretizingHashableStateFactory
Initializes with non hash code caching
DiscretizingMaskedHashableStateFactory - Class in burlap.statehashing.maskeddiscretized
A class for producing HashableState objects that computes hash codes and tests for State equality by discretizing real-valued attributes and by masking (ignoring) either state variables and/or OOState clasees.
DiscretizingMaskedHashableStateFactory(double) - Constructor for class burlap.statehashing.maskeddiscretized.DiscretizingMaskedHashableStateFactory
Initializes with object identifier independence, no hash code caching and object class or attribute masks.
DiscretizingMaskedHashableStateFactory(boolean, double) - Constructor for class burlap.statehashing.maskeddiscretized.DiscretizingMaskedHashableStateFactory
Initializes with non hash code caching and object class or attribute masks
DiscretizingMaskedHashableStateFactory(boolean, double, boolean, String...) - Constructor for class burlap.statehashing.maskeddiscretized.DiscretizingMaskedHashableStateFactory
Initializes with a specified attribute or object class mask.
displayPlots - Variable in class burlap.behavior.singleagent.auxiliary.performance.LearningAlgorithmExperimenter
Whether the performance should be visually plotted (by default they will)
displayPlots - Variable in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentExperimenter
Whether the performance should be visually plotted (by default they will)
distance(double[], double[]) - Method in interface burlap.behavior.functionapproximation.dense.rbf.DistanceMetric
Returns the distance between state s0 and state s1.
distance(double[], double[]) - Method in class burlap.behavior.functionapproximation.dense.rbf.metrics.EuclideanDistance
 
DistanceMetric - Interface in burlap.behavior.functionapproximation.dense.rbf
An interface for defining the distance between two states that are represented with double arrays.
domain - Variable in class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
 
domain - Variable in class burlap.behavior.singleagent.auxiliary.StateEnumerator
The domain whose states will be enumerated
domain - Variable in class burlap.behavior.singleagent.interfaces.rlglue.RLGlueAgent
The BURLAP Domain specifying the RLGlue problem representation and action space.
domain - Variable in class burlap.behavior.singleagent.learnfromdemo.IRLRequest
The domain in which IRL is to be performed
domain - Variable in class burlap.behavior.singleagent.learnfromdemo.RewardValueProjection
 
domain - Variable in class burlap.behavior.singleagent.learning.actorcritic.actor.BoltzmannActor
The domain in which this agent will act
domain - Variable in class burlap.behavior.singleagent.MDPSolver
The domain to solve
domain - Variable in class burlap.behavior.stochasticgames.agents.madp.MADPPlanAgentFactory
 
domain - Variable in class burlap.behavior.stochasticgames.agents.madp.MADPPlannerFactory.MAVIPlannerFactory
The domain in which planning is to be performed
domain - Variable in class burlap.behavior.stochasticgames.agents.maql.MAQLFactory
 
domain - Variable in class burlap.behavior.stochasticgames.agents.naiveq.history.SGQWActionHistoryFactory
The stochastic games domain in which the agent will act
domain - Variable in class burlap.behavior.stochasticgames.agents.naiveq.SGNaiveQFactory
The stochastic games domain in which the agent will act
domain - Variable in class burlap.behavior.stochasticgames.agents.SetStrategySGAgent.SetStrategyAgentFactory
The domain in which the agent will play
domain - Variable in class burlap.behavior.stochasticgames.agents.twoplayer.repeatedsinglestage.GrimTrigger.GrimTriggerAgentFactory
The domain in which the agent will play
domain - Variable in class burlap.behavior.stochasticgames.agents.twoplayer.repeatedsinglestage.TitForTat.TitForTatAgentFactory
The domain in which the agent will play
domain - Variable in class burlap.behavior.stochasticgames.auxiliary.GameSequenceVisualizer
 
domain - Variable in class burlap.behavior.stochasticgames.madynamicprogramming.MADynamicProgramming
The domain in which planning is to be performed
domain - Variable in class burlap.domain.singleagent.rlglue.RLGlueEnvironment
The BURLAP domain
Domain - Interface in burlap.mdp.core
This is a marker interface for a problem domain.
domain(Object) - Method in interface burlap.mdp.core.state.vardomain.StateDomain
Returns the numeric domain of the variable for the given key.
domain - Variable in class burlap.mdp.singleagent.common.VisualActionObserver
The domain this visualizer is rendering
domain - Variable in class burlap.mdp.singleagent.pomdp.beliefstate.TabularBeliefState
The POMDP domain with which this belief state is associated.
domain - Variable in class burlap.mdp.singleagent.pomdp.beliefstate.TabularBeliefUpdate
 
domain - Variable in class burlap.mdp.stochasticgames.agent.SGAgentBase
 
domain - Variable in class burlap.mdp.stochasticgames.common.VisualWorldObserver
The domain this visualizer is rendering
domain - Variable in class burlap.mdp.stochasticgames.tournament.common.ConstantWorldGenerator
 
domain - Variable in class burlap.mdp.stochasticgames.world.World
 
domain - Variable in class burlap.shell.BurlapShell
 
domain - Variable in class burlap.shell.command.env.AddStateObjectCommand
 
domain - Variable in class burlap.shell.command.env.ExecuteActionCommand
 
domain - Variable in class burlap.shell.command.world.AddStateObjectSGCommand
 
domain - Variable in class burlap.shell.visual.SGVisualExplorer
 
domain - Variable in class burlap.shell.visual.VisualExplorer
 
DomainGenerator - Interface in burlap.mdp.auxiliary
This class provides a simple interface for constructing domains, but it is not required to create domains.
domains - Variable in class burlap.behavior.functionapproximation.dense.NormalizedVariableFeatures
 
domainSet - Variable in class burlap.behavior.singleagent.interfaces.rlglue.RLGlueAgent
A variable for synchronized checking if the domain has been set.
door - Variable in class burlap.domain.singleagent.pomdp.tiger.TigerState
 
DOOR_RESET - Static variable in class burlap.domain.singleagent.pomdp.tiger.TigerDomain
The observation value for when reaching a new pair of doors (occurs after opening a door)
dot(double[], double[]) - Static method in class burlap.behavior.stochasticgames.solvers.GeneralBimatrixSolverTools
Returns the dot product of two vectors
doubleEpislon - Static variable in class burlap.behavior.stochasticgames.agents.twoplayer.singlestage.equilibriumplayer.BimatrixEquilibriumSolver
The epislon difference used to test for double equality.
doubleEquality(double, double) - Static method in class burlap.behavior.stochasticgames.agents.twoplayer.singlestage.equilibriumplayer.BimatrixEquilibriumSolver
Returns true if |a - b| < BimatrixEquilibriumSolver.doubleEpislon; false otherwise.
dp - Variable in class burlap.behavior.singleagent.planning.deterministic.DDPlannerPolicy
 
dp - Variable in class burlap.behavior.singleagent.planning.deterministic.SDPlannerPolicy
 
DPOperator - Interface in burlap.behavior.singleagent.planning.stochastic.dpoperator
Defines a function for applying a dynamic programming operator (e.g., reducing the Q-values into a state value).
DPPInit(SADomain, double, HashableStateFactory) - Method in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableDP
 
DPPInit(SADomain, double, HashableStateFactory) - Method in class burlap.behavior.singleagent.planning.stochastic.DynamicProgramming
Common init method for DynamicProgramming instances.
DPrint - Class in burlap.debugtools
A class for managing debug print statements.
dwidth - Variable in class burlap.domain.singleagent.gridworld.GridWorldVisualizer.CellPainter
 
dwidth - Variable in class burlap.domain.singleagent.gridworld.GridWorldVisualizer.LocationPainter
 
dwidth - Variable in class burlap.domain.singleagent.gridworld.GridWorldVisualizer.MapPainter
 
DynamicProgramming - Class in burlap.behavior.singleagent.planning.stochastic
A class for performing dynamic programming operations: updating the value function using a Bellman backup.
DynamicProgramming() - Constructor for class burlap.behavior.singleagent.planning.stochastic.DynamicProgramming
 
DynamicWeightedAStar - Class in burlap.behavior.singleagent.planning.deterministic.informed.astar
Dynamic Weighted A* [1] uses a dynamic heuristic weight that is based on depth of the current search tree and based on an expected depth of the search.
DynamicWeightedAStar(SADomain, StateConditionTest, HashableStateFactory, Heuristic, double, int) - Constructor for class burlap.behavior.singleagent.planning.deterministic.informed.astar.DynamicWeightedAStar
Initializes
A B C D E F G H I J K L M N O P Q R S T U V W X Y Z 
Skip navigation links