Skip navigation links
A B C D E F G H I J K L M N O P Q R S T U V W X Y Z 

P

p - Variable in class burlap.mdp.core.StateTransitionProb
 
p - Variable in class burlap.mdp.singleagent.model.TransitionProb
The probability of the transition
p - Variable in class burlap.mdp.singleagent.pomdp.observations.ObservationProbability
The probability of the observation
p0 - Variable in class burlap.mdp.stochasticgames.tournament.common.AllPairWiseSameTypeMS
 
p1 - Variable in class burlap.mdp.stochasticgames.tournament.common.AllPairWiseSameTypeMS
 
pad - Variable in class burlap.domain.singleagent.lunarlander.state.LLState
 
PadPainter(LunarLanderDomain.LLPhysicsParams) - Constructor for class burlap.domain.singleagent.lunarlander.LLVisualizer.PadPainter
 
paint(Graphics2D, float, float) - Method in interface burlap.behavior.singleagent.auxiliary.valuefunctionvis.StaticDomainPainter
Use to paint general domain information to a 2D graphics context.
paint(Graphics2D, State, float, float) - Method in class burlap.domain.singleagent.cartpole.CartPoleVisualizer.CartPolePainter
 
paint(Graphics2D, State, float, float) - Method in class burlap.domain.singleagent.gridworld.GridWorldVisualizer.MapPainter
 
paint(Graphics2D, State, float, float) - Method in class burlap.domain.singleagent.mountaincar.MountainCarVisualizer.AgentPainter
 
paint(Graphics2D, State, float, float) - Method in class burlap.domain.singleagent.mountaincar.MountainCarVisualizer.HillPainter
 
paint(Graphics2D, State, float, float) - Method in class burlap.visualizer.OOStatePainter
 
paint(Graphics2D, State, float, float) - Method in interface burlap.visualizer.StatePainter
Paints general state information not to graphics context g2
paintComponent(Graphics) - Method in class burlap.visualizer.MultiLayerRenderer
 
painter - Variable in class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
 
painter - Variable in class burlap.behavior.stochasticgames.auxiliary.GameSequenceVisualizer
 
painter - Variable in class burlap.mdp.singleagent.common.VisualActionObserver
The visualizer that will render states
painter - Variable in class burlap.mdp.stochasticgames.common.VisualWorldObserver
The visualizer that will render states
painter - Variable in class burlap.shell.visual.SGVisualExplorer
 
painter - Variable in class burlap.shell.visual.VisualExplorer
 
paintGlyph(Graphics2D, float, float, float, float) - Method in interface burlap.behavior.singleagent.auxiliary.valuefunctionvis.common.ActionGlyphPainter
Called to paint a glyph in the rectangle defined by the top left origin (x,y) with the given width and height.
paintGlyph(Graphics2D, float, float, float, float) - Method in class burlap.behavior.singleagent.auxiliary.valuefunctionvis.common.ArrowActionGlyph
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.blockdude.BlockDudeVisualizer.AgentPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.blockdude.BlockDudeVisualizer.BlockPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.blockdude.BlockDudeVisualizer.BricksPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.blockdude.BlockDudeVisualizer.ExitPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.blocksworld.BlocksWorldVisualizer.BlockPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.frostbite.FrostbiteVisualizer.AgentPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.frostbite.FrostbiteVisualizer.IglooPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.frostbite.FrostbiteVisualizer.PlatformPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.gridworld.GridWorldVisualizer.CellPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.gridworld.GridWorldVisualizer.LocationPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.lunarlander.LLVisualizer.AgentPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.lunarlander.LLVisualizer.ObstaclePainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in class burlap.domain.singleagent.lunarlander.LLVisualizer.PadPainter
 
paintObject(Graphics2D, OOState, ObjectInstance, float, float) - Method in interface burlap.visualizer.ObjectPainter
Paints object instance ob to graphics context g2
paintStatePolicy(Graphics2D, State, Policy, float, float) - Method in class burlap.behavior.singleagent.auxiliary.valuefunctionvis.common.PolicyGlyphPainter2D
 
paintStatePolicy(Graphics2D, State, Policy, float, float) - Method in interface burlap.behavior.singleagent.auxiliary.valuefunctionvis.StatePolicyPainter
Paints a representation of the given policy for a specific state to a 2D graphics context.
paintStateValue(Graphics2D, State, double, float, float) - Method in class burlap.behavior.singleagent.auxiliary.valuefunctionvis.common.StateValuePainter2D
 
paintStateValue(Graphics2D, State, double, float, float) - Method in class burlap.behavior.singleagent.auxiliary.valuefunctionvis.StateValuePainter
Paints the representation of a value function for a specific state.
parameterClasses - Variable in class burlap.mdp.core.oo.propositional.PropositionalFunction
 
parameterClasses - Variable in class burlap.mdp.singleagent.oo.ObjectParameterizedActionType
The object classes each parameter of this action can accept; empty list for a parameter-less action (which is the default)
parameterId - Variable in class burlap.behavior.functionapproximation.FunctionGradient.PartialDerivative
 
parameterOrderGroup - Variable in class burlap.mdp.core.oo.propositional.PropositionalFunction
 
parameterOrderGroup - Variable in class burlap.mdp.singleagent.oo.ObjectParameterizedActionType
Specifies the parameter order group each parameter.
parameters - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.commonrfs.LinearStateActionDifferentiableRF
The parameters of this reward function
parameters - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.commonrfs.LinearStateDifferentiableRF
The parameters of this reward function
parameters - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.diffvinit.LinearDiffRFVInit
 
parameters - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.diffvinit.LinearStateDiffVF
 
ParametricFunction - Interface in burlap.behavior.functionapproximation
An interface for defining a parametric function.
ParametricFunction.ParametricStateActionFunction - Interface in burlap.behavior.functionapproximation
A parametric function that operations on state-actions
ParametricFunction.ParametricStateFunction - Interface in burlap.behavior.functionapproximation
A Parametric function that operates on states.
params - Variable in class burlap.mdp.core.oo.propositional.GroundedProp
 
params - Variable in class burlap.mdp.singleagent.oo.ObjectParameterizedActionType.SAObjectParameterizedAction
 
parse(String) - Static method in class burlap.behavior.stochasticgames.GameEpisode
 
parseEpisode(String) - Static method in class burlap.behavior.singleagent.Episode
 
parseEpisodeFiles(String) - Method in class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
 
parser - Variable in class burlap.shell.command.env.AddStateObjectCommand
 
parser - Variable in class burlap.shell.command.env.EpisodeRecordingCommands.EpisodeBrowserCommand
 
parser - Variable in class burlap.shell.command.env.EpisodeRecordingCommands.RecordCommand
 
parser - Variable in class burlap.shell.command.env.ExecuteActionCommand
 
parser - Variable in class burlap.shell.command.env.IsTerminalCommand
 
parser - Variable in class burlap.shell.command.env.ListActionsCommand
 
parser - Variable in class burlap.shell.command.env.ListPropFunctions
 
parser - Variable in class burlap.shell.command.env.ObservationCommand
 
parser - Variable in class burlap.shell.command.env.RemoveStateObjectCommand
 
parser - Variable in class burlap.shell.command.env.RewardCommand
 
parser - Variable in class burlap.shell.command.env.SetVarCommand
 
parser - Variable in class burlap.shell.command.reserved.AliasCommand
 
parser - Variable in class burlap.shell.command.reserved.AliasesCommand
 
parser - Variable in class burlap.shell.command.reserved.CommandsCommand
 
parser - Variable in class burlap.shell.command.world.AddStateObjectSGCommand
 
parser - Variable in class burlap.shell.command.world.GameCommand
 
parser - Variable in class burlap.shell.command.world.GenerateStateCommand
 
parser - Variable in class burlap.shell.command.world.IsTerminalSGCommand
 
parser - Variable in class burlap.shell.command.world.JointActionCommand
 
parser - Variable in class burlap.shell.command.world.LastJointActionCommand
 
parser - Variable in class burlap.shell.command.world.ManualAgentsCommands.ListManualAgents
 
parser - Variable in class burlap.shell.command.world.ManualAgentsCommands.LSManualAgentActionsCommands
 
parser - Variable in class burlap.shell.command.world.ManualAgentsCommands.RegisterAgentCommand
 
parser - Variable in class burlap.shell.command.world.ManualAgentsCommands.SetAgentAction
 
parser - Variable in class burlap.shell.command.world.RemoveStateObjectSGCommand
 
parser - Variable in class burlap.shell.command.world.RewardsCommand
 
parser - Variable in class burlap.shell.command.world.SetVarSGCommand
 
parser - Variable in class burlap.shell.command.world.WorldObservationCommand
 
parser - Variable in class burlap.shell.visual.VisualExplorer.LivePollCommand
 
PartialDerivative(int, double) - Constructor for class burlap.behavior.functionapproximation.FunctionGradient.PartialDerivative
 
payouts - Variable in class burlap.domain.stochasticgames.normalform.SingleStageNormalFormGame
The pay out function for each player.
peek() - Method in class burlap.datastructures.HashIndexedHeap
Returns a pointer to the head of the heap without removing it
peekAtLearningRate(State, Action) - Method in class burlap.behavior.functionapproximation.dense.fourier.FourierBasisLearningRateWrapper
 
peekAtLearningRate(int) - Method in class burlap.behavior.functionapproximation.dense.fourier.FourierBasisLearningRateWrapper
 
peekAtLearningRate(State, Action) - Method in class burlap.behavior.learningrate.ConstantLR
 
peekAtLearningRate(int) - Method in class burlap.behavior.learningrate.ConstantLR
 
peekAtLearningRate(State, Action) - Method in class burlap.behavior.learningrate.ExponentialDecayLR
 
peekAtLearningRate(int) - Method in class burlap.behavior.learningrate.ExponentialDecayLR
 
peekAtLearningRate(State, Action) - Method in interface burlap.behavior.learningrate.LearningRate
A method for looking at the current learning rate for a state-action pair without having it altered.
peekAtLearningRate(int) - Method in interface burlap.behavior.learningrate.LearningRate
A method for looking at the current learning rate for a state (-action) feature without having it altered.
peekAtLearningRate(State, Action) - Method in class burlap.behavior.learningrate.SoftTimeInverseDecayLR
 
peekAtLearningRate(int) - Method in class burlap.behavior.learningrate.SoftTimeInverseDecayLR
 
peekAtTime() - Method in class burlap.debugtools.MyTimer
Returns the current elapsed time since the timer was started.
percolateWeightChange(StochasticTree<T>.STNode, double) - Method in class burlap.datastructures.StochasticTree
A recursive method for percolating a weight change of a node
PerformanceMetric - Enum in burlap.behavior.singleagent.auxiliary.performance
Enumerator for the types of statistics that can be plotted by PerformancePlotter.
PerformancePlotter - Class in burlap.behavior.singleagent.auxiliary.performance
This class is an action observer used to collect and plot performance data of a learning agent either by itself or against another learning agent.
PerformancePlotter(String, int, int, int, int, TrialMode, PerformanceMetric...) - Constructor for class burlap.behavior.singleagent.auxiliary.performance.PerformancePlotter
Initializes a performance plotter.
PerformancePlotter.AgentDatasets - Class in burlap.behavior.singleagent.auxiliary.performance
A datastructure for maintain the plot series data in the current agent
PerformancePlotter.MutableBoolean - Class in burlap.behavior.singleagent.auxiliary.performance
A class for a mutable boolean
PerformancePlotter.Trial - Class in burlap.behavior.singleagent.auxiliary.performance
A datastructure for maintaining all the metric stats for a single trial.
performBackup(State, int, List<SGAgentType>, AgentQSourceMap) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.backupOperators.CoCoQ
 
performBackup(State, int, List<SGAgentType>, AgentQSourceMap) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.backupOperators.CorrelatedQ
 
performBackup(State, int, List<SGAgentType>, AgentQSourceMap) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.backupOperators.MaxQ
 
performBackup(State, int, List<SGAgentType>, AgentQSourceMap) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.backupOperators.MinMaxQ
 
performBackup(State, int, List<SGAgentType>, AgentQSourceMap) - Method in interface burlap.behavior.stochasticgames.madynamicprogramming.SGBackupOperator
 
performBellmanUpdateOn(State) - Method in class burlap.behavior.singleagent.planning.stochastic.DynamicProgramming
Performs a Bellman value function update on the provided state.
performBellmanUpdateOn(HashableState) - Method in class burlap.behavior.singleagent.planning.stochastic.DynamicProgramming
Performs a Bellman value function update on the provided (hashed) state.
performDPValueGradientUpdateOn(HashableState) - Method in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableDP
Performs the Boltzmann value function gradient backup for the given HashableState.
performFixedPolicyBellmanUpdateOn(State, EnumerablePolicy) - Method in class burlap.behavior.singleagent.planning.stochastic.DynamicProgramming
Performs a fixed-policy Bellman value function update (i.e., policy evaluation) on the provided state.
performFixedPolicyBellmanUpdateOn(HashableState, EnumerablePolicy) - Method in class burlap.behavior.singleagent.planning.stochastic.DynamicProgramming
Performs a fixed-policy Bellman value function update (i.e., policy evaluation) on the provided state.
performIRL() - Method in class burlap.behavior.singleagent.learnfromdemo.mlirl.MLIRL
Runs gradient ascent.
performIRL() - Method in class burlap.behavior.singleagent.learnfromdemo.mlirl.MultipleIntentionsMLIRL
Performs multiple intention inverse reinforcement learning.
performOrderedBellmanUpdates(List<HashableState>) - Method in class burlap.behavior.singleagent.planning.stochastic.rtdp.RTDP
Performs ordered Bellman updates on the list of (hashed) states provided to it.
performReachabilityFrom(State) - Method in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableVI
This method will find all reachable states that will be used by the DifferentiableVI.runVI() method and will cache all the transition dynamics.
performReachabilityFrom(State) - Method in class burlap.behavior.singleagent.planning.stochastic.policyiteration.PolicyEvaluation
This method will find all reachable states that will be used when computing the value function.
performReachabilityFrom(State) - Method in class burlap.behavior.singleagent.planning.stochastic.policyiteration.PolicyIteration
This method will find all reachable states that will be used when computing the value function.
performReachabilityFrom(State) - Method in class burlap.behavior.singleagent.planning.stochastic.valueiteration.PrioritizedSweeping
 
performReachabilityFrom(State) - Method in class burlap.behavior.singleagent.planning.stochastic.valueiteration.ValueIteration
This method will find all reachable states that will be used by the ValueIteration.runVI() method and will cache all the transition dynamics.
performStateReachabilityFrom(State) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.dpplanners.MAValueIteration
Finds and stores all states that are reachable from input state s.
pf - Variable in class burlap.mdp.core.oo.propositional.GroundedProp
 
PF_AT_EXIT - Static variable in class burlap.domain.singleagent.blockdude.BlockDude
Name for the propositional function that tests whether the agent is at an exit
PF_AT_LOCATION - Static variable in class burlap.domain.singleagent.gridworld.GridWorldDomain
Constant for the name of the at location propositional function
PF_CLEAR - Static variable in class burlap.domain.singleagent.blocksworld.BlocksWorld
Constant for the propositional function "clear" name
PF_HOLDING_BLOCK - Static variable in class burlap.domain.singleagent.blockdude.BlockDude
Name for the propositional function that tests whether the agent is holding a block.
PF_IGLOO_BUILT - Static variable in class burlap.domain.singleagent.frostbite.FrostbiteDomain
Constant for the name of the propositional function "igloo is built"
PF_IN_P_GOAL - Static variable in class burlap.domain.stochasticgames.gridgame.GridGame
A constant for the name of a propositional function that evaluates whether an agent is in a personal goal location for just them.
PF_IN_U_GOAL - Static variable in class burlap.domain.stochasticgames.gridgame.GridGame
A constant for the name of a propositional function that evaluates whether an agent is in a universal goal location.
PF_IN_WATER - Static variable in class burlap.domain.singleagent.frostbite.FrostbiteDomain
Constant for the name of the propositional function "agent is in water"
PF_ON_BLOCK - Static variable in class burlap.domain.singleagent.blocksworld.BlocksWorld
Constant for the propositional function "on" name
PF_ON_GROUND - Static variable in class burlap.domain.singleagent.lunarlander.LunarLanderDomain
Constant for the name of the propositional function that indicates whether the agent/lander is on the ground
PF_ON_ICE - Static variable in class burlap.domain.singleagent.frostbite.FrostbiteDomain
Constant for the name of the propositional function "agent is on ice"
PF_ON_PAD - Static variable in class burlap.domain.singleagent.lunarlander.LunarLanderDomain
Constant for the name of the propositional function that indicates whether the agent/lander is on a landing pad
PF_ON_PLATFORM - Static variable in class burlap.domain.singleagent.frostbite.FrostbiteDomain
Constant for the name of the propositional function "agent is on platform"
PF_ON_TABLE - Static variable in class burlap.domain.singleagent.blocksworld.BlocksWorld
Constant for the propositional function "on table" name
PF_PLATFORM_ACTIVE - Static variable in class burlap.domain.singleagent.frostbite.FrostbiteDomain
Constant for the name of the propositional function "platform is active"
PF_TOUCH_SURFACE - Static variable in class burlap.domain.singleagent.lunarlander.LunarLanderDomain
Constant for the name of the propositional function that indicates whether the agent/lander is touching an obstacle surface.
PF_TOUTCH_PAD - Static variable in class burlap.domain.singleagent.lunarlander.LunarLanderDomain
Constant for the name of the propositional function that indicates whether the agent/lander is *touching* a landing pad.
PF_WALL_EAST - Static variable in class burlap.domain.singleagent.gridworld.GridWorldDomain
Constant for the name of the wall to east propositional function
PF_WALL_NORTH - Static variable in class burlap.domain.singleagent.gridworld.GridWorldDomain
Constant for the name of the wall to north propositional function
PF_WALL_SOUTH - Static variable in class burlap.domain.singleagent.gridworld.GridWorldDomain
Constant for the name of the wall to south propositional function
PF_WALL_WEST - Static variable in class burlap.domain.singleagent.gridworld.GridWorldDomain
Constant for the name of the wall to west propositional function
PFFeatures - Class in burlap.behavior.functionapproximation.dense
Binary features that are determined from a list of PropositionalFunctions.
PFFeatures(OODomain) - Constructor for class burlap.behavior.functionapproximation.dense.PFFeatures
Initializes using all propositional functions that belong to the domain
PFFeatures(List<PropositionalFunction>) - Constructor for class burlap.behavior.functionapproximation.dense.PFFeatures
Initializes using the list of given propositional functions.
PFFeatures(PropositionalFunction[]) - Constructor for class burlap.behavior.functionapproximation.dense.PFFeatures
Initializes using the array of given propositional functions.
pfsToUse - Variable in class burlap.behavior.functionapproximation.dense.PFFeatures
 
phiConstructor(double[], int) - Method in class burlap.behavior.singleagent.learning.lspi.LSPI
Constructs the state-action feature vector as a SimpleMatrix.
physParams - Variable in class burlap.domain.singleagent.cartpole.CartPoleDomain
An object specifying the physics parameters for the cart pole domain.
physParams - Variable in class burlap.domain.singleagent.cartpole.InvertedPendulum
 
physParams - Variable in class burlap.domain.singleagent.cartpole.model.CPClassicModel
 
physParams - Variable in class burlap.domain.singleagent.cartpole.model.CPCorrectModel
 
physParams - Variable in class burlap.domain.singleagent.lunarlander.LunarLanderDomain
An object for holding the physics parameters of this domain.
physParams - Variable in class burlap.domain.singleagent.mountaincar.MountainCar.MCModel
 
physParams - Variable in class burlap.domain.singleagent.mountaincar.MountainCar
The physics parameters for mountain car.
pickupBlock(BlockDudeState) - Method in class burlap.domain.singleagent.blockdude.BlockDudeModel
Modifies state s to be the result of the pick up action.
planContainsOption(SearchNode) - Method in class burlap.behavior.singleagent.planning.deterministic.DeterministicPlanner
Returns true if a solution path uses an option in its solution.
planEndNode(SearchNode) - Method in class burlap.behavior.singleagent.planning.deterministic.informed.astar.IDAStar
Returns true if the search node wraps a goal state.
planFromState(State) - Method in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableSparseSampling
Plans from the input state and returns a BoltzmannQPolicy following the Boltzmann parameter used for value Botlzmann value backups in this planner.
planFromState(State) - Method in class burlap.behavior.singleagent.learnfromdemo.mlirl.differentiableplanners.DifferentiableVI
Plans from the input state and returns a BoltzmannQPolicy following the Boltzmann parameter used for value Botlzmann value backups in this planner.
planFromState(State) - Method in class burlap.behavior.singleagent.learning.actorcritic.ActorCritic
 
planFromState(State) - Method in class burlap.behavior.singleagent.learning.lspi.LSPI
Plans from the input state and then returns a GreedyQPolicy that greedily selects the action with the highest Q-value and breaks ties uniformly randomly.
planFromState(State) - Method in class burlap.behavior.singleagent.learning.tdmethods.QLearning
Plans from the input state and then returns a GreedyQPolicy that greedily selects the action with the highest Q-value and breaks ties uniformly randomly.
planFromState(State) - Method in class burlap.behavior.singleagent.learning.tdmethods.vfa.GradientDescentSarsaLam
Plans from the input state and then returns a GreedyQPolicy that greedily selects the action with the highest Q-value and breaks ties uniformly randomly.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.deterministic.informed.astar.DynamicWeightedAStar
Plans and returns a SDPlannerPolicy.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.deterministic.informed.astar.IDAStar
Plans and returns a SDPlannerPolicy.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.deterministic.informed.BestFirst
Plans and returns a SDPlannerPolicy.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.deterministic.uninformed.bfs.BFS
Plans and returns a SDPlannerPolicy.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.deterministic.uninformed.dfs.DFS
Plans and returns a SDPlannerPolicy.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.deterministic.uninformed.dfs.LimitedMemoryDFS
Plans and returns a SDPlannerPolicy.
planFromState(State) - Method in interface burlap.behavior.singleagent.planning.Planner
This method will cause the Planner to begin planning from the specified initial State.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.stochastic.montecarlo.uct.UCT
Plans from the input state and then returns a GreedyQPolicy that greedily selects the action with the highest Q-value and breaks ties uniformly randomly.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.stochastic.policyiteration.PolicyIteration
Plans from the input state and then returns a GreedyQPolicy that greedily selects the action with the highest Q-value and breaks ties uniformly randomly.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.stochastic.rtdp.BoundedRTDP
Plans from the input state and then returns a GreedyQPolicy that greedily selects the action with the highest Q-value and breaks ties uniformly randomly.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.stochastic.rtdp.RTDP
Plans from the input state and then returns a GreedyQPolicy that greedily selects the action with the highest Q-value and breaks ties uniformly randomly.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.stochastic.sparsesampling.SparseSampling
Plans from the input state and then returns a GreedyQPolicy that greedily selects the action with the highest Q-value and breaks ties uniformly randomly.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.stochastic.valueiteration.ValueIteration
Plans from the input state and then returns a GreedyQPolicy that greedily selects the action with the highest Q-value and breaks ties uniformly randomly.
planFromState(State) - Method in class burlap.behavior.singleagent.planning.vfa.fittedvi.FittedVI
Plans from the input state and then returns a GreedyQPolicy that greedily selects the action with the highest Q-value and breaks ties uniformly randomly.
planFromState(State) - Method in class burlap.behavior.singleagent.pomdp.qmdp.QMDP
 
planFromState(State) - Method in class burlap.behavior.singleagent.pomdp.wrappedmdpalgs.BeliefSparseSampling
 
planFromState(State) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.dpplanners.MAValueIteration
 
planFromState(State) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.MADynamicProgramming
Calling this method causes planning to be performed from State s.
planHasDupilicateStates(SearchNode) - Method in class burlap.behavior.singleagent.planning.deterministic.DeterministicPlanner
Returns true if a solution path visits the same state multiple times.
planner - Variable in class burlap.behavior.singleagent.learnfromdemo.IRLRequest
The planning algorithm used to compute the policy for a given reward function
Planner - Interface in burlap.behavior.singleagent.planning
 
planner - Variable in class burlap.behavior.stochasticgames.agents.madp.MultiAgentDPPlanningAgent
The valueFunction this agent will use to estiamte the value function and thereby determine its policy.
plannerFactory - Variable in class burlap.behavior.singleagent.learnfromdemo.mlirl.MultipleIntentionsMLIRLRequest
plannerFactory - Variable in class burlap.behavior.stochasticgames.agents.madp.MADPPlanAgentFactory
 
plannerReferece - Variable in class burlap.behavior.stochasticgames.agents.madp.MADPPlannerFactory.ConstantMADPPlannerFactory
 
planningCollector - Variable in class burlap.behavior.singleagent.learning.lspi.LSPI
The data collector used by the LSPI.planFromState(State) method.
planningDepth - Variable in class burlap.behavior.singleagent.planning.vfa.fittedvi.FittedVI
The SparseSampling planning depth used for computing Bellman operators during value iteration.
PlanningFailedException() - Constructor for exception burlap.behavior.singleagent.planning.deterministic.DeterministicPlanner.PlanningFailedException
 
planningStarted - Variable in class burlap.behavior.stochasticgames.madynamicprogramming.MADynamicProgramming
Whether planning has begun or not.
PlatformActivePF(String) - Constructor for class burlap.domain.singleagent.frostbite.FrostbiteDomain.PlatformActivePF
Initializes to be evaluated on an agent object and platform object.
PlatformPainter(int) - Constructor for class burlap.domain.singleagent.frostbite.FrostbiteVisualizer.PlatformPainter
 
platforms - Variable in class burlap.domain.singleagent.frostbite.state.FrostbiteState
 
platformSpeed - Variable in class burlap.domain.singleagent.frostbite.FrostbiteModel
 
pLayer - Variable in class burlap.behavior.singleagent.auxiliary.valuefunctionvis.ValueFunctionVisualizerGUI
The policy renderer
player - Variable in class burlap.domain.stochasticgames.gridgame.state.GGAgent
 
playerIndex(String) - Method in class burlap.domain.stochasticgames.normalform.NFGameState
 
players - Variable in class burlap.domain.stochasticgames.normalform.NFGameState
 
plotCISignificance - Variable in class burlap.behavior.singleagent.auxiliary.performance.LearningAlgorithmExperimenter
The signficance value for the confidence interval in the plots.
plotCISignificance - Variable in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentExperimenter
The signficance value for the confidence interval in the plots.
plotRefresh - Variable in class burlap.behavior.singleagent.auxiliary.performance.LearningAlgorithmExperimenter
The delay in milliseconds between autmatic refreshes of the plots
plotRefresh - Variable in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentExperimenter
The delay in milliseconds between autmatic refreshes of the plots
plotter - Variable in class burlap.behavior.singleagent.auxiliary.performance.LearningAlgorithmExperimenter
The PerformancePlotter used to collect and plot results
plotter - Variable in class burlap.behavior.stochasticgames.auxiliary.performance.MultiAgentExperimenter
The performance plotter object
poDomain - Variable in class burlap.mdp.singleagent.pomdp.BeliefAgent
The POMDP Domain defining the environment mechanics.
poDomain - Variable in class burlap.mdp.singleagent.pomdp.BeliefMDPGenerator.BeliefModel
 
podomain - Variable in class burlap.mdp.singleagent.pomdp.BeliefMDPGenerator
The input POMDP domain
PODomain - Class in burlap.mdp.singleagent.pomdp
A class for defining POMDP domains.
PODomain() - Constructor for class burlap.mdp.singleagent.pomdp.PODomain
 
poDomain - Variable in class burlap.mdp.singleagent.pomdp.SimulatedPOEnvironment
 
poleFriction - Variable in class burlap.domain.singleagent.cartpole.CartPoleDomain.CPPhysicsParams
The friction between the pole and the joint on the cart.
poleLength - Variable in class burlap.domain.singleagent.cartpole.InvertedPendulum.IPPhysicsParams
The length of the pole
poleMass - Variable in class burlap.domain.singleagent.cartpole.CartPoleDomain.CPPhysicsParams
The mass of the pole.
poleMass - Variable in class burlap.domain.singleagent.cartpole.InvertedPendulum.IPPhysicsParams
The mass of the pole.
Policy - Interface in burlap.behavior.policy
An interface for defining a Policy.
policy - Variable in class burlap.behavior.singleagent.auxiliary.valuefunctionvis.PolicyRenderLayer
The policy to use for visualizing the policy
policy - Variable in class burlap.behavior.singleagent.learning.modellearning.artdp.ARTDP
the policy to follow
policy(State, Episode) - Method in class burlap.behavior.singleagent.options.MacroAction
 
policy(State, Episode) - Method in interface burlap.behavior.singleagent.options.Option
 
policy - Variable in class burlap.behavior.singleagent.options.SubgoalOption
The policy of the options
policy(State, Episode) - Method in class burlap.behavior.singleagent.options.SubgoalOption
 
policy - Variable in class burlap.behavior.singleagent.pomdp.BeliefPolicyAgent
The policy that the agent will follow.
policy - Variable in class burlap.behavior.stochasticgames.agents.madp.MADPPlanAgentFactory
 
policy - Variable in class burlap.behavior.stochasticgames.agents.madp.MultiAgentDPPlanningAgent
The policy dervied from a joint policy derived from the valueFunction's value function estimate that this agent will follow.
policy - Variable in class burlap.behavior.stochasticgames.agents.naiveq.SGNaiveQLAgent
The policy this agent follows
policy - Variable in class burlap.behavior.stochasticgames.agents.SetStrategySGAgent
The policy encoding the strategy this agent will follow
policy - Variable in class burlap.behavior.stochasticgames.agents.SetStrategySGAgent.SetStrategyAgentFactory
The strategy this agent will follow
policyCount - Variable in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearningRequest
The maximum number of times a policy is rolled out and evaluated
policyDistribution(State) - Method in class burlap.behavior.policy.BoltzmannQPolicy
 
policyDistribution(State) - Method in class burlap.behavior.policy.CachedPolicy
 
policyDistribution(State) - Method in interface burlap.behavior.policy.EnumerablePolicy
This method will return action probability distribution defined by the policy.
policyDistribution(State) - Method in class burlap.behavior.policy.EpsilonGreedy
 
policyDistribution(State) - Method in class burlap.behavior.policy.GreedyDeterministicQPolicy
 
policyDistribution(State) - Method in class burlap.behavior.policy.GreedyQPolicy
 
policyDistribution(State) - Method in class burlap.behavior.policy.RandomPolicy
 
policyDistribution(State) - Method in class burlap.behavior.singleagent.learnfromdemo.apprenticeship.ApprenticeshipLearning.StationaryRandomDistributionPolicy
 
policyDistribution(State) - Method in class burlap.behavior.singleagent.learning.actorcritic.actor.BoltzmannActor
 
policyDistribution(State) - Method in class burlap.behavior.singleagent.learning.modellearning.rmax.UnmodeledFavoredPolicy
 
policyDistribution(State, Episode) - Method in class burlap.behavior.singleagent.options.MacroAction
 
policyDistribution(State, Episode) - Method in interface burlap.behavior.singleagent.options.Option
 
policyDistribution(State, Episode) - Method in class burlap.behavior.singleagent.options.SubgoalOption
 
policyDistribution(State) - Method in class burlap.behavior.singleagent.planning.deterministic.DDPlannerPolicy
 
policyDistribution(State) - Method in class burlap.behavior.singleagent.planning.deterministic.SDPlannerPolicy
 
policyDistribution(State) - Method in class burlap.behavior.singleagent.planning.stochastic.montecarlo.uct.UCTTreeWalkPolicy
 
policyDistribution(State) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.policies.ECorrelatedQJointPolicy
 
policyDistribution(State) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyJointPolicy
 
policyDistribution(State) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.policies.EGreedyMaxWellfare
 
policyDistribution(State) - Method in class burlap.behavior.stochasticgames.madynamicprogramming.policies.EMinMaxPolicy
 
policyDistribution(State) - Method in class burlap.behavior.stochasticgames.PolicyFromJointPolicy
 
PolicyEvaluation - Class in burlap.behavior.singleagent.planning.stochastic.policyiteration
This class is used to compute the value function under some specified policy.
PolicyEvaluation(SADomain, double, HashableStateFactory, double, double) - Constructor for class burlap.behavior.singleagent.planning.stochastic.policyiteration.PolicyEvaluation
Initializes.
PolicyFromJointPolicy - Class in burlap.behavior.stochasticgames
This class defines a single agent's policy that is derived from a joint policy.
PolicyFromJointPolicy(JointPolicy) - Constructor for class burlap.behavior.stochasticgames.PolicyFromJointPolicy
Initializes with the underlying joint polciy
PolicyFromJointPolicy(JointPolicy, boolean) - Constructor for class burlap.behavior.stochasticgames.PolicyFromJointPolicy
Initializes with the underlying joint polciy and whether actions should be synchronized with other agents following the same underlying joint policy.
PolicyFromJointPolicy(int, JointPolicy) - Constructor for class burlap.behavior.stochasticgames.PolicyFromJointPolicy
Initializes with the acting agent name whose actions from the underlying joint policy will be returned.
PolicyFromJointPolicy(int, JointPolicy, boolean) - Constructor for class burlap.behavior.stochasticgames.PolicyFromJointPolicy
Initializes with the acting agent name whose actions from the underlying joint policy will be returned and whether actions should be synchronized with other agents following the same underlying joint policy.
PolicyGlyphPainter2D - Class in burlap.behavior.singleagent.auxiliary.valuefunctionvis.common
An class for rendering the policy for states by painting different glyphs for different actions.
PolicyGlyphPainter2D() - Constructor for class burlap.behavior.singleagent.auxiliary.valuefunctionvis.common.PolicyGlyphPainter2D
 
PolicyGlyphPainter2D.PolicyGlyphRenderStyle - Enum in burlap.behavior.singleagent.auxiliary.valuefunctionvis.common
MAXACTION paints only glphys for only those actions that have the highest likelihood
PolicyIteration - Class in burlap.behavior.singleagent.planning.stochastic.policyiteration
 
PolicyIteration(SADomain, double, HashableStateFactory, double, int, int) - Constructor for class burlap.behavior.singleagent.planning.stochastic.policyiteration.PolicyIteration
Initializes the valueFunction.
PolicyIteration(SADomain, double, HashableStateFactory, double, double, int, int) - Constructor for class burlap.behavior.singleagent.planning.stochastic.policyiteration.PolicyIteration
Initializes the valueFunction.
PolicyRenderLayer - Class in burlap.behavior.singleagent.auxiliary.valuefunctionvis
 
PolicyRenderLayer(Collection<State>, StatePolicyPainter, Policy) - Constructor for class burlap.behavior.singleagent.auxiliary.valuefunctionvis.PolicyRenderLayer
 
PolicyUndefinedException - Exception in burlap.behavior.policy.support
RuntimeException to be thrown when a Policy is queried for a state in which the policy is undefined.
PolicyUndefinedException() - Constructor for exception burlap.behavior.policy.support.PolicyUndefinedException
 
PolicyUtils - Class in burlap.behavior.policy
 
policyValue(QProvider, State, EnumerablePolicy) - Static method in class burlap.behavior.valuefunction.QProvider.Helper
Returns the state value under a given policy for a state and QProvider.
poll() - Method in class burlap.datastructures.HashIndexedHeap
Returns a pointer to the head of the heap and removes it
poll() - Method in class burlap.datastructures.StochasticTree
Samples an element according to a probability defined by the relative weight of objects, removes it from the tree, and returns it.
pollInterval - Variable in class burlap.shell.visual.VisualExplorer
 
pollLearningRate(int, State, Action) - Method in class burlap.behavior.functionapproximation.dense.fourier.FourierBasisLearningRateWrapper
 
pollLearningRate(int, int) - Method in class burlap.behavior.functionapproximation.dense.fourier.FourierBasisLearningRateWrapper
 
pollLearningRate(int, State, Action) - Method in class burlap.behavior.learningrate.ConstantLR
 
pollLearningRate(int, int) - Method in class burlap.behavior.learningrate.ConstantLR
 
pollLearningRate(int, State, Action) - Method in class burlap.behavior.learningrate.ExponentialDecayLR
 
pollLearningRate(int, int) - Method in class burlap.behavior.learningrate.ExponentialDecayLR
 
pollLearningRate(int, State, Action) - Method in interface burlap.behavior.learningrate.LearningRate
A method for returning the learning rate for a given state action pair and then decaying the learning rate as defined by this class.
pollLearningRate(int, int) - Method in interface burlap.behavior.learningrate.LearningRate
A method for returning the learning rate for a given state (-action) feature and then decaying the learning rate as defined by this class.
pollLearningRate(int, State, Action) - Method in class burlap.behavior.learningrate.SoftTimeInverseDecayLR
 
pollLearningRate(int, int) - Method in class burlap.behavior.learningrate.SoftTimeInverseDecayLR
 
polyDegree - Variable in class burlap.behavior.singleagent.auxiliary.valuefunctionvis.common.LandmarkColorBlendInterpolation
The power to raise the normalized distance
pos - Variable in class burlap.domain.stochasticgames.gridgame.state.GGWall
 
postPlanPrep() - Method in class burlap.behavior.singleagent.planning.deterministic.informed.astar.AStar
 
postPlanPrep() - Method in class burlap.behavior.singleagent.planning.deterministic.informed.astar.DynamicWeightedAStar
 
postPlanPrep() - Method in class burlap.behavior.singleagent.planning.deterministic.informed.BestFirst
This method is called at the end of the BestFirst.planFromState(State) method and can be used clean up any special data structures needed by the subclass.
potentialFunction - Variable in class burlap.behavior.singleagent.learning.modellearning.rmax.RMaxModel
 
PotentialFunction - Interface in burlap.behavior.singleagent.shaping.potential
Defines an interface for reward potential functions.
potentialFunction - Variable in class burlap.behavior.singleagent.shaping.potential.PotentialShapedRF
The potential function that can be used to return the potential reward from input states.
PotentialShapedRF - Class in burlap.behavior.singleagent.shaping.potential
This class is used to implement Potential-based reward shaping [1] which is guaranteed to preserve the optimal policy.
PotentialShapedRF(RewardFunction, PotentialFunction, double) - Constructor for class burlap.behavior.singleagent.shaping.potential.PotentialShapedRF
Initializes the shaping with the objective reward function, the potential function, and the discount of the MDP.
PotentialShapedRMax - Class in burlap.behavior.singleagent.learning.modellearning.rmax
Potential Shaped RMax [1] is a generalization of RMax in which a potential-shaped reward function is used to provide less (but still admissible) optimistic views of unknown state transitions.
PotentialShapedRMax(SADomain, double, HashableStateFactory, double, int, double, int) - Constructor for class burlap.behavior.singleagent.learning.modellearning.rmax.PotentialShapedRMax
Initializes for a tabular model, VI valueFunction, and standard RMax paradigm
PotentialShapedRMax(SADomain, double, HashableStateFactory, PotentialFunction, int, double, int) - Constructor for class burlap.behavior.singleagent.learning.modellearning.rmax.PotentialShapedRMax
Initializes for a tabular model, VI valueFunction, and potential shaped function.
PotentialShapedRMax(SADomain, HashableStateFactory, PotentialFunction, KWIKModel, ModelLearningPlanner) - Constructor for class burlap.behavior.singleagent.learning.modellearning.rmax.PotentialShapedRMax
Initializes for a given model, model learning planner, and potential shaped function.
PotentialShapedRMax.RMaxPotential - Class in burlap.behavior.singleagent.learning.modellearning.rmax
A potential function for vanilla RMax; all states have a potential value of R_max/(1-gamma)
potentialValue(State) - Method in class burlap.behavior.singleagent.learning.modellearning.rmax.PotentialShapedRMax.RMaxPotential
 
potentialValue(State) - Method in interface burlap.behavior.singleagent.shaping.potential.PotentialFunction
Returns the reward potential from the given state.
preferenceLength() - Method in class burlap.datastructures.BoltzmannDistribution
Returns the number of elements on which there are preferences
preferences - Variable in class burlap.behavior.singleagent.learning.actorcritic.actor.BoltzmannActor
A map from (hashed) states to Policy nodes; the latter of which contains the action preferences for each applicable action in the state.
preferences - Variable in class burlap.datastructures.BoltzmannDistribution
The preference values to turn into probabilities
prePlanPrep() - Method in class burlap.behavior.singleagent.planning.deterministic.informed.astar.AStar
 
prePlanPrep() - Method in class burlap.behavior.singleagent.planning.deterministic.informed.astar.DynamicWeightedAStar
 
prePlanPrep() - Method in class burlap.behavior.singleagent.planning.deterministic.informed.BestFirst
This method is called at the start of the BestFirst.planFromState(State) method and can be used initialize any special data structures needed by the subclass.
printDebug - Variable in class burlap.behavior.singleagent.interfaces.rlglue.RLGlueAgent
Whether to print debug statements.
printOutResults() - Method in class burlap.mdp.stochasticgames.tournament.Tournament
Prints the tournament results by agent index and their cumulative reward received in the tournament.
PrioritizedSearchNode - Class in burlap.behavior.singleagent.planning.deterministic.informed
An extension of the SearchNode class that includes a priority value.
PrioritizedSearchNode(HashableState, double) - Constructor for class burlap.behavior.singleagent.planning.deterministic.informed.PrioritizedSearchNode
Initializes a PrioritizedSearchNode for a given (hashed) input state and priority value.
PrioritizedSearchNode(HashableState, Action, SearchNode, double) - Constructor for class burlap.behavior.singleagent.planning.deterministic.informed.PrioritizedSearchNode
Constructs a SearchNode for the input state and priority and sets the generating action and back pointer to the provided elements.
PrioritizedSearchNode.PSNComparator - Class in burlap.behavior.singleagent.planning.deterministic.informed
A class for comparing the priority of two PrioritizedSearchNodes.
PrioritizedSweeping - Class in burlap.behavior.singleagent.planning.stochastic.valueiteration
An implementation of Prioritized Sweeping as DP planning algorithm as described by Li and Littman [1].
PrioritizedSweeping(SADomain, double, HashableStateFactory, double, int) - Constructor for class burlap.behavior.singleagent.planning.stochastic.valueiteration.PrioritizedSweeping
Initializes
PrioritizedSweeping.BPTR - Class in burlap.behavior.singleagent.planning.stochastic.valueiteration
A back pointer and its max action probability of transition.
PrioritizedSweeping.BPTRNode - Class in burlap.behavior.singleagent.planning.stochastic.valueiteration
A node for state thar contains a list of its back pointers, their max probability of transition to this state, and the priority of this nodes state.
PrioritizedSweeping.BPTRNodeComparator - Class in burlap.behavior.singleagent.planning.stochastic.valueiteration
Comparator for the the priority of BPTRNodes
priority - Variable in class burlap.behavior.singleagent.planning.deterministic.informed.PrioritizedSearchNode
The priority of the node used to order it for expansion.
priority - Variable in class burlap.behavior.singleagent.planning.stochastic.valueiteration.PrioritizedSweeping.BPTRNode
 
priorityCompare - Variable in class burlap.datastructures.HashIndexedHeap
A comparator to compare objects
priorityNodes - Variable in class burlap.behavior.singleagent.planning.stochastic.valueiteration.PrioritizedSweeping
The priority queue of states
probabilities(State, Action) - Method in class burlap.domain.singleagent.pomdp.tiger.TigerObservations
 
probabilities(State, Action) - Method in interface burlap.mdp.singleagent.pomdp.observations.DiscreteObservationFunction
Returns the observation probability mass/density function for all observations that have non-zero mass/density conditioned on the true MDP state and previous action taken that led to the state.
probabilitiesByEnumeration(DiscreteObservationFunction, State, Action) - Static method in class burlap.mdp.singleagent.pomdp.observations.ObservationUtilities
A helper method for easily implementing the DiscreteObservationFunction.probabilities(State, Action) method that computes observation probability distribution by enumerating all possible observations (as defined by the DiscreteObservationFunction.allObservations() method) and assigning their probability according to the ObservationFunction.probability(State, State, Action) method.
probability - Variable in class burlap.behavior.singleagent.options.model.BFSMarkovOptionModel.OptionScanNode
the *un*-discounted probability of reaching this search node
probability - Variable in class burlap.domain.singleagent.graphdefined.GraphDefinedDomain.NodeTransitionProbability
The probability of transitioning to the resulting state
probability(State, State, Action) - Method in class burlap.domain.singleagent.pomdp.tiger.TigerObservations
 
probability(State, State, Action) - Method in interface burlap.mdp.singleagent.pomdp.observations.ObservationFunction
Returns the probability that an observation will be observed conditioned on the MDP state and previous action taken that led to the state.
probabilityOfTermination(State, Episode) - Method in class burlap.behavior.singleagent.options.MacroAction
 
probabilityOfTermination(State, Episode) - Method in interface burlap.behavior.singleagent.options.Option
 
probabilityOfTermination(State, Episode) - Method in class burlap.behavior.singleagent.options.SubgoalOption
 
probs - Variable in class burlap.datastructures.BoltzmannDistribution
The output probabilities
produceRandomOffset(boolean[], double[]) - Method in class burlap.behavior.functionapproximation.sparse.tilecoding.TileCodingFeatures
Creates and returns a random tiling offset for the given widths and required dimensions.
produceUniformTilingsOffset(boolean[], double[], int, int) - Method in class burlap.behavior.functionapproximation.sparse.tilecoding.TileCodingFeatures
Creates and returns an offset that is uniformly spaced from other tilings.
projectionType - Variable in class burlap.behavior.singleagent.learnfromdemo.RewardValueProjection
 
propFunction(String) - Method in interface burlap.mdp.core.oo.OODomain
Returns the PropositionalFunction with the given name
propFunction(String) - Method in class burlap.mdp.singleagent.oo.OOSADomain
 
propFunction(String) - Method in class burlap.mdp.stochasticgames.oo.OOSGDomain
 
propFunctionMap - Variable in class burlap.mdp.singleagent.oo.OOSADomain
 
propFunctionMap - Variable in class burlap.mdp.stochasticgames.oo.OOSGDomain
 
propFunctions() - Method in interface burlap.mdp.core.oo.OODomain
Returns a list of the propositional functions that define this domain.
propFunctions() - Method in class burlap.mdp.singleagent.oo.OOSADomain
 
propFunctions() - Method in class burlap.mdp.stochasticgames.oo.OOSGDomain
 
PropositionalFunction - Class in burlap.mdp.core.oo.propositional
The propositional function class defines evaluations of object instances in an OO-MDP state and are part of the definition for an OO-MDP domain.
PropositionalFunction(String, String[]) - Constructor for class burlap.mdp.core.oo.propositional.PropositionalFunction
Initializes a propositional function with the given name and parameter object classes.
PropositionalFunction(String, String[], String[]) - Constructor for class burlap.mdp.core.oo.propositional.PropositionalFunction
Initializes a propositional function with the given name, parameter object classes, and the parameter order groups of the parameters.
propViewer - Variable in class burlap.behavior.singleagent.auxiliary.EpisodeSequenceVisualizer
 
propViewer - Variable in class burlap.behavior.stochasticgames.auxiliary.GameSequenceVisualizer
 
propViewer - Variable in class burlap.shell.visual.SGVisualExplorer
 
propViewer - Variable in class burlap.shell.visual.VisualExplorer
 
providesStateEnumerator() - Method in class burlap.mdp.singleagent.pomdp.PODomain
Indicates whether this domain has a StateEnumerator defined for it.
pSelection - Variable in class burlap.behavior.policy.support.ActionProb
The probability of the action being selected.
PSNComparator() - Constructor for class burlap.behavior.singleagent.planning.deterministic.informed.PrioritizedSearchNode.PSNComparator
 
put(int, double) - Method in interface burlap.behavior.functionapproximation.FunctionGradient
Adds the partial derivative for a given weight
put(int, double) - Method in class burlap.behavior.functionapproximation.FunctionGradient.SparseGradient
 
put(int, int) - Method in class burlap.behavior.functionapproximation.sparse.SparseCrossProductFeatures.FeaturesMap
 
put(String, int) - Method in class burlap.domain.stochasticgames.normalform.SingleStageNormalFormGame.ActionNameMap
 
putdownBlock(BlockDudeState) - Method in class burlap.domain.singleagent.blockdude.BlockDudeModel
Modifies state s to put down the block the agent is holding.
A B C D E F G H I J K L M N O P Q R S T U V W X Y Z 
Skip navigation links