Planning and acting in partially observable stochastic domains

From MaRDI portal
Publication:72343

DOI10.1016/s0004-3702(98)00023-xzbMath0908.68165OpenAlexW2168359464WikidataQ56602944 ScholiaQ56602944MaRDI QIDQ72343

Anthony R. Cassandra, Leslie Pack Kaelbling, Michael L. Littman, Leslie Pack Kaelbling, Michael L. Littman, Anthony R. Cassandra

Publication date: May 1998

Published in: Artificial Intelligence (Search for Journal in Brave)

Full work available at URL: https://doi.org/10.1016/s0004-3702(98)00023-x




Related Items (only showing first 100 items - show all)

The Concept of Opposition and Its Use in Q-Learning and Q(λ) TechniquesUnnamed ItemLarge-scale financial planning via a partially observable stochastic dual dynamic programming frameworkLearning-based state estimation and control using MHE and MPC schemes with imperfect modelsReward prediction errors, not sensory prediction errors, play a major role in model selection in human reinforcement learningApproximability and efficient algorithms for constrained fixed-horizon POMDPs with durative actionsA conflict-directed approach to chance-constrained mixed logical linear programmingRisk-aware shielding of partially observable Monte Carlo planning policiesSimultaneous perception-action design via invariant finite belief setsBridging Commonsense Reasoning and Probabilistic Planning via a Probabilistic Action LanguageReward Maximization Through Discrete Active InferencePOMDP controllers with optimal budgetEpistemic uncertainty aware semantic localization and mapping for inference and belief space planningA Markovian model for the spread of the SARS-CoV-2 virusBehavioral model summarisation for other agents under uncertaintyReward tampering problems and solutions in reinforcement learning: a causal influence diagram perspectiveOff-policy evaluation in partially observed Markov decision processes under sequential ignorabilityUnnamed ItemUsing Machine Learning for Decreasing State Uncertainty in PlanningUnnamed ItemUnnamed ItemMinimax real-time heuristic searchUnnamed ItemModel-Based Reinforcement Learning for Partially Observable Games with Sampling-Based State EstimationUnnamed ItemA reinforcement learning scheme for a partially-observable multi-agent gameA reinforcement learning scheme for a partially-observable multi-agent gameRepresentation and Timing in Theories of the Dopamine SystemGeneral Value Function NetworksA Sufficient Statistic for Influence in Structured Multiagent EnvironmentsConstrained Multiagent Markov Decision Processes: a Taxonomy of Problems and AlgorithmsInduction and Exploitation of Subgoal Automata for Reinforcement LearningStrategy Graphs for Influence DiagramsTask-Aware Verifiable RNN-Based Policies for Partially Observable Markov Decision ProcessesAnalyzing generalized planning under nondeterminismAn evidential approach to SLAM, path planning, and active explorationMeeting a deadline: shortest paths on stochastic directed acyclic graphs with information gatheringGoal-directed learning of features and forward modelsEnforcing almost-sure reachability in POMDPsFinite-horizon LQR controller for partially-observed Boolean dynamical systemsMyopic Bounds for Optimal Policy of POMDPs: An Extension of Lovejoy’s Structural ResultsOptimal speech motor control and token-to-token variability: a Bayesian modeling approachDynamic multiagent probabilistic inferenceAutonomous agents modelling other agents: a comprehensive survey and open problemsA two-state partially observable Markov decision process with three actionsGradient-descent for randomized controllers under partial observabilityA synthesis of automated planning and reinforcement learning for efficient, robust decision-makingLearning Where to Attend with Deep Architectures for Image TrackingRecognizing and learning models of social exchange strategies for the regulation of social interactions in open agent societiesActive inference and agency: optimal control without cost functionsTutorial series on brain-inspired computing. IV: Reinforcement learning: machine learning and natural learningOpen problems in universal induction \& intelligenceCost-sensitive feature acquisition and classificationPlanning for multiple measurement channels in a continuous-state POMDPSimultaneous learning and planning in a hierarchical control system for a cognitive agentPartially observable multistage stochastic programmingMulti-stage classifier designLearning to steer nonlinear interior-point methodsRepresentations for robot knowledge in the \textsc{KnowRob} frameworkRobotic manipulation of multiple objects as a POMDPGeometric backtracking for combined task and motion planning in robotic systemsSupervisor synthesis of POMDP via automata learningExact decomposition approaches for Markov decision processes: a surveyTesting probabilistic equivalence through reinforcement learningBOUNDED-PARAMETER PARTIALLY OBSERVABLE MARKOV DECISION PROCESSES: FRAMEWORK AND ALGORITHMPlanning in partially-observable switching-mode continuous domainsOptimal management of stochastic invasion in a metapopulation with Allee effectsQuantitative controller synthesis for consumption Markov decision processesRationalizing predictions by adversarial information calibrationUnnamed ItemProbabilistic may/must testing: retaining probabilities by restricted schedulersAffect control processes: intelligent affective interaction using a partially observable Markov decision processDynamic optimization over infinite-time horizon: web-building strategy in an orb-weaving spider as a case studyMulti-goal motion planning using traveling salesman problem in belief spacePrivacy stochastic games in distributed constraint reasoningPolicy iteration for bounded-parameter POMDPsProbabilistic reasoning about epistemic action narrativesReasoning and predicting POMDP planning complexity via covering numbersComputing rank dependent utility in graphical models for sequential decision problemsExploiting symmetries for single- and multi-agent partially observable stochastic domainsDecentralized MDPs with sparse interactionsState observation accuracy and finite-memory policy performanceAn online multi-agent co-operative learning algorithm in POMDPsMarkov limid processes for representing and solving renewal problemsComputation of weighted sums of rewards for concurrent MDPsStrong planning under partial observabilitySolving for Best Responses and Equilibria in Extensive-Form Games with Reinforcement Learning MethodsAn affective mobile robot educator with a full-time jobMarkov decision processes with sequential sensor measurementsCounterexample-guided inductive synthesis for probabilistic systemsAlgorithms and conditional lower bounds for planning problemsA survey of inverse reinforcement learning: challenges, methods and progressAn integrated approach to solving influence diagrams and finite-horizon partially observable decision processesTask-structured probabilistic I/O automataPosterior Weighted Reinforcement Learning with State UncertaintyIntegration of Reinforcement Learning and Optimal Decision-Making Theories of the Basal GangliaThe value of information for populations in varying environmentspomdpSolveConformant plans and beyond: principles and complexityPerformance prediction of an unmanned airborne vehicle multi-agent system


Uses Software


Cites Work


This page was built for publication: Planning and acting in partially observable stochastic domains