Corpus ID: 14203192

Probabilistic State-Dependent Grammars for Plan Recognition

@inproceedings{Pynadath2000ProbabilisticSG,
  title={Probabilistic State-Dependent Grammars for Plan Recognition},
  author={David V. Pynadath and Michael P. Wellman},
  booktitle={UAI},
  year={2000}
}
Techniques for plan recognition under uncertainty require a stochastic model of the plangeneration process. We introduce probabilistic state-dependent grammars (PSDGs) to represent an agent's plan-generation process. The PSDG language model extends probabilistic contextfree grammars (PCFGs) by allowing production probabilities to depend on an explicit model of the planning agent's internal and external state. Given a PSDG description of the plan-generation process, we can then use inference… Expand
A probabilistic plan recognition algorithm based on plan tree grammars
TLDR
It is shown that this clarifies several difficult issues in plan recognition including the execution of multiple interleaved root goals, partially ordered plans, and failing to observe actions. Expand
Controlling the Hypothesis Space in Probabilistic Plan Recognition
TLDR
A heuristic weighted model counting algorithm is presented that limits the number of generated plan execution models in order to recognize goals quickly by computing their lower and upper bound likelihoods. Expand
Abstract Hidden Markov Models for Online Probabilistic Plan Recognition
TLDR
The structure of the stochastic model rep- resenting the execution of the general AMP is analyzed and an efficient hybrid Rao-Blackwellised sampling method for policy recognition that scales well with the number of levels in the plan hierarchy is provided. Expand
SLIM: Semi-Lazy Inference Mechanism for Plan Recognition
TLDR
A new and efficient algorithm for online plan recognition called SLIM (Semi-Lazy Inference Mechanism), which combines both a bottom-up and top-down parsing processes, which allow it to commit only to the minimum necessary actions in real-time, but still provide complete hypotheses post factum. Expand
Chapter 3 – Plan Recognition Using Statistical–Relational Models
TLDR
This chapter introduces two new approaches to abductive plan recognition using Bayesian logic programs and Markov Logic Networks, and presents an extensive evaluation of these approaches on three benchmark datasets on plan recognition, comparing them with existing state-of-the-art methods. Expand
Extending Bayesian Logic Programs for Plan Recognition and Machine Reading
TLDR
This work develops an approach to abductive plan recognition using BLPs and extends BLPs to use logical abduction to construct Bayesian networks and calls the resulting model Bayesian Abductive Logic Programs (BALPs), demonstrating that BALPs outperform the existing state-of-art methods like Markov Logic Networks (MLNs). Expand
Recognizing Plans with Loops Represented in a Lexicalized Grammar
TLDR
It is shown how the loop-handling methods from formal grammars can be extended to the more general plan recognition problem and provide a method for encoding loops in an existing plan recognition system that scales linearly in the number of loop iterations. Expand
Plan Recognition in Continuous Domains
TLDR
This work provides a formalization of recognition problems which admits continuous environments, as well as discrete domains, and shows that through mirroring— generalizing plan-recognition by planning—the authors can apply continuous-world motion planners in plan recognition. Expand
Considering State in Plan Recognition with Lexicalized Grammars
  • C. Geib
  • Computer Science
  • CogRob@AAAI
  • 2012
TLDR
This paper documents extending the ELEXIR (Engine for LEXicalized Intent Recognition) system with a world model and allows a number of additions to the algorithm, most significantly conditioning probabilities for recognized plans on the state of the world during execution. Expand
Automatic Generation of Plan Libraries for Plan Recognition Performance Evaluation
TLDR
A mechanism to automatically generate arbitrarily complex plan libraries is developed, such plan library generation can be directed through a number of parameters to allow for systematic experimentation. Expand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 28 REFERENCES
Probabilistic grammars for plan recognition
TLDR
New inference algorithms that generate a Bayesian network representation of the PCFG distribution and specialized algorithms that exploit the particular independence properties of the PSDG language to maintain a more efficient summary of evidence in the form of a belief state are presented. Expand
A New Model of Plan Recognition
TLDR
A new abductive, probabilistic theory of plan recognition that accounts for phenomena omitted from most previous plan recognition theories: notably the cumulative effect of a sequence of observations of partially-ordered, interleaved plans and the effect of context on plan adoption. Expand
A Bayesian Model of Plan Recognition
Abstract We argue that the problem of plan recognition, inferring an agent's plan from observations, is largely a problem of inference under conditions of uncertainty. We present an approach to theExpand
Accounting for Context in Plan Recognition, with Application to Traffic Monitoring
TLDR
A general Bayesian framework encompassing this view is presented, and it is shown how context can be exploited in plan recognition on a problem in traffic monitoring. Expand
Generalized Queries on Probabilistic Context-Free Grammars
TLDR
An algorithm for constructing Bayesian networks fromPCFGs is presented, and it is shown how queries or patterns of queries on the network correspond to interesting queries on PCFGs. Expand
Towards History-based Grammars: Using Richer Models for Probabilistic Parsing
TLDR
HBG incorporates lexical, syntactic, semantic, and structural information from the parse tree into the disambiguation process in a novel way and significantly outperforms P-CFG, increasing the parsing accuracy rate from 60% to 75%, a 37% reduction in error. Expand
A Message Passing Algorithm for Plan Recognition
TLDR
A general framework for plan recognition whose formulation is motivated by a general purpose algorithm for effective abduction, and the class of plans for which the method is applicable is wider than those previously proposed, as both recursive and optional plan components can be represented. Expand
Effective Bayesian Inference for Stochastic Programs
TLDR
This paper proposes a stochastic version of a general purpose functional programming language that contains random choices, conditional statements, structured values, defined functions, and recursion, and provides an exact algorithm for computing conditional probabilities of the form Pr(P(x) | Q(x)) where x is chosen randomly from this distribution. Expand
Simulation-based inference for plan monitoring
TLDR
This work investigates and extend simulation algorithms for approximate inference on Bayesian networks and proposes a new algorithm, called Rewind/Replay, for generating a set of simulations weighted by their likelihood given past observations. Expand
Generalized Plan Recognition
TLDR
A new theory of plan recognition that is significantly more powerful than previous approaches and employs circumscription to transform a first-order theory of action into an action taxonomy, which can be used to logically deduce the complex action(s) an agent is performing. Expand
...
1
2
3
...