Next Article in Journal
Brain Network Modeling Based on Mutual Information and Graph Theory for Predicting the Connection Mechanism in the Progression of Alzheimer’s Disease
Next Article in Special Issue
Where Was Past Low-Entropy?
Previous Article in Journal
Guessing with Distributed Encoders
Previous Article in Special Issue
From an Entropic Measure of Time to Laws of Motion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Goal Identification Control Using an Information Entropy-Based Goal Uncertainty Metric †

College of System Engineering, National University of Defense Technology, Changsha 410000, China
*
Author to whom correspondence should be addressed.
This paper is an extended version of the conference materials in the 6th Goal Reasoning Workshop at IJCAI/FAIM-2018.
Entropy 2019, 21(3), 299; https://doi.org/10.3390/e21030299
Submission received: 14 February 2019 / Revised: 11 March 2019 / Accepted: 18 March 2019 / Published: 20 March 2019
(This article belongs to the Special Issue Entropy Production and Its Applications: From Cosmology to Biology)

Abstract

:
Recent research has found situations where the identification of agent goals could be purposefully controlled, either by changing the underlying environment to make it easier, or exploiting it during agent planning to delay the opponent’s goal recognition. The paper tries to answer the following questions: what kinds of actions contain less information and more uncertainty about the agent’s real goal, and how to describe this uncertainty; what is the best way to control the process of goal identification. Our contribution is the introduction of a new measure we call relative goal uncertainty (rgu) with which we assess the goal-related information that each action contains. The rgu is a relative value associated with each action and represents the goal uncertainty quantified by information entropy after the action is taken compared to other executable ones in each state. After that, we show how goal vagueness could be controlled either for one side or for both confronting sides, and formulate this goal identification control problem as a mixed-integer programming problem. Empirical evaluation shows the effectiveness of the proposed solution in controlling goal identification process.

1. Introduction

Goal recognition—the ability to recognize the plans and goals of other agents—enables humans, AI agents, or command and control systems to reason about what the others are doing, why they are doing it, and what they will do next [1]. Until now, goal recognition system has worked well in many applications such as human–robot interaction [2], intelligent tutoring [3], system-intrusion detection [4] and security applications [5].
Though this technique has been successfully applied to many application domains, new problems arise when goal recognition encounters a goal uncertainty discovered in certain environmental settings. Figure 1 offers a simple example that is first mentioned in [6] and will help to clarify the concepts of our work, which is extended from our original [7].
The model consists of a simple room (or airport) with a single entry point, marked as ‘Start’ and two possible exit points (boarding gates), marked as ‘Goal 1’ (domestic flights) and ‘Goal 2’ (international flights). An agent can move vertically or horizontally from ‘Start’ to one of the goals. Notice that in this model, the goal of the agent becomes clear once turning left or right, while moving vertically would impose goal uncertainty on goal recognizers and postpone the goal identification. Therefore, in the worst case an optimal agent can move up 4 steps before it is obliged to turn towards its goal.
The goal uncertainty showed in the above example could be viewed as an inherent property of one particular goal recognition task, and pose a new challenge to effective goal reasoning. The problem finds itself relevant in many of the similar applications of goal recognition tasks where goal uncertainty exists. Also, it should be noted that apart from the friendly situation in Figure 1 and works in [6], the goal uncertainty could further be used as a potential backdoor for the adversary to delay goal identification.
Therefore, this paper focuses on effective control of the goal identification process. As a first stage of our exploration, we assume agents are optimal and that the actions of the agent are fully observable and deterministic. To achieve our objective, we introduce a new concept called relative goal uncertainty (rgu), with which we assess the goal-related information that each action contains. The rgu is a relative value associated with each action that agent would take and represents the goal uncertainty quantified by information entropy after the action is taken compared to other executable ones in each state.
The rgu value associated with each action helps in assessing the uncertainty that exists in goal recognition task, and our goal is to provide methods for agents to control it. Towards this end, we define two optimization models based on mixed-integer programming, one of which reduces the goal uncertainty through limiting the set of available actions an agent can perform, the other one delays goal identification through balancing the reward and goal uncertainty during mission planning.
Goal identification control, while relevant to goal recognition [1] or goal reasoning, is a different task. While goal recognition aims at discovering the goals of an agent according to observations, goal identification control rather offers an offline and online combined solution for assessing and controlling the goal uncertainty to assure the goal of any optimal agent in the system is recognized. Also, different from goal recognition design problem [6,8,9] which purely focuses on facilitating goal recognition offline, goal identification control provides a more compact solution for agents to control the goal uncertainty and allows this additional information to be incorporated into ongoing missions.
Finally, considering the situation where both confronting sides are trying to control the goal identification process (i.e., both sides are behaving in a strategic game-theoretic manner), the paper then proposes the Shortest-Path Largest-Uncertainty Network Interdiction model (SPLUNI), which could be viewed as a bilevel mixed-integer program.
The SPLUNI model is transformed into its dual form using KKT conditions and computed using mixed-integer programming method.
The paper is organized as follows. We start by providing the necessary background on probabilistic goal recognition. We continue by introducing the formal model representing the goal identification control problem, and the rgu value. The following sections present the methods we developed for calculating rgu and controlling the goal uncertainty of a given goal identification control problem. We conclude with an empirical evaluation, a discussion of related work, and a conclusion.

2. Background and Related Work

2.1. Probabilistic Goal Recognition

The goal recognition problem has been formulated and addressed in many ways, as a graph covering problem upon a plan graph [10], a parsing problem over grammar [11,12,13,14], a deductive and probabilistic inference task over a static or Dynamic Bayesian Network [15,16,17,18,19,20] and an inverse planning problem over planning models [21,22,23,24,25].
Among the approaches viewing the goal or plan recognition as an uncertainty problem, two formulations appear and solve the problem from different perspectives. One focuses on constructing a suitable library of plans or policies [16,17], while the other one takes the domain theory as an input and use planning algorithms to generate problem solutions [21,23]. Before reviewing these two formulations, we would first give the definition of probabilistic goal recognition (or plan recognition) as follows:
Definition 1.
A probabilistic goal recognition problem or theory is a tuple T = D , G , O , P r o b where D is the problem domain, G is the set of possible goals G, O = o 1 , , o m is an observation sequence (which may or may not be adjacent) with each o i being an action, and P r o b represents the prior probabilities over G .
And generally, the posterior goal distribution P ( G | O ) will be computable from the Bayes Rule as:
P ( G | O ) = α P ( O | G ) P ( G ) ,
where α is a normalizing constant, and P ( G ) is P r o b ( G ) .
The uncertainty problem is first addressed by the work [16] which phrases the plan recognition problem as the inference problem in a Bayesian network representing the process of executing the actor’s plan. It is followed by more work considering dynamic models for performing plan recognition online [12,13,26,27]. While this offers a coherent way of modelling and dealing with various sources of uncertainty in the plan execution model, the computational complexity and scalability of inference is the main issue, especially for dynamic models. In [17], Bui et al. proposed a framework for online probabilistic plan recognition based on the Abstract Hidden Markov Models (AHMM), which is a stochastic model for representing the execution of a hierarchy of contingent plans (termed policies). Scalability in policy recognition in the AHMM, and the inference upon its transformed Dynamic Bayesian Network (DBN), is achieved by using an approximate inference scheme known as the Rao-Blackwellized Particle Filter (RBPF) [28]. Generally, in a particle system, the posterior distribution is empirically represented using a weighted sum of N p samples g i drawn independent identically distributed from the proposal distribution [29]:
p ( g | O ) i = 1 N p W i δ ( g g i )
where the importance weights W i should be updated recursively.
Within this formulation, following research continues to improve the model’s expressiveness and computational attractiveness, as in [18,20,30,31,32,33,34,35,36]. Notably, research in [33,36] proposes a CSSMs model which allow simple system model definition in the form of planning problem and can reason in large state spaces by using approximate inference. Also, recently machine learning methods including reinforcement learning [35], deep learning [3,37] and inverse reinforcement learning [38,39] have already been successfully applied in learning the agents’ decision models for goal recognition tasks. These efforts once again extend the usability of policy-based probabilistic goal recognition methods by constructing agents’ behavior models from real data.
Recently, the work [21,22] shows that plan recognition can be formulated and solved using off-the-shelf planners, follows by work considering suboptimality [23] and partial observability [24]. Not working over libraries but over the domain theory where a set G of possible goals is given, this generative approach solves the probabilistic goal recognition efficiently, provided that the probability of goals is defined in terms of the cost difference of achieving the goal under two conditions: complying with the observations, and not complying with them.
costdif R G ( s , g , O ) = optc ( s , O , g ) optc ( s , O , g )
By comparing cost differences across goals, a probability distribution is generated that conforms to the intuition: the lower the cost difference, the higher the probability. Concretely, they propose the assumption of a Boltzmann distribution and take P ( O | G ) = sigmoid ( β ( optc ( s , G , O ) optc ( s , G , O ) ) ) , thus arriving at the following posterior:
P ( G | O ) = α e β X 1 + e β X ,
where X = costdif R G ( s , G , O ) , α is a normalizing constant across all goals, and β a positive constant which captures a ‘soft rationality’ assumption.
Several works followed this approach [25,40,41,42,43,44] by using various automated planning techniques to analyze and solve goal or plan recognition problems. Its basic ideas have also been applied in new problems such as Goal Recognition Design [6], Deceptive Path-Planning [45], etc. The advantages of the latter formulation are two-fold: one is that by using plenty of off-the-shelf model-based planners the approach scales up well, handling domains with hundreds of actions and fluents quite efficiently; the other one lies in the fact that the model-based method has no concerns about the recognition of joint plans for achieving goal conjunctions. Joint plans come naturally in the generative approach from conjunctive goals, but harder to handle in library-based methods.
According to the relationships between the observer and the observed agent, the goal or plan recognition could be further divided into keyhole, intended and adversarial types [4,46,47,48]. In keyhole plan recognition, the observed agent is indifferent to the fact that its plans are being observed and interpreted. The presence of a recognizer who is watching the activity of the planning agent does not affect the way he plans and acts [26,49]. Intended recognition arises, for example, in cooperative problem-solving and in understanding indirect speech acts. In these cases, recognizing the intentions of the agent allows us to help or respond appropriately [46]. In adversarial recognition, the observed agent is hostile to the observation of his actions and attempts to thwart the recognition [50,51].

2.2. Goal Recognition Design

Goal recognition design was first introduced in [6] to reduce goal uncertainty and advance the correct recognition through redesigning the underlying environment. It could be seen as an inverse problem to deceptive path-planning. Standing on the side of the observer, the GRD problem tries to reduce goal uncertainty and advance the correct recognition through redesigning the domain layout. To do so, they introduce a concept named worst-case distinctiveness (wcd), measuring the maximal length of a prefix of a plan an agent may take within a domain before its real goal has been revealed. As in their first case [6] shown in Figure 2, the goal of the agent becomes clear once turning left or right, while it maintains the longest ambiguity if the agent moves straight up 4 steps ( w c d = 4 ) before it is obliged to turn towards its real goal. Thus, the blockade of the action moving the agent from C 1 to C 2 (Figure 2b) successfully reduces the wcd from 4 to 0, and prohibits the agent from hiding key information.
Since then, lots of research has been carried out. The work described in [6,52] extends the previous work by accounting for agents that behave non-optimally and non-observably. The work in [53] further allows the outcomes of agents’ actions to be non-deterministic, and proposes a Stochastic GRD problem. Apart from the relaxation of assumptions, researchers try to solve the GRD from different aspects, or use newly established metrics. Answer Set Programming [8] was first used to address the same problem, resulting in higher scalability and efficiency. Mirsky et al. [54] extend GRD to the Plan Recognition Design (PRD), which is the task of designing a domain using plan libraries to facilitate fast identification of an agent’s plan. Noticing the inconsistency between the original wcd (worst-case distinctiveness) and Stochastic GRD model, a new metric, namely all-goals wcd ( w c d a g ), is proposed by [9]. Also, based on new assumption of the availability of prior information about the agent’s true goal, they further propose a expected-case distinctiveness (ecd) metric that weighs the possible goals based on their likelihoods. Mirsky et al. [55] further study the goal and PRD for plan libraries.
The general GRD problem [6] is represented as a tuple P = D , G , where D = S , s 0 , A , f , C captures the domain information and G is a set of possible goal states of the agent. Except that all actions have the same cost of 1, the elements in the tuple D are as the definitions in classical planning where S is a finite and discrete state space; s 0 is the start state of the agent; A is the set of actions; f : S × A S is a deterministic state transition function; and G is a set of goal states. The wcd of problem P is the length of a longest sequence of actions π = a 1 , , a k that is the prefix in cost-minimal plans π g 1 and π g 2 to distinct goals g 1 , g 2 G . GRD’s objective is to find a subset of actions A ^ A such that if they are removed from the set of actions A, then the wcd of the resulting problem is minimized. As summarized in [53], the GRD task could be formulated into an optimization problem which is subject to the constraint that the cost of cost-minimal plans to achieve each goal g G is the same before and after removing the subset of actions:
A ^ = arg min A ^ A w c d ( P ^ )
subject to C ( π g ) = C ( π ^ g ) g G
where P ^ = D ^ , G is the problem with the resulting domain D ^ = S , s 0 , A \ A ^ , f , C after removing actions A ^ , π g is a cost-minimal plan to achieve goal g in the original problem P, and π ^ g is a cost-minimal plan to achieve goal g in P ^ .

3. Probabilistic Goal Recognition

The authors in [20] shows a simple yet effective probabilistic goal recognizer based on Markov Decision Process (MDP). The probabilistic goal recognition model is defined by a tuple D = S , s 0 , A , f , G , e , O , where S is a finite and discrete state space; s 0 is the start state of the agent; A is the set of actions; f : S × A S is a deterministic state transition function; G is a set of goal states; e is the goal termination variable and O is a non-empty observation set. Essentially, the model is a DBN, in which all causalities could be depicted. We introduce a full DBN structure depicting two time slices is presented in Figure 3.
The behaviors of system evolution are described using functions or parameters.
  • state transition function T : S × A × S [ 0 , 1 ] is P s τ = p ( s τ | s τ 1 , a τ ) ,
  • observation function S × O [ 0 , 1 ] is P o τ = p ( o τ | s τ ) ,
  • agent action policy P a τ = p ( a τ | s τ 1 , g τ ) ,
  • goal transition probability P g τ = p ( g τ | e τ 1 , g τ 1 ) ,
  • goal termination probability P e τ = p ( e τ | g τ , s τ ) .
Recognizing the evader’s goal is an inference problem trying to find the real goal behind agent actions based on observations online. In essence, the task is to compute the posterior distribution P ( g τ | o τ ) of goal g τ given observation o τ . This could be achieved either by accurate inference or by approximate methods. Accurate inference, however, is not scalable when state space of the domain problem becomes large, nor can it tackle partially missing or noisy data. Widely applied in sequential state estimation, particle filter is a kind of approximate inference methods designed to handle non-Gaussian and nonlinear problems [29]. As for the problem with a large-scale state space, methods such as [28,56] are preferred. In this work, we would like to simply show how the probabilistic goal recognition works, and thus the basic particle filter would be enough. In this work, the MDP or agent action model is assumed to be known by both the evader and the indicator, except for the current goal g τ of the evader. Instead, the set of possible goals is given along with the priors P ( G ) . Similar assumptions also exist in [24] in which the posterior goal probabilities P ( G | O ) is obtained from Bayes Rule
P ( G | O ) = α P ( O | G ) P ( G )
where α is a normalizing constant. In particle filter however, a posterior distribution is empirically represented using a weighted sum of N p samples [29] drawn from the proposal distribution:
p ( g τ | o τ ) i = 1 N p W τ ( i ) δ ( g τ g τ ( i ) )
where g τ ( i ) are assumed to be i.i.d drawn from q ( g τ | o i ) . The importance weights W τ ( i ) should be updated recursively
W τ ( i ) W τ 1 ( i ) p ( o τ | g τ ( i ) ) p ( g τ ( i ) | g τ 1 ( i ) ) q ( g τ ( i ) | g 0 : τ 1 ( i ) , o τ )
As we use simplest sampling, the q ( g τ ( i ) | g 0 : τ 1 ( i ) , o τ ) is set to be p ( g τ ( i ) | g τ 1 ( i ) ) , which could be computed directly using the agent action model:
p ( g τ ( i ) | g τ 1 ( i ) ) = a τ 1 ( i ) s τ 1 ( i ) e τ 1 ( i ) p g τ ( i ) p e τ 1 ( i ) p s τ 1 ( i ) p a τ 1 ( i )
Thus, the g τ in Equation (7) would be sampled from p ( g τ ( i ) | g τ 1 ( i ) ) . As the observation o τ only depends on s τ , the importance weights W τ ( i ) can be updated by
W τ ( i ) = W τ 1 ( i ) · p ( o τ | s τ ( i ) ) .
In this work, a sequential importance sampling (SIS) filter with resampling is used in evader’s goal inference. Notably, in the resampling process, we apply the estimated effective sample size, N ^ e f f , according to
N ^ e f f = 1 i = 1 N p W ˜ t ( i ) 2
where W ˜ t ( i ) is the normalized weight. The resampling process returns if N ^ e f f > N T , where N T is the pre-defined threshold which is set to N p / 3 , otherwise generates a new particle set by resampling with replacement N p times from the previous set with probabilities W ˜ t ( i ) , and then reset the weights to 1 / N p .

4. Relative Goal Uncertainty

Given the set of possible goal states G and the current observation o τ , the probability distribution over possible goals P ( G | o τ ) actually reflects the goal uncertainty the agent will encounter after selecting actions A ( s τ ) and reaches the next state o τ (assuming fully observable and deterministic) according to f. This enlightens us to use the goal inference information one step ahead to measure the goal uncertainty associated with available actions at each state, specifically the adjacent edges directing out of each node in the road network example.
Defined over each possible action a i A ( s ) at state s, the metric rgu is a relative value describing the overall uncertainty performance of state s before and after the action a’s interdiction. Intuitively, this could be done by summing up the information entropy of the pre-computed, next-step goal probability which are reasoned about in the goal recognition problem D = S , s i , A , f , G , e , O = { s i , s j } , where s i S and s j F S ( s i ) , the set of states that could arrive at according to f ( s i ) .
Definition 2.
(Entropy-based rgu) The Entropy-based rgu ( r g u E ) over a goal identification control problem is defined as:
r g u E ( s , a i ) = a j A ( s ) H ( G | a j , s ) | A ( s ) |
where a i A ( s ) , a j A ( s ) , and A ( s ) = A ( s ) \ a i . The entropy H ( G | a j , s ) computes the goal uncertainty contained in the reasoning result, and H ( G | a j , s ) = p ( G | a j , s ) · log p ( G | a j , s ) . Technically, we also need to remove zero entries of p ( G | a j , s ) , followed by normalization guaranteeing that s u m ( p ) = 1 .
The r g u E quantified by entropy H evaluates the goal uncertainty in a natural and compact way. More uncertainty introduced after the agent takes its action gives less information to goal recognizer and postpones the goal identification process. As the example in Figure 1 and assuming the agents are fully optimal, the r g u E for three actions leading to states ( 2 , C ) , ( 1 , B ) and ( 1 , D ) are 0 . 32 , 0 . 00 and 0 . 00 respectively. Also, the definition of r g u E could easily be applied to tasks where | G | > 2 . It should be noted that the definition of r g u E makes sure that this metric can be naturally applied to the situation where the agent changes its goal during the midway.
To illustrate how r g u E would be used in goal identification control problem and for clarity, we compute the values upon a 3 × 3 grid network (Figure 4a) with nodes representing grids and dashed edges connecting nodes. Selecting absorbing nodes No.1 and No.3 as possible goals, we compute r g u E for each action depicted as dashed line, as shown in Figure 4a. Higher r g u E means greater uncertainty associated with that action. Here we discuss two path-planning tasks T and T , where the agent has different starting points ( s 0 = 9 or s 0 = 8 ). The tasks are defined as T = S , A , f , C , G = { 1 , 3 } , s 0 , with C = C + r g u E to model the road barrier, patrolling police or other methods the goal recognizer could use in order to control the goal identification process.
For the first task, the agent would choose h = 9 , 8 , 7 , 4 , 1 for goal No.1 and h = 9 , 6 , 3 for goal No.2. While for the second task, using the definition of r g u E , the agent would choose one of the two optimal routes ( h = 8 , 7 , 4 , 1 or h = 8 , 5 , 4 , 1 ), other than h = 8 , 5 , 2 , 1 for goal No.1, and h = 8 , 9 , 6 , 3 or h = 8 , 5 , 6 , 3 other than h = 8 , 5 , 2 , 3 for goal No.3. For the latter case, we find that r g u E cannot help the agent capture the fact in goal identification control that early interdiction would be more important than a later one.
To solve this problem, we define a discounted r g u d i s so as to reduce the r g u value along with the increase of the number of steps that agent has taken from the start. For agent changing its goal during the midway, we could recompute this value from the original r g u E . Case 3.1 shown in Section 4.1, which is talked about in the next section, gives a good indication of this situation occurrence.
Definition 3. (Discountedrgu rgu)
The discounted rgu ( r g u d i s ) of a goal identification control problem is defined as:
r g u d i s ( s , a i ) = r g u ( s , a i ) × β d
where β is the discount factor and d is the number of steps that agent has taken from the start.
With the discount factor equals to 0.8 , Figure 4b shows the discounted r g u d i s values which have been adjusted according to its importance to the goal identification control task.

5. Goal Identification Control

In this section, we show how to control the r g u using the optimization technique. Instead of requiring the cost of cost-minimal plans to achieve each goal g G be the same before and after removing the subset of actions, as researched about in goal recognition design problem [6], we are more interested in deliberate changing goal uncertainty under adversarial settings where the resource constraint compared to optimality conservation is more applicable. Also, the original way of removing actions permanently is changed “softer” by adding additional costs to actions.

5.1. Reduction and Improvement of rgu

The goal identification control models based on optimization techniques are introduced. We first present models which are individually used for reducing and improving goal uncertainty for the different sides of the goal recognition task. Then we talk about the applicability of the offline-computed r g u in the online goal identification control.
The models could be transformed into the problem of maximizing (minimizing) the expectation of “s–t” path length, with the r g u being proportionally added to the original length. The mathematical-programming formulation of our new goal identification control model is as follows:
Problem : Maximize ( Minimize ) the expectation s t   path length in a directed network by interdicting arcs , Indices : i N , nodes in   G   ( s   is the current source node , t   is the terminus ) , k = ( i , j ) A , arcs in   G , k F S ( i ) ( k R S ( i ) ) , arcs directed out of   ( into ) node i , Data : c k = 1 , normal length of arc   k   ( vector form   c ) , d k = 1 , added integer delay if arc   k   is interdicted   ( vector form   d ) , r g u k , relative goal uncertainty of arc   k   ( vector form   rgu ) r k = 1 , resource required to interdict arc   k   ( vector form   r ) , R , total amount of interdiction resource , Variables : x k = 1   if arc   k   is interdicted by the interdictor ; else x k = 0 , y k = 1   if arc   k   is traversed by the evader ; else y k = 0
The formulation of rgu reduction problem ([RGUR-P]) for the observer is:
[ RGUR P ] max x X k A ( c k + x k d k ( 1 + r g u k ) ) y k
k F S ( i ) y k k R S ( i ) y k = 1 for i = s 0 i N \ { s , t } 1 i = t
x k , y k { 0 , 1 } , k A
where X = { x { 0 , 1 } | A | | r T x R } , Equation (14) is the flow-balance constraint.
While the formulation of rgu improvement problem ([RGUI-P]) for the observed agent is:
[ RGUI P ] min y k A ( c k + x k d k ) y k 1 + r g u k
k F S ( i ) y k k R S ( i ) y k = 1 for i = s 0 i N \ { s , t } 1 i = t
x k , y k { 0 , 1 } , k A
where X = { x { 0 , 1 } | A | | r T x R } . Thus, with additional goal uncertainty information, the problem could be transformed into network interdiction models, which are of a typical mixed-integer program (MIP) and solved using linear programming algorithms.
Figure 5 shows two moving traces of the observed agent under [RGUR-P] and [RGUI-P] problems. The observer could use [RGUR-P] to reduce the enemy’s goal uncertainty and accelerate situation awareness, while the observed agent, after analyzing the observer’s goal recognition task, could actually use [RGUI-P] to cover its real intention to the maximum extent. General results would be given in Section 6.
Please note that until now, the metric r g u used for assessing goal uncertainty is defined over actions available at each state in an offline manner, i.e., p ( G | a i , s ) are computed according to P ( G | O ) = α P ( O | G ) P ( G ) , with prior probability P ( G ) following Uniform distribution, instead of over agent’s full course of actions where the real-time P ( G ) contains history information, this may incur uncertainty inconsistency among individual states and the whole task. Fortunately, our definition of rgu has no effect on problem-solving of the goal identification control,
Case 1: If the real-time prior P ( G ) cannot help observer distinguish the right goal, meaning that H ( P ( G ) ) is relatively high. Then it is naturally established for agents choosing actions with either high or low uncertainty.
Case 2: If P ( G ) distinguishes the goals clearly where H ( P ( G ) ) is relatively low, and agent chooses the action that has high uncertainty, i.e., r g u 0 , then highly uncertain action would input little information to the original beliefs about agent goals. The agent’s belief of the right goal would still maintain. However, in this case, resource needed for action interdiction would be wasted as little information would be given and agent’s intention usually maintains for a period.
Case 3: Also, if P ( G ) distinguishes the goals clearly, e.g., the prior P ( g ) for goal g with the largest value, while agent chooses the action that has low uncertainty rgu, then two cases should be considered. Firstly, agent’s belief of the right goal has changed according to the posterior P ( G | o ) (Case 3.1). This happens when the agent is changing its goal from g to g during the midway, and thus chooses the action that will lead to g . Secondly, the computed posterior P ( G | o ) still capture the right goal g (Case 3.2). For both cases, the rgu works properly.

5.2. The SPLUNI Model and Its Dual Form

After individually talk about rgu reduction and improvement, in this section we propose a SPLUNI model, where uncertainty becomes another optimization objective along with path length. The SPLUNI model could be viewed as a bilevel mixed-integer program (BLMIP).
The problem description of SPLUNI is similar to [RGUR-P] and [RGUI-P], whereas it is described as maximizing the expectation of the shortest s t path length while minimizing the largest uncertainty and 0 c k , d k , r k < . Its mathematical-programming formulation is given as:
[ SPLUNI P ] max x X min y k A ( c k + x k d k ( 1 + α × r g u k ) ) y k 1 + β × r g u k
with the same constraints as Equations (14) and (15) and r T x R . α , β are parameters controlling the importance of goal uncertainty between two objectives.
As a BLMIP problem, which cannot be solved directly using the MIP approaches, we propose a dual reformulation of SPLUNI. Fix the outer variable x and take the dual of the inner minimization using KKT conditions, the final MIP formulation is given as:
[ SPLUNI D ] max x X , π b T π
s . t . K T π = c + D x
π s = 0
where K is the network matrix controlling y as in Equation (14), r T x R , b = ( 1 , 0 , , 0 , 1 ) T , c k = c k / ( 1 + β × r g u k ) and d k k = d k k × ( 1 + α × r g u k ) / ( 1 + β × r g u k ) .

6. Experiments

For simplicity, we select a reduced Chicago Sketch Road Network [20] expanded from the vertex No.368 to its neighbors and neighbors’ neighbors for 5 times, consisting of 51 vertexes and 113 edges. The computations of the RGUR and RGUI are formulated into a BLMIP, and SPLUNI into a BLMIP and solved using the MIP solvers of CPLEX 11.5 and YALMIP toolbox of MATLAB [57]. Nontrivial details are omitted.

6.1. Tests on Uncertainty Reduction

Upon the reduced Chicago Sketch Road Network, a goal recognition task is defined where s 0 = s t a r t , G = { g o a l 1 , g o a l 2 } as shown in Figure 6c. We first show the influence of uncertainty on the goal recognition system and how rgu reduction model could be used in advancing situation awareness.
In this test, a dataset consisting of 100 labeled traces for each goal is collected using agent decision model, and we use F-measure which is a frequently used metric to measure the overall accuracy of the recognizer [1], computed as:
F measure = 2 · precision · recall precision + recall
precision = 1 N i = 1 N TP i TI i
recall = 1 N i = 1 N TP i TT i
where N is the number of possible goals, TP i , TI i and TT i are the true positives, total of true labels, and the total of inferred labels for class i respectively. Equations (20) to (22) show that F-measure is an integration of precision and recall, where precision is used to scale the reliability of the recognized results and recall to scale the efficiency of the algorithm applied in the test dataset. The value of F-measure will be between 0 and 1, and a higher metric value means a better performance. As to evaluate traces with different lengths, the paper applies the method in [35], and partitions each trace into k ( k = 1 , 2 , . . . , N stage ) stages. The corresponding sequences are y t i : k * length j / N stage j 1 : N trace .
From the results shown in Figure 6a,b, goal uncertainty associated with certain domains indeed seriously impedes effective goal identification. Under [RGUR-P] model, the values of F-measure after the network redesignation increases to almost 1 the moment agent selects its first action, compared to values under tasks with no network changes and the most ambiguous situation ([RGUI-P]). While for the RGUI-P model, its F-measure values are equal or greater than those of the original task in most times. Note the big difference of F-measure values between two models, i.e., [RGUR-P] and [RGUI-P], this simple case not only proves that the goal uncertainty could be purposefully controlled in both directions, but also shows the potential advantage that the observer could take off during the goal identification process.
Next, we statistically evaluate the maximum early prediction that our model enables according to Convergence Point metric [1], defined as 0 < τ = C P o i n t T , s . t . p τ ( g t r u e | o τ ) γ . C P o i n t denote the time step at which the posterior of the agent’s true goal is greater or equal to β . For tests in Figure 6d, γ is set to 0 . 8 . We test recognition tasks D = { s 0 , G = { g o a l 1 , g o a l 2 } } for s 0 S \ { G , U } , where U is the set of nodes with no available paths to G, and compare the early convergence ability of goal identification between [RGUR-P] and [RGUI-P] using a new metric R e l C o n v e r g e R a t i o , which describes the relative convergence ratio ( r c r ( % ) ) between two models given the s 0 g t r u e path length,
R e l C o n v e r g e R a t i o = C P o i n t I C P o i n t R | h = s 0 , , g t r u e | 1
where | h = s 0 , , g t r u e | is the length of path starting from s 0 while ending at the real goal g t r u e .
Clearly, both for tasks where g t r u e equals to g o a l 1 and g o a l 2 , there are considerable reductions of goal uncertainty. For example, there are total 5 cases in these two tasks whose relative convergence ratio is greater than or equal to 80 % . In addition, 7 cases in total fall within the 60%–80% interval, 16 at 40%–60% and 9 at 20%–40%. It should also be noted that there are more than 30 cases whose RelConvergeRatio equals to 0. This means that although the goal uncertainty widely exists in many identification tasks, it does not exist in every case.
Further for generality, we test 3 more cases where G 1 = { g o a l 2 , g o a l 3 } , G 2 = { g o a l 1 , g o a l 3 } and G 3 = { g o a l 1 , g o a l 2 , g o a l 3 } , as shown in Table 1. From these results, we could conclude that:
  • goal uncertainty exists in normal goal recognition task, and it could be purposefully controlled to affect the agent’s situation awareness.
  • the possible reduction of goal uncertainty differs in each scenario.

6.2. Performance of SPLUNI

Finally, we test the performance of SPLUNI (set α = β = 1 ), which is the game between the observer (interdictor) and the observed agent (evader), in maximizing the expectation of the shortest s t path length while minimizing the largest uncertainty, as shown in Table 2. We first define the path interdiction efficiency as e = Δ L / ( d T × x ) where Δ L is the increased path length for evader after interdiction and d T × x is the total length increase in the network. The tests are conducted over those ambiguous cases in three goal settings. E ( e ) and E ( r c r ) help us understand the impact of SPLUNI model on both network interdiction and goal uncertainty reduction.
Clearly, SPLUNI model works well for observer by adding the evader’s path length and at the same time reducing the goal uncertainty during the process. In all three goal settings, the average path interdiction efficiencies are more than 50%. Especially when G = { g 2 , g 3 } , E ( e ) reaches to almost 80 % for g t r u e = g 2 and more than 90 % for g t r u e = g 3 . This proves the effectiveness of SPLUNI model in network interdiction. Also, the model successfully reduces the goal uncertainty to a certain level in three different settings, as we could find in the data of E ( r c r ) . The above results show that the SPLUNI model that we propose is of great value for security domains where high-value targets need to be timely recognized meanwhile the intruder’s actions be delayed.

7. Conclusions

We present a new perspective of goal identification control using off-the-shelf probabilistic goal recognizer, and introduce the relative goal uncertainty (rgu) value for actions in a goal recognition task. We present ways of controlling goal uncertainty, followed by a presentation of the method in an adversarial setting using a SPLUNI model. Empirical evaluation shows the effectiveness of the proposed solution in controlling goal uncertainty in recognition tasks.
The work grounds the behaviors of the agents on clear, well-founded computational models. Inference over graphic models and mixed-integer programming is brought together to provide a compelling solution to the goal identification problem. Though this work is quite suggestive but still simple, which means there are many open problems in our future work. Currently, as there exists inconsistency of goal uncertainty between the offline P ( G ) which follows Uniform distribution and the real-time P ( G | O ) assessed online according to the agent’s observations, how to incorporate the goal uncertainty information into the goal identification control process online has not yet been solved. Also, we will also try to incorporate rgu definition into the plan identification process to improve the method’s applicability.

Author Contributions

Conceptualization, K.X.; supervision, Q.Y.

Funding

This work is partially sponsored by the National Natural Science Foundation of China under Grants No.61473300.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AHMMAbstract Hidden Markov Models
DBNDynamic Bayesian Network
GRDGoal Recognition Design
PRDPlan Recognition Design
MDPMarkov Decision Process
MIPMixed-Integer Programming
rgurelative goal uncertainty
KKTKarush-Kuhn-Tucker condition
wcdworst-case distinctiveness
SPLUNIShortest-Path Largest-Uncertainty Network Interdiction
RGURrgu Reduction
RGUIrgu Improvement

References

  1. Sukthankar, G.; Geib, C.; Bui, H.H.; Pynadath, D.; Goldman, R.P. Plan, Activity, and Intent Recognition: Theory and Practice; Elsevier: Waltham, MA, USA, 2014. [Google Scholar]
  2. Hofmann, A.G.; Williams, B.C. Intent Recognition for Human-Robot Interaction. In Proceedings of the AAAI 2007 Spring Symposium Interaction Challenges for Intelligent Assistants, Stanford, CA, USA, 26–28 March 2007; pp. 60–61. [Google Scholar]
  3. Min, W.; Ha, E.; Rowe, J.P.; Mott, B.W.; Lester, J.C. Deep Learning-Based Goal Recognition in Open-Ended Digital Games. AIIDE 2014, 14, 3–7. [Google Scholar]
  4. Geib, C.W.; Goldman, R.P. Plan recognition in intrusion detection systems. In Proceedings of the DARPA Information Survivability Conference & Exposition II. DISCEX’01, Anaheim, CA, USA, 12–14 June 2001; pp. 46–55. [Google Scholar]
  5. Jarvis, P.A.; Lunt, T.F.; Myers, K.L. Identifying terrorist activity with AI plan recognition technology. AI Mag. 2005, 26, 73. [Google Scholar]
  6. Keren, S.; Gal, A.; Karpas, E. Goal Recognition Design. In Proceedings of the Twenty-Fourth International Conference on Automated Planning and Scheduling, Portsmouth, NH, USA, 21–26 June 2014. [Google Scholar]
  7. Xu, K.; Yin, Q.; Qi, Z. A New Metric and Method for Goal Identification Control. In Proceedings of the IJCAI Workshop on Goal Reasoning, Stockholm, Sweden, 13–18 July 2018. [Google Scholar]
  8. Son, T.C.; Sabuncu, O.; Schulz-Hanke, C.; Schaub, T.; Yeoh, W. Solving Goal Recognition Design Using ASP. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 3181–3187. [Google Scholar]
  9. Wayllace, C.; Hou, P.; Yeoh, W. New Metrics and Algorithms for Stochastic Goal Recognition Design Problems. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 4455–4462. [Google Scholar]
  10. Kautz, H.A.; Allen, J.F. Generalized Plan Recognition. In Proceedings of the Fifth AAAI National Conference on Artificial Intelligence, Philadelphia, PA, USA, 11–15 August 1986; pp. 32–37. [Google Scholar]
  11. Pynadath, D.V.; Wellman, M.P. Generalized queries on probabilistic context-free grammars. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 65–77. [Google Scholar] [CrossRef] [Green Version]
  12. Pynadath, D.V. Probabilistic Grammars for Plan Recognition; University of Michigan: Ann Arbor, MI, USA, 1999. [Google Scholar]
  13. Pynadath, D.V.; Wellman, M.P. Probabilistic state-dependent grammars for plan recognition. In Proceedings of the Sixteenth Conference on Uncertainty in Artificial Intelligence, Stanford, CA, USA, 30 June–3 July 2000; pp. 507–514. [Google Scholar]
  14. Geib, C.W.; Goldman, R.P. A probabilistic plan recognition algorithm based on plan tree grammars. Artif. Intell. 2009, 173, 1101–1132. [Google Scholar] [CrossRef] [Green Version]
  15. Wellman, M.P.; Breese, J.S.; Goldman, R.P. From knowledge bases to decision models. Knowl. Eng. Rev. 1992, 7, 35–53. [Google Scholar] [CrossRef]
  16. Charniak, E.; Goldman, R.P. A Bayesian model of plan recognition. Artif. Intell. 1993, 64, 53–79. [Google Scholar] [CrossRef]
  17. Bui, H.H.; Venkatesh, S.; West, G. Policy recognition in the abstract hidden markov model. J. Artif. Intell. Res. 2002, 17, 451–499. [Google Scholar] [CrossRef]
  18. Bui, H.H. A general model for online probabilistic plan recognition. In Proceedings of the 18th International Ioint Conference on Artificial Intelligence, Acapulco, Mexico, 9–15 August 2003; pp. 1309–1315. [Google Scholar]
  19. Liao, L.; Patterson, D.J.; Fox, D.; Kautz, H. Learning and inferring transportation routines. Artif. Intell. 2007, 171, 311–331. [Google Scholar] [CrossRef] [Green Version]
  20. Xu, K.; Xiao, K.; Yin, Q.; Zha, Y.; Zhu, C. Bridging the gap between observation and decision making: Goal recognition and flexible resource allocation in dynamic network interdiction. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 4477–4483. [Google Scholar]
  21. Baker, C.L.; Saxe, R.; Tenenbaum, J.B. Action understanding as inverse planning. Cognition 2009, 113, 329–349. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Ramırez, M.; Geffner, H. Plan recognition as planning. In Proceedings of the 21st International Joint Conference on Artifical Intelligence, Pasadena, CA, USA, 11–17 July 2009; pp. 1778–1783. [Google Scholar]
  23. Ramırez, M.; Geffner, H. Probabilistic plan recognition using off-the-shelf classical planners. In Proceedings of the Conference of the Association for the Advancement of Artificial Intelligence (AAAI 2010), Atlanta, GA, USA, 11–15 July 2010; pp. 1121–1126. [Google Scholar]
  24. Ramırez, M.; Geffner, H. Goal recognition over POMDPs: Inferring the intention of a POMDP agent. In Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence, Barcelona, Spain, 16–22 July 2011; pp. 2009–2014. [Google Scholar]
  25. Sohrabi, S.; Riabov, A.V.; Udrea, O. Plan Recognition as Planning Revisited. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016; pp. 3258–3264. [Google Scholar]
  26. Albrecht, D.W.; Zukerman, I.; Nicholson, A.E. Bayesian models for keyhole plan recognition in an adventure game. User Model. User-Adapted Interact. 1998, 8, 5–47. [Google Scholar] [CrossRef]
  27. Goldman, R.P.; Geib, C.W.; Miller, C.A. A new model of plan recognition. In Proceedings of the Fifteenth Conference on Uncertainty in Artificial Intelligence, Stockholm, Sweden, 30 July–1 August 1999; pp. 245–254. [Google Scholar]
  28. Doucet, A.; De Freitas, N.; Murphy, K.; Russell, S. Rao-Blackwellised particle filtering for dynamic Bayesian networks. In Proceedings of the Sixteenth Conference on Uncertainty in Artificial Intelligence, San Francisco, CA, USA, 30 June–3 July 2000; pp. 176–183. [Google Scholar]
  29. Chen, Z. Bayesian filtering: From Kalman filters to particle filters, and beyond. Statistics 2003, 182, 1–69. [Google Scholar] [CrossRef]
  30. Saria, S.; Mahadevan, S. Probabilistic Plan Recognition in Multiagent Systems. In Proceedings of the Fourteenth International Conference on International Conference on Automated Planning and Scheduling, Whistler, BC, Canada, 3–7 June 2004; pp. 287–296. [Google Scholar]
  31. Blaylock, N.; Allen, J. Fast hierarchical goal schema recognition. In Proceedings of the National Conference on Artificial Intelligence, Boston, MA, USA, 16–20 July 2006; pp. 796–801. [Google Scholar]
  32. Singla, P.; Mooney, R.J. Abductive Markov Logic for Plan Recognition. In Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 7–11 August 2011; pp. 1069–1075. [Google Scholar]
  33. Krüger, F.; Nyolt, M.; Yordanova, K.; Hein, A.; Kirste, T. Computational state space models for activity and intention recognition. A feasibility study. PLoS ONE 2014, 9, e109381. [Google Scholar] [CrossRef] [PubMed]
  34. Yin, Q.; Yue, S.; Zha, Y.; Jiao, P. A semi-Markov decision model for recognizing the destination of a maneuvering agent in real time strategy games. Math. Probl. Eng. 2016, 2016, 1907971. [Google Scholar] [CrossRef]
  35. Yue, S.; Yordanova, K.; Krüger, F.; Kirste, T.; Zha, Y. A Decentralized Partially Observable Decision Model for Recognizing the Multiagent Goal in Simulation Systems. Discr. Dyn. Nat. Soc. 2016, 2016, 5323121. [Google Scholar] [CrossRef]
  36. Yordanova, K.; Lüdtke, S.; Whitehouse, S.; Krüger, F.; Paiement, A.; Mirmehdi, M.; Craddock, I.; Kirste, T. Analysing Cooking Behaviour in Home Settings: Towards Health Monitoring. Sensors 2019, 19, 646. [Google Scholar] [CrossRef] [PubMed]
  37. Bisson, F.; Larochelle, H.; Kabanza, F. Using a Recursive Neural Network to Learn an Agent’s Decision Model for Plan Recognition. In Proceedings of the 24th International Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015; pp. 918–924. [Google Scholar]
  38. Tastan, B.; Chang, Y.; Sukthankar, G. Learning to intercept opponents in first person shooter games. In Proceedings of the 2012 IEEE Conference on Computational Intelligence and Games (CIG), Granada, Spain, 11–14 September 2012; pp. 100–107. [Google Scholar]
  39. Zeng, Y.; Xu, K.; Yin, Q.; Qin, L.; Zha, Y.; Yeoh, W. Inverse Reinforcement Learning Based Human Behavior Modeling for Goal Recognition in Dynamic Local Network Interdiction. In Proceedings of the AAAI Workshops on Plan, Activity and Intent Recognition, New Orleans, LA, USA, 2–3 February 2018. [Google Scholar]
  40. Agotnes, T. Domain independent goal recognition. In Proceedings of the 2010 Conference on STAIRS 2010: Proceedings of the Fifth Starting AI Researchers’ Symposium, Lisbon, Portugal, 16–20 August 2010; pp. 238–250. [Google Scholar]
  41. Pattison, D.; Long, D. Accurately determining intermediate and terminal plan states using bayesian goal recognition. In Proceedings of the First Workshop on Goal, Activity and Plan Recognition, Freiburg, Germany, 11–16 June 2011. [Google Scholar]
  42. Yolanda, E.; R-Moreno, M.D.; Smith, D.E. A fast goal recognition technique based on interaction estimates. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015. [Google Scholar]
  43. Pereira, R.F.; Oren, N.; Meneguzzi, F. Landmark-based heuristics for goal recognition. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17), San Francisco, CA, USA, 4–9 February 2017; pp. 3622–3628. [Google Scholar]
  44. Masters, P.; Sardina, S. Cost-Based Goal Recognition in Navigational Domains. J. Artif. Intell. Res. 2019, 64, 197–242. [Google Scholar] [CrossRef]
  45. Masters, P.; Sardina, S. Deceptive Path-Planning. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 4368–4375. [Google Scholar]
  46. Cohen, P.R.; Perrault, C.R.; Allen, J.F. Beyond question answering. In Strategies for Natural Language Processing; Psychologh Press: Hove, UK, 1981; pp. 245–274. [Google Scholar]
  47. Jensen, R.M.; Veloso, M.M.; Bowling, M.H. OBDD-based optimistic and strong cyclic adversarial planning. In Proceedings of the Sixth European Conference on Planning, Toledo, Spain, 12–14 September 2014. [Google Scholar]
  48. Avrahami-Zilberbrand, D.; Kaminka, G.A. Keyhole adversarial plan recognition for recognition of suspicious and anomalous behavior. In Plan, Activity, and Intent Recognition: Theory and Practice; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 2014; pp. 87–121. [Google Scholar]
  49. Avrahami-Zilberbrand, D.; Kaminka, G.A. Incorporating observer biases in keyhole plan recognition (efficiently!). In Proceedings of the Twenty-Second National Conference on Artificial Intelligence, Vancouver, BC, Canada, 22–26 July 2007; pp. 944–949. [Google Scholar]
  50. Braynov, S. Adversarial planning and plan recognition: Two sides of the same coin. In Proceedings of the Secure Knowledge Management Workshop, Brooklyn, NY, USA, 28–29 September 2006; pp. 67–70. [Google Scholar]
  51. Le Guillarme, N.; Mouaddib, A.I.; Gatepaille, S.; Bellenger, A. Adversarial Intention Recognition as Inverse Game-Theoretic Planning for Threat Assessment. In Proceedings of the 2016 IEEE 28th International Conference on Tools with Artificial Intelligence (ICTAI), San Jose, CA, USA, 6–8 November 2016; pp. 698–705. [Google Scholar]
  52. Keren, S.; Gal, A.; Karpas, E. Goal Recognition Design with Non-Observable Actions. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 3152–3158. [Google Scholar]
  53. Wayllace, C.; Hou, P.; Yeoh, W.; Son, T.C. Goal recognition design with stochastic agent action outcomes. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016; pp. 3279–3285. [Google Scholar]
  54. Mirsky, R.; Stern, R.; Ya’akov (Kobi) Gal, M.K.; Kalech, M. Plan Recognition Design. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 4971–4972. [Google Scholar]
  55. Mirsky, R.; Gal, K.; Stern, R.; Kalech, M. Goal and Plan Recognition Design for Plan Libraries. ACM Trans. Intell. Syst. Technol. 2019, 10, 14. [Google Scholar] [CrossRef]
  56. Nyolt, M.; Krüger, F.; Yordanova, K.; Hein, A.; Kirste, T. Marginal filtering in large state spaces. Int. J. Approx. Reason. 2015, 61, 16–32. [Google Scholar] [CrossRef]
  57. Lofberg, J. YALMIP: A toolbox for modeling and optimization in MATLAB. In Proceedings of the 2004 IEEE International Symposium on Computer Aided Control Systems Design, Taipei, Taiwan, 2–4 September 2004; pp. 284–289. [Google Scholar]
Figure 1. An example where goal identification process could be controlled, where C1 is the Start, and A5 and E5 are exit points denoted as Goal 1 and Goal 2, respectively. In this example, path ‘C1-C2-C3-C4-C5-B5-A5’ contains the largest goal uncertainty, while ‘C1-B1-A1-A2-A3-A4-A5’ the least.
Figure 1. An example where goal identification process could be controlled, where C1 is the Start, and A5 and E5 are exit points denoted as Goal 1 and Goal 2, respectively. In this example, path ‘C1-C2-C3-C4-C5-B5-A5’ contains the largest goal uncertainty, while ‘C1-B1-A1-A2-A3-A4-A5’ the least.
Entropy 21 00299 g001
Figure 2. The motivating example of the GRD problem, where the wcd value is 4 and the blockade of the action moving the agent from C 1 to C 2 successfully reduces the wcd from 4 to 0.
Figure 2. The motivating example of the GRD problem, where the wcd value is 4 and the blockade of the action moving the agent from C 1 to C 2 successfully reduces the wcd from 4 to 0.
Entropy 21 00299 g002
Figure 3. The DBN structure depicting causalities within two time slices of the model.
Figure 3. The DBN structure depicting causalities within two time slices of the model.
Entropy 21 00299 g003
Figure 4. r g u E and the discounted r g u d i s of each action in a 3 × 3 grid network
Figure 4. r g u E and the discounted r g u d i s of each action in a 3 × 3 grid network
Entropy 21 00299 g004
Figure 5. The moving traces (the solid line for agents moving after the observer’s interdiction; dashed line for the observed agent selecting the most ambiguous path) and their corresponding recognition results p ( G | O ) for the example shown in Figure 1.
Figure 5. The moving traces (the solid line for agents moving after the observer’s interdiction; dashed line for the observed agent selecting the most ambiguous path) and their corresponding recognition results p ( G | O ) for the example shown in Figure 1.
Entropy 21 00299 g005
Figure 6. The statistical evaluation of uncertainty influence on goal recognition upon a reduced sketch road network.
Figure 6. The statistical evaluation of uncertainty influence on goal recognition upon a reduced sketch road network.
Entropy 21 00299 g006
Table 1. The number of cases (discard states unreachable to g t r u e ) fall in intervals of relative convergence ratio ( r c r ( % ) ). “/” separates results with elements in G being the true goal.
Table 1. The number of cases (discard states unreachable to g t r u e ) fall in intervals of relative convergence ratio ( r c r ( % ) ). “/” separates results with elements in G being the true goal.
G \ rcr ( % ) 0 20 40 60 80
{ g o a l 2 , g o a l 3 } 23/241/33/32/01/0
{ g o a l 1 , g o a l 3 } 25/312/46/00/00/0
{ g 1 , g 2 , g 3 } 15/21/193/6/14/1/43/1/10/0/0
Table 2. The expectations of path interdiction efficiency and relative convergence ratio ( r c r ( % ) ) for different goal setting.
Table 2. The expectations of path interdiction efficiency and relative convergence ratio ( r c r ( % ) ) for different goal setting.
{ g 1 , g 2 } { g 1 , g 3 } { g 2 , g 3 }
E ( e ) 56.0/73.663.4/67.077.4/90.7
E ( r c r ) 19.7/13.37.4/9.12.3/1.9

Share and Cite

MDPI and ACS Style

Xu, K.; Yin, Q. Goal Identification Control Using an Information Entropy-Based Goal Uncertainty Metric. Entropy 2019, 21, 299. https://doi.org/10.3390/e21030299

AMA Style

Xu K, Yin Q. Goal Identification Control Using an Information Entropy-Based Goal Uncertainty Metric. Entropy. 2019; 21(3):299. https://doi.org/10.3390/e21030299

Chicago/Turabian Style

Xu, Kai, and Quanjun Yin. 2019. "Goal Identification Control Using an Information Entropy-Based Goal Uncertainty Metric" Entropy 21, no. 3: 299. https://doi.org/10.3390/e21030299

APA Style

Xu, K., & Yin, Q. (2019). Goal Identification Control Using an Information Entropy-Based Goal Uncertainty Metric. Entropy, 21(3), 299. https://doi.org/10.3390/e21030299

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop