Observation Subset Selection as Lo cal Compilation of Performance Profiles Yan Radovilsky Computer Science Dept. Ben-Gurion University 84105 Beer-Sheva, Israel yanr@cs.bgu.ac.il Solomon Eyal Shimony Computer Science Dept. Ben-Gurion University 84105 Beer-Sheva, Israel shimony@cs.bgu.ac.il To tackle this problem we present OSS as a variant of the following well-known meta-reasoning problem. In systems composed of several computational components (CCs), the meta-level controller should reason about allocation of available computational resources for different CCs in order to optimize the overall performance of the entire system. This task is usually referred to in the research literature as the meta-level resource al location (MRA) problem (see for example [11]). The standard approach used to optimize the MRA task was proposed by S. Zilberstein [10, 11, 12], the technique of local compilation (LC). This technique is applied to individual CCs, represented in a form of conditional performance profiles, and generates the optimal time al location scheme for the entire system (see Section 2). However, local compilation requires the input monotonicity assumption and is, therefore, restricted to deterministic performance profiles with scalar output quality. In this paper we relax the input monotonicity assumption and extend the LC technique to more general classes of PPs. We then apply the extended approximation scheme (Section 4) to find an approximate solution to the OSS problem, under two settings, both for dependencies modeled as a tree-shaped BN: a) finding a maximum expectation variable in a binary valued BN (Section 4.1), and b) minimizing the worst variance in a Gaussian BN (Section 4.2). Abstract Deciding what to sense is a crucial task, made harder by dependencies and by a nonadditive utility function. We develop approximation algorithms for selecting an optimal set of measurements, under a dependency structure modeled by a tree-shaped Bayesian network (BN). Our approach is a generalization of composing anytime algorithm represented by conditional performance profiles. This is done by relaxing the input monotonicity assumption, and extending the local compilation technique to more general classes of performance profiles (PPs). We apply the extended scheme to selecting a subset of measurements for choosing a maximum expectation variable in a binary valued BN, and for minimizing the worst variance in a Gaussian BN. 1 Introduction A typical diagnostic system consists of two types of variables: tests (observable) and hypotheses (unobservable), with statistical dependencies among variables. Each test, if performed, consumes resources (time or money), and provides a measurement of one or more test variables. After obtaining the values of the selected tests, the distribution of the model is updated. An ob jective function specifies a reward given to the system for the posterior distribution. The system should make its selection so as to optimize the ob jective function, a hard problem in the general case. Observation subset selection (OSS) is a restricted version of this problem, where all measurements must be selected in advance, prior to any observations. In this paper we develop approximation algorithms for some settings of the OSS problem for tree-shaped dependency structures. 2 Background Flexible computation refers to procedures that allow a graceful trade-off to be made between the quality of results and allocation of costly resources, such as time, memory, or information [3]. Since time is usually the main computational resource, there are several alternative terms used for reference to flexible computation in the research literature: continual computation [4, 5], anytime computation [8], and anytime algorithms [1, 10, 12]. To predict the quality of the result which depends on the amount of allocated time, a statistical model called a performance profile (PP) is employed. The most simple version of such a PP called an expected performance profile (EPP), a mapping from consumed time to an expected result quality, Q : T Q. Given anytime algorithm A described by EPP QA and time-dependent utility function U : T × Q R, optimal time allocation t can be derived as follows: t = arg max U (t, QA (t)) t where operator · denotes a summatii n over all eleo ^ ments of its argument vector: t = ^ [t]i . An intuitive approach to tackle the MRA problem has been proposed in [10]. This approach, called compilation of performance profiles, involves construction of an appropriate PP for the entire system: Qc (t, qin ) = ^ tTn : t =t ^ max ^ (t, qin ), (3) (1) When an algorithm operates with some inexact inputs, the quality of its result may also strongly depend on the quality of the inputs. To address this point, a more flexible model, known as a conditional performance profile (CPP), is used. For example, an anytime algorithm with 2 inputs and one output can be represented by a CPP Q : T × Q2 Q. Each entry in Qc is additionally associated with a cor^ ^ responding TAS (t, qin ) = t s.t. (t, qin ) = Qc (t, qin ) and t = t. ^ Once such a model is constructed, the optimal solution for the MRA problem can be derived from this model ^ as t = (t , qin ), where t is the optimal total time allocation computed using equation 1. Algorithm 1: method RLC for in-tree shaped composite systems Input : s - a node in a composition graph. Output: Composed PP for the subtree rooted in s. if (Pred (s) = ) then return Qs ; foreach si Pred (s) do Let Qci RLC (si ); s Let L {Qci : si Pred (s)}; s return Compose (Qs , L); In general, the task of (global) compilation is computationally hard even for approximate solution [10]. However, for some restricted topologies (e.g. trees) S. Zilberstein [10] proposed an efficient algorithm based on the local compilation (LC) technique, summarized as RLC in Algorithm 1. In this technique a composite PP of each subtree is generated based on the CPP of its root and composed PPs of all its predecessors. Method Compose performs the basic composition operation, which results in the following composite PP: Qc (t, q1 , . . . , qk ) = (4) s c c maxP Qs (t0 , Qs1 (t1 , q1 ), . . . , Qsk (tk , qk )) t0 ,...,tk T: ti =t Figure 1: Composition graph In a complex system, several CCs (or anytime algorithms ) can be composed to achieve a required result. In this case, their PPs should be compiled in order to obtain an appropriate performance prediction for the entire system. Such a system is usually described in a graphical form by a directed acyclic graph (DAG), where each node corresponds to a single CC and is associated with an appropriate PP, while edges represent dependencies between input/output qualities of different CCs (Figure 1). This model is referred to as a composition graph (CG). ^ A time al location scheme (TAS) t Tn specifies allocation of time for each of n elementary CCs of a composite system. Output quality of the whole system can be expressed as a function of a TAS in a form of composite expression (CE) : Tn × Qk Q, where k is a number of external inputs. For example, the system in figure 1 has the following CE: ^ ^ ^ ^ ^ (t, qin ) = Q1 ([t]1 , Q2 ([t]2 , Q3 ([t]3 , qin )), Q3 ([t]3 , qin )) The MRA problem for a composite system is to optimize a time-dependent utility function by selecting an appropriate TAS. Formally, given a composite system represented by CE , external input quality qin , and ^ utility function U , the goal is to find TAS t Tn which maximizes the utility: ^ ^ t = arg max U ( t , (t, qin )), ^ ^ tTn where s1 , . . . , sk are the predecessors of node s. For systems, where all CCs are represented by deterministic CPPs, RLC is proved to produce the correct result (equals to one obtained by global compilation ) when the following assumptions hold: · A tree-structured CG - each node has only one output which serves as an input for one successor node except for the root node, whose output is the resulting system output. (2) · Input monotonicity (IM) - the utility function and all involved CPPs are non-decreasing functions of input quality, i.e. q p : Q(t, q ) Q(t, p). Such restrictive assumptions are required to enable independent compilation of each subtree followed by greedy selection of one local TAS for each total allocation without harming optimality of the resulting solution. However, in practice, the IM assumption may be too restrictive. For example, a diagnostic system with a dependency model represented by Gaussian BN and a utility function depending on posterior marginal precision of several hypothesis nodes, violates this assumption. In this work we propose an extension to the CPP model, which allows relaxing the IM assumption. Recent work exists in another line directly related to the OSS task. In [7] the authors proved that the OSS problem is NP PP -hard even for tree-structured BNs. They proposed a polynomial time algorithm, based on dynamic programming, which constructs an optimal solution to a version of OSS restricted to chain topology, exact observation, and additive reward. In [9] a similar technique was used beyond the exact observation assumption, and determined a theoretical bound for the worst-case loss in expected reward. Our work can be seen as an extention of the latter research. Another approximation method based on greedy selection of test nodes is applicable when the reward function exhibits property of sub-modularity [6]. can contain more than one output quality for the same pair of total time allocation and input quality (each with a different local TAS). We further extend our model by allowing backward conditioning, which means that output quality of a CC may depend on output quality of its successor. The general form of a CP in this case is ^ (t, (psuc , ppred ), (qsuc , qpred )), where the p-component has two parts: psuc is a required output quality of the successor, and ppred is a vector of required output qualities of all the predecessors; the q -component has two parts as well: an output quality towards its successor (qsuc ), and a vector of qualities towards all its predecessors (qpred ). Such a form of CPs is useful in applications for BNs, where posterior probability distribution of a node (and its local reward) depends on the messages coming from all its neighbors (details appear in section 4). In order to adapt the RLC algorithm to the extended settings, we need to modify the Compose method. This method is now applied to a list of profiles in the RPP representation, and its output should be in the RPP form as well. Moreover, the greedy selection reflected in the max operator in equation 3 strongly relies on the IM assumption. Since this assumption fails in the RPP case, we propose another approach. The composition process comprises two parts: the first part (Construct ) considers all combinations of input CPs, and collects appropriate resulting CPs. ^ ^ ^ ^ Q = {(t0 + t1 + . . . + tk , p0 , q0 ) : (ti , pi , qi ) Qsi , ^ (t0 , (p0 , q1 , . . . , qk ), (q0 , p1 , . . . , pk )) Qs }. (6) The second part, called Purge is applied to the set Q. Purge exploits domination and equivalence among reachable CPs in order to filter out irrelevant CPs.1 The idea of pruning irrelevant CPs is very similar to one used in the Incremental Pruning algorithm [2] for filtering irrelevant -vectors. The resulting RPP Qc keeps one representative for each CP in Q: a Q b Qc s.t. (b a) (b a). where and are domination and equivalence operators respectively. These operators can be defined based on partial orders within each performance component. Assuming only a natural partial order in the t-component we obtain: ^ ^ ^ ^ (t, p, q ) (t , p , q ) ( t = t ) (p p ) (q q ) and ^ ^ ^ ^ (t, p, q ) (t , p , q ) ( t < t ) (p p ) (q q ) 1 The filtering can b e applied after Construct terminates, but it is usually more efficient to p erform filtering on-the-fly. 3 Generalized Local Compilation This section generalizes the notion of a performance profile to a reachable performance profile, and adapts the LC technique to handle the extended model. Definition 3.1 (Conditional performance). Conditional performance (CP) of a CC (either elementary ^ ^ or composite) is a tuple (t, p, q ), where t is a local TAS (t-component), p is a required input quality (pcomponent), and q is an expected output quality, obtained by the CC when applied to inputs of quality p ^ with TAS t (q -component). Either p or q may be represented by scalars or by vec^ tors, while t is assumed to be a complete assignment of time allocations to all CCs of the system. Definition 3.2 (Reachable performance profile). Reachable performance profile (RPP) of a CC is a set of CPs achievable by this CC. The RPP model can be derived from a CPP as follows: RA = {(A (t, p), p, QA (t, p))}, (5) P Q P Q where QA is a CPP, and RA is an appropriate RPP. The converse is not always possible, because a RPP where and are equivalence operators defined over the p and q components, respectively. P Q notation to refer to other relevant subsets of nodes: XO = X \ XI ; s s I Ys = {Yi : Xi (XM XI )}; s O I Ys = Y \ Ys ; I I Es = E Ys ; O O Es = E Ys . 4 Observation Selection in BNs. In this section we describe how our framework can be applied to OSS in BNs. We use the following notation: · X = {Xi : 1 i n} - set of all state variables; · XH X - set of hypothesis state variables; · XM X - set of measurable state variables; · Y = {Yi : Xi XM } - set of test variables; · N - BN over X Y; · R : Pr(XH ) R - reward function; · T (E ) = Y i E 4. 1 OSS in discrete BNs We now apply our approach to OSS in tree-shaped BNs with discrete variables. Since solving this problem for the general setting (even for the tree-structured topology) is proved to be NP PP -hard [7], we restrict our focus to BNs with boolean variables (Dom(Xi ) = Dom(Yi ) = {0, 1}), and consider the following reward function, defined for an arbitrary set of nodes A: R(A|E = e) = max Ri (Xi |E = e), Xi A i - additive time consumption; · B - time budget (maximum time for observation). Definition 4.1 (OSS optimization problem). The OSS optimization problem is: given a tuple (N , R, T , B ), select a subset of observation variables E Y which maximizes the expected reward: ^ R(XH |E ) = e Pr(E = e)R(XH |E = e) (7) where each Ri : Pr(Xi ) R is a local reward function: P r(Xi = 1|E = e) : if(Xi XH ), Ri (Xi |E = e) = 0 : otherwise. (8) We refer to this version of OSS as Boolean OSS (BOSS). The BN can be specified by the following parameters: P r(Xi = 1) : if Xi is a root, i = Pr(Xi = 1|XP rev(i) = 1) : otherwise; P r(Xi = 1) : if Xi is a root, i = Pr(Xi = 1|XP rev(i) = 0) : otherwise; P r(Yi = 0|Xi = 1) : if(Xi XM ), i = 1 : otherwise; P r(Yi = 1|Xi = 0) : if(Xi XM ), i = 0 : otherwise; We make the following simplifying assumptions about the involved observation process: 1. Probability of a false positive observation result for all nodes is bounded by a small constant max (that is i : i max ); D o m(E ) subject to the budget limit T (E ) B . Figure 2: Out-tree BN topology. We assume BNs with an out-tree shaped dependency graph rooted in node X1 (as shown in figure 2). Let XI s denote a subset of X, which consists of all nodes Xi in the subtree rooted in Xs (all descendants of Xs including Xs itself ), and let E Y be any given subset of observation ("evidence") nodes. We use the following 2. Only hypothesis nodes have directly attached observations (XM XH ). Henceforth this set of assumptions is called the restricted false positive property. In the extreme case (max = 0) we get a false-positive-free observation process. Despite the relatively restricted setting, we have the following complexity result: Theorem 1 (Hardness of BOSS). Finding an exact solution to the BOSS problem is NP -hard even when al l state variables are independent (i = i ) and al l observations are exact (i = i = 0). Proof is by reduction from Knapsack, which is a wellknown NP -complete problem. Below, we show how the BOSS problem can be reduced to a special case of MRA and then solved (approximately) by the RLC algorithm. In order to apply the RLC algorithm we must specify the problem in terms of a composite system. Deriving the corresponding CG is straightforward: the graph has in-tree form and can be obtained from the dependency graph by simply reversing directions of all arcs. Careful inspection of the false-positive-free property yields that observing one positive value at any observation node (Yi = 1) provides a sufficient evidence for determining the reward value (R(XH |Yi = 1, E = e) = Ri (Xi |Yi = 1) = 1), regardless of other observations. We employ this fact to obtain a recursive decomposition of the expected reward. We specify output quality (q -component) of exploring subset E Y w.r.t. subtree XI as the triple (f , g , r): s f = Pr(eI |Xs = 1), ^s g = Pr(eI |Xs = 0), ^s r = R(XI |e), s^ where e, ^ and denote assignments of all zeros to I O E , Es , and Es respectively. While f and g components depend only on observations inside the subtree (eI ), to determine the value of the r-component we ^s need additional information from outside the subtree, provided by the p-component: p = Pr(Xs = 1|eO ). ^s For each quality component we define one corresponding domain set (of relevant values): P F O Ps = r(Xs = 1|Es = eO ) : E Y ^s P G I r(Es = eI |Xs = 1) : E Y ^s s= P R I r(Es = eI |Xs = 0) : E Y ^s s= RI . (Xs |E = e) : E Y ^ s= We also define combined domain sets Qs = Fs × Gs × Rs . Finally, all alternative assignments to a number of observations (measurements) in node Xs is represented by set Ms . In the basic OSS setting we have at most one observation per node: { 0, 1} : if (Xs XM ), Ms = : otherwise. {0} However, the model can be easily extended to multiple observations (by specifying appropriate Ms sets). eI , ^s eO ^s RPPs of observation nodes contain CPs with no condition (denoted by in the p-component): QY = {(ms , , m) : m Ms }, ^ s (9) where s is an assignment of s time units to s and 0 ^ to all the other CCs. All leaf X -nodes are associated with RPPs of the following form: QX = {(^, u, s (u)) : u Ps × Ms } 0 s (10) where ^ denotes a zero time allocation (to all CCs), 0 and s : Ps × Ms Qs is a vector function defined as follows: s (p, m) = (f , g , r), f g = 0) = (1 - s )m , p f : if (Xs XH ), L(f ,g,p) r = Rs (Xs |e) = ^ 0 : otherwise In our notation L(·, ·, ·) stands for the operator of linear interpolation defined as follows: L(a, b, c) = ca + (1 - c)b. (12) = Pr(eI |Xs ^s = Pr(eI |Xs ^s = 1) = m s , (11) RPP of any non-leaf node Xs with k children (Xs1 , . . . , Xsk ) is specified as follows: 0 QX = {(^, u, s (u)) : u Ps × Ms × Qs1 × · · · × Qsk } s (13) where s : Ps × Ms × Qs1 × · · · × Qsk Qs × Ps1 × · · · × Psk is a vector function defined as follows: s (p, m, (f1 , g1 , r1 ), . . . , (fk , gk , rk )) = = ((f , g , r), p1 , . . . , pk ), r = max{r0 , r1 , . . . , rk }, p 0 f L(f ,g,p) (14) r0 = Rs (Xs |e) = ^ : if (Xs XH ), : otherwise fi , 1 gi , m f = Pr(eI |Xs = 1) = s ^s g= Pr(eI |Xs ^s = 0) = (1 - s )m 1|eO ) ^si 1 ik ik pi = Pr(Xsi = fi = Pr(eIi |Xs = 1) = L(fi , gi , si ), ^s gi = Pr(eIi |Xs = 0) = L(fi , gi , si ). ^s = j m L(si , si , ps =i fj ) j j = , , (1 - )m m L(s s =i fj =i gj , p) While all QY RPPs can be specified explicitly (as a list s of CPs), the QX RPPs generally cannot, due to a poss sibly exponential size of domains Ps , Fs , Gs and Rs in their input condition. Instead, we represent them in an implicit form by providing (parameters of ) the involved s functions. During the compilation process (according to the RLC algorithm) method Compose is applied to lists of RPPs. At each application one composed RPP Qc s (which represents the whole subtree XI ) is generated. s Due to backward conditioning on elements of Ps which can contain exponential number of elements), such a RPP cannot be derived exactly (and explicitly). There is also no obvious way to specify it implicitly by providing some predefined functions (as we do in case of individual RPPs). To address this problem, we propose to approximate all Ps sets by one uniform grid Dp defined as follows: ( k 1 15) Dp = + p : k 0, . . . , dp - 1 2 1 i where 0 < p < 1 is a small constant, and dp = p s a number of intervals of size p in the range [0, 1]. Thus, set Dp has a fixed number of elements (dp ) which makes efficient enumeration possible. We define discretization function p , which maps any value p Ps to an appropriate point in set Dp : p+ 1 (16) p (p) = p p 2 This discretization induces approximate equivalence relation among elements of Ps defined as follows: p p p (p) = p (p ) P P After compiling RPPs of the entire tree, the resulting RPP Qc can be used to select a near-optimal TAS 1 w.r.t. this utility function. Let E denote the optimal observation subset, and let E be a subset corresponding to the TAS selected based on the resulting RPP. Theorem 2 (Approximation quality of RLC for BOSS). The RLC algorithm applied to a transformed instance of BOSS with out-tree topology and a restricted false positive observation process approximates the optimal solution within additive factor of u , which is bounded as fol lows: ^ ^ u = R(X|E ) - R(X|E ) hp + 2ns + r + nmax (19) where h is a height of the tree, and s = max{f , g }. Selection of the appropriate values for the grid steps depends on the required precision of the solution. To ensure approximation with u + nmax (in worst case) we can select p = 3 , f = g = 6 , and r = 3 . h n Due to monotonicity in total time, any composed RPP can be represented by the 4-dimensional table (Dp × Df × Dg × Dr ), where each entry is associated with at most one appropriate partial TAS. The number of entries (CPs) in such a table is bounded as follows: 6 23. 3 n h c (20) |Qs | dp df dg dr = Worn t-case complexitysfor the complete run of RLC is st n3 4 ime, O 4 pace for a chain topology, and O 4 n2c+1 t n2c 2 s h O 2c+2 c ime, O 2ch 2c pace for a tree with a + maximum branching factor of c. 4. 2 OSS in Gaussian Bayesian networks (17) We apply a similar discretization technique to other domain sets (Fs , Gs , and Rs ) with discretization steps f , g and r respectively. The appropriate grids (Df , Dg , Dr ), discretization functions (f , g , r ) and equivalence operators (, , ) are defined accordingly. The composed equivalence operator is defined as follows: (f , g , r) (f , g , r ) (f f ) (g g ) (r r ) To complete the specification, we define the following comprehensive utility function: ^ U (t, p, (f , g , r)) = L (r, 1, L(f , g , 1 )) = - (18) : if (p 1 t B ), ^ : otherwise, P Q F G R Q F G R Gaussian Bayesian network (GBN) is a special case of BN, where the conditional probability distributions of the variables are Normal (Gaussian) distributions: X 2 ai,j (Xj - µj ), i Xi |P a(Xi ) N µi + j P a(Xi ) We parametrize our GBN model as follows: i = a2,P rev(i) , i 1 i = P rec(Xi |XP rev(i) ) = 2 , i P rec(Yi |Xi ) : if (Xi XM ), i = 0 : otherwise. In our notation P rec(·) denotes the precision operator, which is reciprocal to variance: P rec(Xi |E ) = 1 . V ar(Xi |E ) (21) For Gaussian OSS (GOSS) we consider the following reward function: R(A|E = e) = min Ri (Xi |E ), Xi A For the non-leaf nodes we have: s (p, m, (f1 , r1 ), . . . , (fk , rk )) = ((f , r), p1 , . . . , pk ); 1 I f = P rec(Xs |Es ) = ms + fi , ik where Ri : Pr(Xi ) R are local reward functions: l oga,b (P rec(Xi |E )) : if(Xi XH ), Ri (Xi |E ) = 1 : otherwise. (22) Here a and b are two parameters that determine a range of distinguishable (for reward) values of precision, and loga,b (·) denotes a normalized (and truncated at its extreme points) log operator, defined as follows: 0 loga,b (p) = 1 log p-log a log b-log a r = min{r0 , r1 , . . . , rk }, l oga,b (p + f ) : if (Xs XH ), r0 = Rs (Xs |E ) = 1 : otherwise, j O pi = P rec(Xsi |Esi ) = J (si , si (ms + p + fj )), =i I fi = P rec(Xs |Esi ) = J (fi , si ) , si (29) (30) : if (p a), : if (p b), : otherwise. (23) In our notation J (·, ·) stands for the operator of precision propagation defined as follows: 0 : if (a = b = 0), (31) J (a, b) = ab : otherwise. a+b As in case of BOSS, to prevent exponential growth of the composed RPPs we apply discretization to all domains by appropriate grids. Grid Dr and corresponding discretization function r are defined exactly as in BOSS. To define Dp we use its pro jection D to the p [0, 1] interval (D is defined exactly as Dp in the BOSS p case): Dp = {p : loga,b (p) D }, (32) p We express the discretization function p through its pro jected version (which is defined as p in BOSS): p p (p) = (loga,b (p)). p (33) Theorem 3 (Hardness of GOSS). Finding an exact solution for a general instance of the GOSS problem is NP -hard even for a tree-shaped GBN. Proof is by reduction from Knapsack. A polynomial scheme for approximate solution of GOSS, similar to one presented for BOSS, follows. To apply the RLC algorithm we specify the problem in terms of a composite system. The CG is as for BOSS. All domain sets (except for Ms , that remain the same) should be redefined as follows: O Ps = {P rec(Xs |Es ) : E Y}, Grid Df and the corresponding discretization function f are similarly defined. Equivalence operators , , and are defined as in BOSS. The composed equivalence operator is: (f , r) (f , r ) (f f ) (r r ) The comprehensive utility function is as follows: r P : if ((p 1 ) ( t B )), ^ ^ U (t, p, (f , r)) = - : otherwise (34) After compiling the composite system (using the RLC algorithm), a near-optimal TAS can be selected from the resulting RPP Qc w.r.t. this utility function. Let 1 E denote an optimal observation subset, and E a subset corresponding to the TAS selected from Qc . 1 Q F R Q P F R (24) (25) (26) (27) Fs = : E Y }, Rs = {Rs (Xs |E ) : E Y}, Qs = Fs × Rs . I {P rec(Xs |Es ) We need to reformulate, in the definition of RPPs, the specification of the s functions. For the leaf nodes s is defined as follows: s (p, m) = (f , r), f= I P rec(Xs |Es ) (28) = m s , l oga,b (p + f ) : if (Xs XH ), r = Rs (Xs |E ) = 1 : otherwise, Theorem 4 (Approximation quality of RLC for GOSS). The RLC algorithm applied to a transformed instance of GOSS problem with out-tree topology approximates the optimal solution E within additive factor of u , bounded as fol lows: ^ ^ u = R(XH |E ) - R(XH |E ) hp + hf + r (35) To ensure approximation with u (in worst case) we can select p = f = h , and r = . Any composed RPP Qc can be represented by a 3s dimensional (Dp × Df × Dr ) table with a number of entries bounded as follows: h2 1. c (36) |Qs | dp df dr = The appropriate worst-case complexity fon the t omr c h2 plete run of the RLC algorithm is O 2 ime, t h2 s n c+ 1 h pace for a chain topology, and O c+2 c O 3 h c+ 2 s ime, O c+2c pace for a tree with maximum branching factor of c. [2] A. Cassandra, M. L. Littman, and N. L. Zhang. Incremental Pruning: A simple, fast, exact method for partially observable Markov decision processes. In Proceedings of (UAI­97), pages 54­ 61, 1997. [3] E. Horvitz. Reasoning about beliefs and actions under computational resource constraints. In UAI, pages 301­324, 1987. [4] E. Horvitz. Models of continual computation. In AAAI/IAAI, pages 286­293, 1997. [5] E. Horvitz. Continual computation policies for allocating offline and real-time resources. In IJCAI, pages 1280­1287, 1999. [6] A. Krause and C. Guestrin. Near-optimal nonmyopic value of information in graphical models. In UAI, pages 324­331, 2005. [7] A. Krause and C. Guestrin. Optimal nonmyopic value of information in graphical models - efficient algorithms and theoretical limits. In IJCAI, pages 1339­1345, 2005. [8] A. I. Mouaddib and S. Zilberstein. Knowledgebased anytime computation. In IJCAI, pages 775­783, 1995. [9] Y. Radovilsky, G. Shattah, and E. S. Shimony. Efficient deterministic approximation algorithm for nonmyopic value of information in graphical models. In SMC Conference, Taipei, Taiwan, 2006. [10] S. Zilberstein. Operational rationality through compilation of anytime algorithms. Technical report, Computer Science Division, University of California at Berkeley, 1993. PhD Dissertation. [11] S. Zilberstein. Optimizing decision quality with contract algorithms. In IJCAI, pages 1576­1582, 1995. [12] S. Zilberstein. Using anytime algorithms in intelligent systems. AI Magazine, 17(3):73­83, 1996. 5 Summary In this paper we extended the concept of CPP, and presented an efficient technique for compiling a composite system beyond the input monotonicity assumption. The extended scheme has been applied to optimizing a set of measurements in two different settings (for choosing a maximum expectation variable in a binary valued BN, and for minimizing the worst variance in a Gaussian BN). Polynomial time methods have been presented for both problems, and quality of approximation has been theoretically determined. Applying our framework to real-world domains as an empirical evaluation is underway. Further extending the framework to deal with more general system topologies, tractable strategies for active monitoring are possible directions for future research. Acknowledgements Partially supported by the IMG4 consortium (under the Ministry of Industry, Trade and Labor of Israel MAGNET program), by the Lynn and William Frankel Center for Computer Sciences, and by the Paul Ivanier Center for Robotics. References [1] M. Boddy and T. Dean. Solving time-dependent planning problems. In IJCAI, pages 979­984, 1989.