index
int64 0
20.3k
| text
stringlengths 0
1.3M
| year
stringdate 1987-01-01 00:00:00
2024-01-01 00:00:00
| No
stringlengths 1
4
|
|---|---|---|---|
2,000
|
MIME: Mutual Information Minimization and Entropy Maximization for Bayesian Belief Propagation Anand Rangarajan Dept. of Computer and Information Science and Engineering University of Florida Gainesville, FL 32611-6120, US anand@cise.ufl.edu Alan L. Yuille Smith-Kettlewell Eye Research Institute 2318 Fillmore St. San Francisco, CA 94115, US yuille@ski.org Abstract Bayesian belief propagation in graphical models has been recently shown to have very close ties to inference methods based in statistical physics. After Yedidia et al. demonstrated that belief propagation fixed points correspond to extrema of the so-called Bethe free energy, Yuille derived a double loop algorithm that is guaranteed to converge to a local minimum of the Bethe free energy. Yuille’s algorithm is based on a certain decomposition of the Bethe free energy and he mentions that other decompositions are possible and may even be fruitful. In the present work, we begin with the Bethe free energy and show that it has a principled interpretation as pairwise mutual information minimization and marginal entropy maximization (MIME). Next, we construct a family of free energy functions from a spectrum of decompositions of the original Bethe free energy. For each free energy in this family, we develop a new algorithm that is guaranteed to converge to a local minimum. Preliminary computer simulations are in agreement with this theoretical development. 1 Introduction In graphical models, Bayesian belief propagation (BBP) algorithms often (but not always) yield reasonable estimates of the marginal probabilities at each node [6]. Recently, Yedidia et al. [7] demonstrated an intriguing connection between BBP and certain inference methods based in statistical physics. Essentially, they demonstrated that traditional BBP algorithms can be shown to arise from approximations of the extrema of the Bethe and Kikuchi free energies. Next, Yuille [8] derived new double-loop algorithms which are guaranteed to minimize the Bethe and Kikuchi energy functions while continuing to have close ties to the original BBP algorithms. Yuille’s approach relies on a certain decomposition of the Bethe and Kikuchi free energies. In the present work, we begin with a new principle—pairwise mutual information minimization and marginal entropy maximization (MIME)—and derive a new energy function which is shown to be equivalent to the Bethe free energy. After demonstrating this connection, we derive a family of free energies closely related to the MIME principle which also shown to be equivalent, when constraint satisfaction is exact, to the Bethe free energy. For each member in this family of energy functions , we derive a new algorithm that is guaranteed to converge to a local minimum. Moreover, the resulting form of the algorithm is very simple despite the somewhat unwieldy nature of the algebraic development. Preliminary comparisons of the new algorithm with BBP were carried out on spin glass-like problems and indicate that the new algorithm is convergent when BBP is not. However, the effectiveness of the new algorithms remains to be seen. 2 Bethe free energy and the MIME principle In this section, we show that the Bethe free energy can be interpreted as pairwise mutual information minimization and marginal entropy maximization. The Bethe free energy for Bayesian belief propagation is written as FBethe({pij, pi, γij, λij}) = P ij:i>j P xi,xj pij(xi, xj) log pij(xi,xj) φij(xi,xj) −P i(ni −1) P xi pi(xi) log pi(xi) ψi(xi) + P ij:i>j P xj λij(xj)[P xi pij(xi, xj) −pj(xj)] + P ij:i>j P xi λji(xi)[P xj pij(xi, xj) −pi(xi)] + P ij:i>j γij(P xi,xj pij(xi, xj) −1) (1) where φij(xi, xj) def = ψij(xi, xj)ψi(xi)ψj(xj) and ni is the number of neighbors of node i. Link functions ψij > 0 are available relational data between nodes i and j. The singleton function ψi is also available at each node i. The double summation P ij:i>j is carried out only over the nodes that are connected. The Lagrange parameters {λij, γij} are needed in the Bethe free energy (1) to satisfy the following constraints relating the joint probabilities {pij} with the marginals {pi}: X xi pij(xi, xj) = pj(xj), X xj pij(xi, xj) = pi(xi), and X xi,xj pij(xi, xj) = 1. (2) The pairwise mutual information is defined as MIij = X xi,xj pij(xi, xj) log pij(xi, xj) pi(xi)pj(xj) (3) The mutual information is minimized when the joint probability pij(xi, xj) = pi(xi)pj(xj) or equivalently when nodes i and j are independent. When nodes i and j are connected via a non-separable link ψij(xi, xj) they will not be independent. We now state the MIME principle. Statement of the MIME principle: Maximize the marginal entropy and minimize the pairwise mutual information using the available marginal and pairwise link function expectations while satisfying the joint probability constraints. The pairwise MIME principle leads to the following free energy: FMIME({pij, pi, γij, λij}) = P ij:i>j P xi,xj pij(xi, xj) log pij(xi,xj) pi(xi)pj(xj) + P i P xi pi(xi) log pi(xi) −P ij:i>j P xi,xj pij(xi, xj) log ψij(xi, xj) −P i P xi pi(xi) log ψi(xi) + P ij:i>j P xj λij(xj)[P xi pij(xi, xj) −pj(xj)] + P ij:i>j P xi λji(xi)[P xj pij(xi, xj) −pi(xi)] + P ij:i>j γij(P xi,xj pij(xi, xj) −1). (4) In the above free energy, we minimize the pairwise mutual information and maximize the marginal entropies. The singleton and pairwise link functions are additional information which do not allow the system to reach its “natural” equilibrium—a uniform i.i.d. distribution on the nodes. The Lagrange parameters enforce the constraints between the pairwise and marginal probabilities. These constraints are the same as in the Bethe free energy (1). Note that the Lagrange parameter terms vanish if the constraints in (2) are exactly satisfied. This is an important point when considering equivalences between different energy functions. Lemma 1 Provided the constraints in (2) are exactly satisfied, the MIME free energy in (4) is equivalent to the Bethe free energy in (1). Proof: Using the fact that constraint satisfaction is exact and using the identity pij(xi, xj) = pji(xj, xi), we may write − X ij:i>j X xi,xj pij(xi, xj) log pi(xi)pj(xj) = − X ij:i̸=j X xi,xj pij(xi, xj) log pi(xi) = − X i ni X xi pi(xi) log pi(xi), and X ij:i>j X xi,xj pij(xi, xj) log ψi(xi)ψj(xj) = X i ni X xi pi(xi) log ψi(xi). (5) We have shown that a marginal entropy term emerges from the mutual information term in (4) when constraint satisfaction is exact. Collecting the marginal entropy terms together and rearranging the MIME free energy in (4), we get the Bethe free energy in (1). 3 A family of decompositions of the Bethe free energy Recall that the Bethe free energy and the energy function resulting from application of the MIME principle were shown to be equivalent. However, the MIME energy function is merely one particular decomposition of the Bethe free energy. As Yuille mentions [8], many decompositions are possible. The main motivation for considering alternative decompositions is for algorithmic reasons. We believe that certain decompositions may be more effective than others. This belief is based on our previous experience with closely related deterministic annealing algorithms [3, 2]. In this section, we derive a family of free energies that are equivalent to the Bethe free energy provided constraint satisfaction is exact. The family of free energies is inspired by and closely related to the MIME free energy in (4). Lemma 2 The following family of energy functions indexed by the free parameters δ > 0 and {ξi} is equivalent to the original Bethe free energy (1) provided the constraints in (2) are exactly satisfied and the parameters q and r are set to {qi = (1 −δ)ni} and {ri = 1 −niξi} respectively. Fequiv({pij, pi, γij, λij}) = P ij:i>j P xi,xj pij(xi, xj) log pij(xi,xj) [P xj pij(xi,xj)]δ[P xi pij(xi,xj)]δ + P i P xi pi(xi) log pi(xi) −P i qi P xi pi(xi) log pi(xi) −P ij:i>j P xi,xj pij(xi, xj) log ψij(xi, xj)ψξi i (xi)ψξj j (xj) −P i ri P xi pi(xi) log ψi(xi) + P ij:i>j P xj λij(xj)[P xi pij(xi, xj) −pj(xj)] + P ij:i>j P xi λji(xi)[P xj pij(xi, xj) −pi(xi)] + P ij:i>j γij(P xi,xj pij(xi, xj) −1). (6) In (6), the first term is no longer the pairwise mutual information as in (4). And unlike (4), pi(xi) no longer appears in the pairwise mutual information-like term. Proof: We selectively substitute P xi pij(xi, xj) = pj(xj) and P xj pij(xi, xj) = pi(xi) to show the equivalence. First X ij:i>j X xi,xj pij(xi, xj) log[ X xj pij(xi, xj)]δ[ X xi pij(xi, xj)]δ = δ X i ni X xj pi(xi) log pi(xi), X ij:i>j X xi,xj pij(xi, xj) log ψ ξi i (xi)ψ ξj j (xj) = X i niξi X xj pi(xi) log ψi(xi). (7) Substituting the identities in (7) into (6), we see that the free energies are algebraically equivalent. 4 A family of algorithms for belief propagation We now derive descent algorithms for the family of energy functions in (6). All the algorithms are guaranteed to converge to a local minimum of (6) under mild assumptions regarding the number of fixed points. For each member in the family of energy functions, there is a corresponding descent algorithm. Since the form of the free energy in (6) is complex and precludes easy minimization, we use algebraic (Legendre) transformations [1] to simplify the optimization. − X xj pij(xi, xj) log X xj pij(xi, xj) = minσji(xi) −P xj pij(xi, xj) log σji(xi) + σji(xi) −P xj pij(xi, xj) − X xi pij(xi, xj) log X xi pij(xi, xj) = minσij(xj) −P xi pij(xi, xj) log σij(xj) + σij(xj) −P xi pij(xi, xj) −pi(xi) log pi(xi) = min ρi(xi) −pi(xi) log ρi(xi) + ρi(xi) −pi(xi). (8) We now apply the above algebraic transforms. The new free energy is (after some algebraic manipulations) Fequiv({pij, pi, σij, ρi, γij, λij}) = X ij:i>j X xi,xj pij(xi, xj) log pij(xi, xj) σδ ji(xi)σδ ij(xj) +δ X ij:i̸=j X xi σij(xj) + X i X xi pi(xi) log pi(xi) ρqi i (xi) + X i qi X xi ρi(xi) − X ij:i>j X xi,xj pij(xi, xj) log ψij(xi, xj)ψ ξi i (xi)ψ ξj j (xj) − X i ri X xi pi(xi) log ψi(xi) + X ij:i>j X xj λij(xj)[ X xi pij(xi, xj) −pj(xj)] + X ij:i>j X xi λji(xi)[ X xj pij(xi, xj) −pi(xi)] + X ij:i>j γij( X xi,xj pij(xi, xj) −1). (9) We continue to keep the parameters {qi} and {ri} in (9). However, from Lemma 2, we know that the equivalence of (9) to the Bethe free energy is predicated upon appropriate setting of these parameters. In the rest of the paper, we continue to use q and r for the sake of notational simplicity. Despite the introduction of new variables via Legendre transforms, the optimization problem in (9) is still a minimization problem over all the variables. The algebraically transformed energy function in (9) is separately convex w.r.t. {pij, pi} and w.r.t. {σij, ρi} provided δ ∈[0, 1]. Since the overall energy function is not convex w.r.t. all the variables, we pursue an alternating algorithm strategy similar to the double loop algorithm in Yuille [8]. The basic idea is to separately minimize w.r.t. the variables {σij, ρi} and the variables {pij, pi}. The linear constraints in (2) are enforced when minimizing w.r.t the latter and do not affect the convergence properties of the algorithm since the energy function w.r.t. {pij, pi} is convex. We evaluate the fixpoints of {σij, ρi}. Note that (9) is convex w.r.t. {σij, ρi}. σij(xj) = X xi pij(xi, xj), σji(xi) = X xj pij(xi, xj), and ρi(xi) = pi(xi). (10) The fixpoints of {pij, pi} are evaluated next. Note that (9) is convex w.r.t. {pij, pi}. pij(xi, xj) = σδ ji(xi)σδ ij(xj)ψij(xi, xj)ψξi i (xi)ψξj j (xj)e−λij(xj)−λji(xi)−γij−1 pi(xi) = ρqi i (xi)ψri i (xi)e P k λki(xi)−1. (11) The constraint satisfaction equations from (2) can be rewritten as X xj pij(xi, xj) = pi(xi) ⇒ e2λji(xi) = P xj σδ ji(xi)σδ ij(xj)ψij(xi,xj)ψξi i (xi)ψ ξj j (xj)e−λij (xj )−γij −1 ρqi i (xi)ψri i (xi)e P k̸=j λki(xi)−1 (12) Similar relations can be obtained for the other constraints in (2). Consider a Lagrange parameter update sequence where the Lagrange parameter currently being updated is tagged as “new” with the rest designated as “old.” We can then rewrite the Lagrange parameter updates using “old” and “new” values. Please note that each Lagrange parameter update corresponds to one of the constraints in (2). It can be shown that the iterative update of the Lagrange parameters is guaranteed to converge to the unique solution of (2) [8]. While rewriting (12), we multiply the left and right sides with e−2λold ji (xi). e2λnew ji (xi)−2λold ji (xi) = P xj σδ ji(xi)σδ ij(xj)ψij(xi,xj)ψ ξi i (xi)ψ ξj j (xj)e −λold ij (xj )−λold ji (xi)−γold ij −1 ρ qi i (xi)ψ ri i (xi)e P k λold ki (xi)−1 . (13) Using (11), we relate each Lagrange parameter update with an update of pij(xi, xj) and pi(xi). We again invoke the “old” and “new” designations, this time on the probabilities. From (11), (12) and (13), we write the joint probability update pnew ij (xi, xj) pold ij (xi, xj) = e−λnew ji (xi)+λold ji (xi) = s pold i (xi) P xj pold ij (xi, xj) (14) and for the marginal probability update pnew i (xi) pold i (xi) = eλnew ji (xi)−λold ji (xi) = sP xj pold ij (xi, xj) pold i (xi) . (15) From (14) and (15), the update equations for the probabilities are pnew ij (xi, xj) = pold ij (xi, xj) s pold i (xi) P xj pold ij (xi, xj), pnew i (xi) = s pold i (xi) X xj pold ij (xi, xj) (16) With the probability updates in place, we may write down new algorithms minimizing the family of Bethe equivalent free energies using only probability updates. The update equations (16) can be seen to satisfy the first constraint in (2). Similar update equations can be derived for the other constraints in (2). For each Lagrange parameter update, an equivalent, simultaneous probability (joint and marginal) update can be derived similar to (16). The overall family of algorithms can be summarized as shown in the pseudocode. Despite the unwieldy algebraic development preceding it, the algorithm is very simple and straightforward. Set free parameters δ ∈[0, 1] and {ξi}. Initialize {pij, pi}. Set {qi = (1 −δ)ni} and {ri = 1 −niξi}. Begin A: Outer Loop σij(xj) ←P xi pij(xi, xj) σji(xi) ←P xj pij(xi, xj) ρi(xi) ←pi(xi) pij(xi, xj) ←σδ ji(xi)σδ ij(xj)ψij(xi, xj)ψξi i (xi)ψξj j (xj) pi(xi) ←ρqi i (xi)ψri i (xi) Begin B: Inner Loop: Do B until 1 N P ij:i>j[(P xj pij(xi, xj) − pi(xi))2 + (P xi pij(xi, xj) −pj(xj))2] < cthr Simultaneously update pij(xi, xj) and pi(xi) below. pij(xi, xj) ←pij(xi, xj) r pi(xi) P xj pij(xi,xj) pi(xi) ← q pi(xi) P xj pij(xi, xj) Simultaneously update pij(xi, xj) and pj(xj) below. pij(xi, xj) ←pij(xi, xj) r pj(xj) P xi pij(xi,xj) pj(xj) ← q pj(xj) P xi pij(xi, xj) Normalize pij(xi, xj). pij(xi, xj) ← pij(xi,xj) P xi,xj pij(xi,xj) End B End A In the above family of algorithms, the MIME algorithm corresponds to free parameter settings δ = 1 and ξi = 0 which in turn lead to parameter settings qi = 0 and ri = 1. The Yuille [8] double loop algorithm corresponds to the free parameter settings δ = 0 and ξi = 0 which in turn leads to parameter settings qi = ni and ri = 1. A crucial point is that the energy function for every valid parameter setting is equivalent to the Bethe free energy provided constraint satisfaction is exact. The inner loop constraint satisfaction threshold parameter cthr setting is very important in this regard. We are obviously not restricted to the MIME parameter settings. At this early stage of exploration of the inter-relationships between Bayesian belief propagation and inference methods based in statistical physics [7], it is premature to speculate regarding the “best” parameter settings for δ and {ξi}. Most likely, the effectiveness of the algorithms will vary depending on the problem setting which enters into the formulation via the link functions {ψij} and the singleton functions {ψi}. 5 Results We implemented the family of algorithms in C++ and conducted tests on locally connected 50 node graphs and binary state variables. The ψi(xi) and ψij(xi, xj) are of the form e±hi and e±hij where hi and hij are drawn from uniform distributions (in the interval [−1, 1]). Provided the constraint satisfaction theshold parameter cthr was set low enough, the algorithm (for δ = 1 and other parameter settings as described in Figure 1) exhibited monotonic convergence. Figure 2 shows the number of inner loop iterations corresponding to different settings of the constraint satisfaction threshold parameter. We also implemented the BBP algorithm and empirically observed that it often did not converge for these graphs. These results are quite preliminary and far more validation experiments are required. However, they provide a proof of concept for our approach. 6 Conclusion We began with the MIME principle and showed the equivalence of the MIMEbased free energy to the Bethe free energy assuming constraint satisfaction to be exact. Then, we derived new decompositions of the Bethe free energy inspired by the MIME principle, and driven by our belief that certain decompositions may be more effective than others. We then derived a convergent algorithm for each member in the family of MIME-based decompositions. It remains to be seen if the MIME-based algorithms are efficient for a reasonable class of problems. While the MIME-based algorithms derived here use closed-form solutions in the constraint satisfaction inner loop, it may turn out that the inner loop is better handled using preconditioned gradient-based descent algorithms. And it is important to explore the inter-relationships between the convergent MIME-based descent algorithms and other recent related approaches with interesting convergence properties [4, 5]. References [1] E. Mjolsness and C. Garrett. Algebraic transformations of objective functions. Neural Networks, 3:651–669, 1990. [2] A. Rangarajan. Self annealing and self annihilation: unifying deterministic annealing and relaxation labeling. Pattern Recognition, 33:635–649, 2000. [3] A. Rangarajan, S. Gold, and E. Mjolsness. A novel optimizing network architecture with applications. Neural Computation, 8(5):1041–1060, 1996. [4] Y. W. Teh and M. Welling. Passing and bouncing messages for generalized inference. Technical Report GCNU 2001-01, Gatsby Computational Neuroscience Unit, University College, London, 2001. [5] M. Wainwright, T. Jaakola, and A. Willsky. Tree-based reparameterization framework for approximate estimation of stochastic processes on graphs with cycles. Technical Report LIDS P-2510, MIT, Cambridge, MA, 2001. [6] Y. Weiss. Correctness of local probability propagation in graphical models with loops. Neural Computation, 12:1–41, 2000. [7] J. S. Yedidia, W. T. Freeman, and Y. Weiss. Bethe free energy, Kikuchi approximations and belief propagation algorithms. In Advances in Neural Information Processing Systems 13, Cambridge, MA, 2001. MIT Press. [8] A. L. Yuille. A double loop algorithm to minimize the Bethe and Kikuchi free energies. Neural Computation, 2001. (submitted). 0 500 1000 1500 −0.5 −0.4 iteration MIME energy 0 500 1000 1500 −0.5 −0.4 iteration MIME energy 0 500 1000 1500 −0.5 −0.4 iteration MIME energy (a) (b) (c) Figure 1: MIME energy versus outer loop iteration: 50 node, local topology, δ = 1. Constraint satisfaction threshold parameter cthr was set to (a) 10−8 (b) 10−4 (c) 10−2 0 500 1000 1500 0 2 4 6 8 10 12 14 16 18 20 outer loop iteration index total # of inner loop iterations 0 500 1000 1500 1 2 3 4 5 6 7 outer loop iteration index total # of inner loop iterations 0 500 1000 1500 1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 2 outer loop iteration index total # of inner loop iterations (a) (b) (c) Figure 2: Inner loop iterations versus outer loop: 50 node, local topology, δ = 1. Constraint satisfaction threshold parameter cthr was set to (a) 10−8 (b) 10−4 (c) 10−2
|
2001
|
186
|
2,001
|
Rates of Convergence of Performance Gradient Estimates Using Function Approximation and Bias in Reinforcement Learning Gregory Z. Grudic University of Colorado, Boulder grudic@cs.colorado.edu Lyle H. Ungar University of Pennsylvania ungar@cis.upenn.edu Abstract We address two open theoretical questions in Policy Gradient Reinforcement Learning. The first concerns the efficacy of using function approximation to represent the state action value function, . Theory is presented showing that linear function approximation representations of can degrade the rate of convergence of performance gradient estimates by a factor of relative to when no function approximation of is used, where is the number of possible actions and is the number of basis functions in the function approximation representation. The second concerns the use of a bias term in estimating the state action value function. Theory is presented showing that a non-zero bias term can improve the rate of convergence of performance gradient estimates by
, where is the number of possible actions. Experimental evidence is presented showing that these theoretical results lead to significant improvement in the convergence properties of Policy Gradient Reinforcement Learning algorithms. 1 Introduction Policy Gradient Reinforcement Learning (PGRL) algorithms have recently received attention because of their potential usefulness in addressing large continuous reinforcement Learning (RL) problems. However, there is still no widespread agreement on how PGRL algorithms should be implemented. In PGRL, the agent’s policy is characterized by a set of parameters which in turn implies a parameterization of the agent’s performance metric. Thus if represents a dimensional parameterization of the agent’s policy and is a performance metric the agent is meant to maximize, then the performance metric must have the form [6]. PGRL algorithms work by first estimating the performance gradient (PG) !"#! and then using this gradient to update the agent’s policy using: $&%('*)+$-,. / ! !0 (1) where . is a small positive step size. If the estimate of !"#! is accurate, then the agent can climb the performancegradient in the parameter space, toward locally optimal policies. In practice, !12! is estimated using samples of the state action value function . The PGRL formulation is attractive because 1) the parameterization of the policy can directly imply a generalization over the agent’s state space (e.g., can represent the adjustable weights in a neural network approximation), which suggests that PGRL algorithms can work well on very high dimensional problems [3]; 2) the computational cost of estimating !0"#! is linear in the number of parameters , which contrasts with the computational cost for most RL algorithms which grows exponentially with the dimension of the state space; and 3) PG algorithms exist which are guaranteed to give unbiased estimates of !0"#! [6, 5, 4, 2, 1]. This paper addresses two open theoretical questions in PGRL formulations. In PGRL formulations performance gradient estimates typically have the following form: / !0 ! ) #'
1' #'
(2) where !"
#! is the estimate of the value of executing action $! in state ! (i.e. the state action value function), # %! the bias subtracted from !& &#! in state ! , ' is the number of steps the agent takes before estimating !12!0 , and the form of the function depends on the PGRL algorithm being used (see Section 2, equation (3) for the form being considered here). The effectiveness of PGRL algorithms strongly depends on how !" &#! is obtained and the form of # %! . The aim of this work is to address these questions. The first open theoretical question addressed here is concerned with the use of function approximation (FA) to represent the state action value function , which is in turn used to estimate the performance gradient. The original formulation of PGRL [6], the REINFORCE algorithm, has been largely ignored because of the slow rate of convergence of the PG estimate. The use of FA techniques to represent based on its observations has been suggested as a way of improving convergence properties. It has been proven that specific linear FA formulations can be incorporated into PGRL algorithms, while still guaranteeing convergence to locally optimal solutions [5, 4]. However, whether linear FA representations actually improves the convergence properties of PGRL is an open question. We present theory showing that using linear basis function representations of , rather than direct observations of it, can slow the rate of convergence of PG estimates by a factor of (see Theorem 1 in Section 3.1). This result suggests that PGRL formulations should avoid the use of linear FA techniques to represent . In Section 4, experimental evidence is presented supporting this conjecture. The second open theoretical question addressed here is can a non-zero bias term # in (2) improve the convergence properties of PG estimates? There has been speculation that an appropriate choice of # can improve convergence properties [6, 5], but theoretical support has been lacking. This paper presents theory showing that if # ) #)(* + &1 , where is the number actions, then the rate of convergence of the PG estimate is improved by # (see Theorem 2 in Section 3.2). This suggests that the convergence properties of PGRL algorithms can be improved by using a bias term that is the average of values in each state. Section 4 gives experimental evidence supporting this conjecture. 2 The RL Formulation and Assumptions The RL problem is modeled as a Markov Decision Process (MDP). The agent’s state at time , . /
0$ 12 is given by $ 43 , 365 87 . At each time step the agent chooses from a finite set of :9 actions $ <; ) ' 1
#= and receives a reward > $ . The dynamics of the environment are characterized by transition probabilities ? * @"@A )B?C>+- $&%(' )D A&E $ ) + & $ )FG2 and expected rewards H * @ )JIK-> $&%(' E $ )L+ & $ )MG2 , NO+
A 3P & ; . The policy followed by the agent is characterized by a parameter vector , and is defined by the probability distribution Q + &R )S?C>+-% $ )D E $ )S+RT2 , NO U3P & U; . We assume that Q + &R is differentiable with respect to . We use the Policy Gradient Theorem of Sutton et al. [5] and limit our analysis to the start state discount reward formulation. Here the reward function Q and state action value function + &1 are defined as: Q )I ( $ ' $ > $ / &Q
+
)I ( (' ' > $&% $ ) + & $ ) "Q
where . Then the exact expression for the performance gradient is: ! ! ) @ = !(' !GQ + & ! R ! + & ! (3) where ) ( $ $ $ ) E "Q2 and # . This policy gradient formulation requires that the state-action value function, , under the current policy be estimated. This estimate, , is derived using the observed value ! @ + & ! . We assume that ! @ + & ! has the following form: ! @ /
#! ) + &#! ,#" + &#! where " +
#! has zero mean and finite variance $&% @!' *)( . Therefore, if +
! is an estimate of /
#! obtained by averaging * observations of ! @ + &#! , then the mean and variance are given by: IB /
#! ) + &#! O ,+B + &#! ).-0/ 132 4 ( 5 (4) In addition, we assume that ! @ /
! are independently distributed. This is consistent with the MDP assumption. 3 Rate of Convergence Results Before stating the convergence theorems, we define the following: $6% 798;: ) <>=0? @!@AB' ! @C ' 'EDEDEDE' =GF $6% @!' * ( H$6% 79I J ) <GKML @N@A' ! @C ' 'EDEDEDE' =>F $6% @!' * ( (5) where $ % @!' * ( is defined in (4) and O 79I J )QP ( @ % = ( !M(' SR BT @!' *)(VU W!X R W %ZY $ % 79I J O 798;: )QP ( @ % = ( !M(' R BT @!' * ( U W!X R W %ZY $6% 798;: (6) 3.1 Rate of Convergence of PIFA Algorithms Consider the PIFA algorithm [5] which uses a basis function representation for estimated state action value function, , of the following form: /
#! ) *)( )\[ ] '&^ *)( ' ]`_ *)( ' ] (7) where ^ * (3' ] are weights and _ * (3' ] are basis functions defined in + 7 . If the weights ^ * (a' ] are chosen based using the observed ! @ + & ! , and the basis functions, _ *)( ' ] , satisfy the conditions defined in [5, 4], then the performance gradient is given by: ! !cb ) @ = !(' !Q + & ! R ! * ( (8) The following theorem establishes bounds on the rate of convergencefor this representation of the performance gradient. Theorem 1: Let R R W b be an estimate of (8) obtained using the PIFA algorithm and the basis function representation (7). Then, given the assumptions defined in Section 2 and equations (5) and (6), the rate of convergence of a PIFA algorithm is bounded below and above by: O 79I J * + ! ! b O 798 : * (9) where is the number of basis functions, is the number of possible actions, and * is the number of independent estimates of the performance gradient. Proof: See Appendix. 3.2 Rate of Convergence of Direct Sampling Algorithms In the previous section, the observed ! @ /
#! are used to build a linear basis function representation of the state action value function, + & ! , which is in turn used to estimate the performance gradient. In this section we establish rate of convergence bounds for performance gradient estimates that directly use the observed @ + &#! without the intermediate step of building the FA representation. These bounds are established for the conditions # ) # ( * /
1 and # ) in (3). Theorem 2: Let / R R W be a estimate of (3), be obtained using direct samples of . Then, if #
) , and given the assumptions defined in Section 2 and equations (5) and (6), the rate of convergence of / R R W is bounded by: O 79I J * + / !0 ! O 798;: * (10) where * is the number of independent estimates of the performance gradient. If # ) is defined as: ) = ' + & (11) then the rate of convergence of the performance gradient R R W is bounded by: O 79I J * + ! !0 O 798 : *
(12) where is the number of possible actions. Proof: See Appendix. Thus comparing (12) and (10) to (9) one can see that policy gradient algorithms such as PIFA which build FA representations of converge by a factor of slower than algorithms which directly sample . Furthermore, if the bias term is as defined in (11), the bounds on the variance are further reduced by . In the next section experimental evidence is given showing that these theoretical consideration can be used to improve the convergence properties of PGRL algorithms. 4 Experiments The Simulated Environment: The experiments simulate an agent episodically interacting in a continuous two dimensional environment. The agent starts each episode in the same state ! , and executes a finite number of steps following a policy to a fixed goal state . The stochastic policy is defined by a finite set of Gaussians, each associated with a specific 0 20 40 60 80 100 −0.1 0 0.1 0.2 0.3 0.4 0.5 0.6 Number of Policy Updates ρ(π) Biased Q No Bias Linear FA Q a) Convergence of Algorithms 0 2 4 6 8 10 12 14 10 0 10 1 10 2 10 3 10 4 Number of Possible Actions (M) V[∂ ρ / ∂ θF] / V[∂ ρ / ∂ θ] b) + / R R W b + / R R W 0 2 4 6 8 10 12 14 10 0 10 1 10 2 Number of Possible Actions (M) V[∂ ρ / ∂ θ] / V[∂ ρ / ∂ θb] c) + / R R W + / R R W Figure 1: Simulation Results action. The Gaussian associated with action is defined as: ) Z? 7 ('
% where ) #' 1
7 87 , is the agents state, *'%
7 is the Gaussian center, and *' 1 7 is the variance along each state space dimension. The probability of executing action in state is Q + & R ) = ( (' where ) '' 1 ' ' ' 1 ' 1 1= ' 1 1= 7 = ' = 7 defines the policy parameters that dictate the agent’s actions. Action ' directs the agent toward the goals state , while the remaining actions (for ) 0T ) direct the agent towards the corresponding Gaussian center *'%
7 . Noise is modeled using a uniform random distribution between $ denoted by $ , such that the noise in dimension is given by: ! @ ) , `) #$2 where 9 is the magnitude of the noise, @ is the state the agent observes and uses to choose actions, and is the actual state of the agent. The agent receives a reward of +1 when it reaches the goal state, otherwise it receives a reward of: > ) ) Z? 7 (' % Thus the agent gets negative rewards the closer it gets to the origin of the state space, and a positive reward whenever it reaches the goal state. Implementation of the PGRL algorithms: All the PGRL formulations studied here require observations (i.e. samples) of the state action value function. @ + & ! is sampled by executing action T! in state and thereafter following the policy. In the episodic formulation, where the agent executes a maximum of ' steps during each episode, at the end of each episode, @ $ & $ for step , can be evaluated as follows: @ $ & $ ) (' ' > $&% E $ )+
$ ) &Q Thus, given that the agent executes a complete episode '
1'
#( following the policy Q , at the completion of the episode we can calculate @ ' & ' 1 @ & . This gives samples of ' state action value pairs. Equation (3) tells us that we require a total of ' state action value function observations to estimate a performance gradient (assuming the agent can execute actions). Therefore, we can obtain the remaining + ' observations of @ by sending the agent out on ' epsisodes, each time allowing it to follow the policy Q for all ' steps, with the exception that action $ )U is executed when ! @ $ & is being observed. This sampling procedure requires a total of ' episodes and gives a complete set of @ state action pairs for any path ' & ' 1
. For the direct sampling algorithms in Section 3.2, these observations are directly used to estimate the performance gradient. For the linear basis function based PGRL algorithm in Section 3.1, these observations are first used to calculate the ^ * ( ' ] as defined in [5, 4], and then the performance gradient is calculated using (8). Experimental Results: Figure 1b shows a plot of average + / !"#! b + / !12!0 values over 10,000 estimates of the performance gradient. For each estimate, the goal state, start state, and Gaussian centers are all chosen using a uniform random distribution + ; the Gaussian variances are sampled from a uniform distribution $ / . As predicted by Theorem 1 in Section 3.1 and Theorem 2 in Section 3.2, as the number of actions increases, this ratio also increases. Note that Figure 1b plots average variance ratios, not the bounds in variance given in Theorem 1 and Theorem 2 (which have not been experimentally sampled), so the ratio predicted by the theorems is supported by the increase in the ratio as increases. Figure 1c shows a plot of average + / !12!0 + / !"#! values over 10,000 estimates of the performance gradient. As above, for each estimate, the goal state, start state, and Gaussian centers are all chosen using a uniform random distribution + ; the Gaussian variances are sampled from a uniform distribution `) + . This also follows the predicted trends of Theorem 1 and Theorem 2. Finally, Figure 1a shows the average reward over 100 runs as the three algorithms converge on a two action problem. Each algorithm is given the same number of ! @ samples to estimate the gradient before each update. Because / !0"#! has the least variance, it allows the policy Q to converge to the highest reward value Q . Similarly, because / !12!0 b has the highest variance, its policy updates converge to the worst Q . Note that because all three algorithms will converge to the same locally optimal policy given enough samples of ! @ , Figure 1a simply demonstrates that / !12! b requires more samples than / !1#! , which in turn requires more samples than / !"#! . 5 Conclusion The theoretical and experimental results presented here indicate that how PGRL algorithms are implemented can substantially affect the number of observations of the state action value function ( ) needed to obtain good estimates of the performance gradient. Furthermore, they suggest that an appropriately chosen bias term, specifically the average value of over all actions, and the direct use of observed values can improve the convergence of PGRL algorithms. In practice linear basis function representations of can significantly degrade the convergence properties of policy gradient algorithms. This leaves open the question of whether any (i.e. nonlinear) function approximation representation of value functions can be used to improve convergence of such algorithms. References [1] Jonathan Baxter and Peter L. Bartlett, Reinforcement learning in pomdp’s via direct gradient ascent, Proceedings of the Seventeenth International Conference on Machine Learning (ICML’2000) (Stanford University, CA), June 2000, pp. 41–48. [2] G. Z. Grudic and L. H. Ungar, Localizing policy gradient estimates to action transitions, Proceedings of the Seventeenth International Conference on Machine Learning, vol. 17, Morgan Kaufmann, June 29 - July 2 2000, pp. 343–350. [3] , Localizing search in reinforcement learning, Proceedings of the Seventeenth National Conference on Artificial Intelligence, vol. 17, Menlo Park, CA: AAAI Press / Cambridge, MA: MIT Press, July 30 - August 3 2000, pp. 590–595. [4] V. R. Konda and J. N. Tsitsiklis, Actor-critic algorithms, Advances in Neural Information Processing Systems (Cambridge, MA) (S. A. Solla, T. K. Leen, and K.-R. Mller, eds.), vol. 12, MIT Press, 2000. [5] R. S. Sutton, D. McAllester, S. Singh, and Y. Mansour, Policy gradient methods for reinforcement learning with function approximation, Advances in Neural Information Processing Systems (Cambridge, MA) (S. A. Solla, T. K. Leen, and K.-R. Mller, eds.), vol. 12, MIT Press, 2000. [6] R. J. Williams, Simple statistical gradient-following algorithms for connectionist reinforcement learning, Machine Learning 8 (1992), no. 3, 229–256. Appendix: Proofs of Theorems 1 and 2 Proof of Theorem 1: Consider the definition of * ( given in (7). In [5] it is shown that there exist ^ *Z( ' ] and _ *)( ' ] such that: I @ = !(' !Q + & ! R ! * ( @ = !M(' !Q + & ! R ! + & ! ) (13) Let R R W ! @ be the observation of R R W (3) after a single episode. Using (13), we get the following: R R W ! @ ) ( @ = ( !(' R BT @N' * ( U W!X R W ! @ /
#! R R W , " ) P ( @ = ( !M ' R BT @!' *Z( U W;X R W + & ! Y , " ) P ( @ = ( !M ' R BT @!' * ( U W;X R W *)( Y ,#" ) P ( @ = ( !M ' R BT @!' * ( U W;X R W [ ( ] ' ^ *)( ' ]`_ *)( ' ] Y ,#" ) P = ( !M(' [ ( ] (' ^ *Z( ' ] ( @ R BT @!' * ( U W X R W _ *Z( ' ] Y , " = ( !M ' [ ( ] (' ^ *)( ' ] ! ] ,#" where the basis functions ! ] have the form ! ] ) @ !GQ /
! R ! _ * ( ' ] and I " ) , with variance + " )+ P ! ! @ Y ) @ % = !(' !Q + & ! R ! % $ % @!' * ( Denoting R R W b as the least squares (LS) estimate of (3), its form is given by: ! !0 b ) = [ (' (14) where are LS estimates of the weights ^ *)( ' ] and correspond to the basis functions ! ] . Then, it can be shown that any linear system of the type given in (14) has a rate of convergence given by: + ! ! b ) * + " ) * @ % = !(' !GQ + &#!R ! % $ % @!' * ( Substituting (5) and (6) into the above equation completes the proof. Proof of Theorem 2: We prove equation (10) first. For * estimates of the performance gradient, we get * independent samples of each ! @ /
#! . These examples are averaged and therefore: I / ! ! ) @ = !(' !GQ /
! R ! /
! Because each ! @ + & ! is independently distributed, the variance of the estimate is given by + / ! ! ) * @ % = !(' !Q + &#!R ! % $ % @N' *)( (15) Given (5) the worst rate of convergence is bounded by: + / ! ! @ % = !(' !Q + & ! R ! % $ % 798 : * ) O 798 : * A similarly argument applies to the lower bound on convergence completing the proof for (10). Following the same argument for (12), we have + ! ! ) * @ % = !(' !Q + & ! R ! % + +
! = (' +
Where + + & ! ' = = ( (' + & )+ = ' = + & ! ' = = ( (' ! /
) = ' =
% $6% @!' *)( , = ( (' ! ' =
% $6% @N' * (16) Given (5) the variance + on the far left of (16) is bounded by: = ' =
% $6% @!' *)( , = ( (' ! ' =
% $6% @!' * 798;: ) = ' =
% $6% 798;: , = ( (' ! ' =
% $6% 798 : ) = ' =
% , ' =
% $6% 798;: ) ' =
$6% 798 : Plugging the above into (16) and inserting O * from (6) completes the proof for the upper bound. The proof for the lower bound in the variance follows similar reasoning.
|
2001
|
187
|
2,002
|
A Generalization of Principal Component Analysis to the Exponential Family Michael Collins Sanjoy Dasgupta Robert E. Schapire AT&T Labs Research 180 Park Avenue, Florham Park, NJ 07932 mcollins, dasgupta, schapire @research.att.com Abstract Principal component analysis (PCA) is a commonly applied technique for dimensionality reduction. PCA implicitly minimizes a squared loss function, which may be inappropriate for data that is not real-valued, such as binary-valued data. This paper draws on ideas from the Exponential family, Generalized linear models, and Bregman distances, to give a generalization of PCA to loss functions that we argue are better suited to other data types. We describe algorithms for minimizing the loss functions, and give examples on simulated data. 1 Introduction Principal component analysis (PCA) is a hugely popular dimensionality reduction technique that attempts to find a low-dimensional subspace passing close to a given set of points
. More specifically, in PCA, we find a lower dimensional subspace that minimizes the sum of the squared distances from the data points to their projections in the subspace, i.e.,
(1) This turns out to be equivalent to choosing a subspace that maximizes the sum of the squared lengths of the projections , which is the same as the (empirical) variance of these projections if the data happens to be centered at the origin (so that ). PCA also has another convenient interpretation that is perhaps less well known. In this probabilistic interpretation, each point is thought of as a random draw from some unknown distribution "! , where denotes a unit Gaussian with mean #$ . The purpose then of PCA is to find the set of parameters % that maximizes the likelihood of the data, subject to the condition that these parameters all lie in a low-dimensional subspace. In other words, are considered to be noise-corrupted versions of some true points & which lie in a subspace; the goal is to find these true points, and the main assumption is that the noise is Gaussian. The equivalence of this interpretation to the ones given above follows simply from the fact that negative log likelihood under this Guassian model is equal (ignoring constants) to Eq. (1). This Gaussian assumption may be inappropriate, for instance if data is binary-valued, or integer-valued, or is nonnegative. In fact, the Gaussian is only one of the canonical distributions that make up the exponential family, and it is a distribution tailored to real-valued data. The Poisson is better suited to integer data, and the Bernoulli to binary data. It seems natural to consider variants of PCA which are founded upon these other distributions in place of the Gaussian. We extend PCA to the rest of the exponential family. Let be any parameterized set of distributions from the exponential family, where is the natural parameter of a distribution. For instance, a one-dimensional Poisson distribution can be parameterized by , corresponding to mean and distribution
Given data & , the goal is now to find parameters which lie in a low-dimensional subspace and for which the log-likelihood !
is maximized. Our unified approach effortlessly permits hybrid dimensionality reduction schemes in which different types of distributions can be used for different attributes of the data. If the data have a few binary attributes and a few integer-valued attributes, then some coordinates of the corresponding can be parameters of binomial distributions while others are parameters of Poisson distributions. (However, for simplicity of presentation, in this abstract we assume all distributions are of the same type.) The dimensionality reduction schemes for non-Gaussian distributions are substantially different from PCA. For instance, in PCA the parameters , which are means of Gaussians, lie in a space which coincides with that of the data . This is not the case in general, and therefore, although the parameters lie in a linear subspace, they typically correspond to a nonlinear surface in the space of the data. The discrepancy and interaction between the space of parameters and the space of the data is a central preoccupation in the study of exponential families, generalized linear models (GLM’s), and Bregman distances. Our exposition is inevitably woven around these three intimately related subjects. In particular, we show that the way in which we generalize PCA is exactly analogous to the manner in which regression is generalized by GLM’s. In this respect, and in others which will be elucidated later, it differs from other variants of PCA recently proposed by Lee and Seung [7], and by Hofmann [4]. We show that the optimization problem we derive can be solved quite naturally by an algorithm that alternately minimizes over the components of the analysis and their coefficients; thus, the algorithm is reminiscent of Csisz´ar and Tusn´ady’s alternating minization procedures [2]. In our case, each side of the minimization is a simple convex program that can be interpreted as a projection with respect to a suitable Bregman distance; however, the overall program is not generally convex. In the case of Gaussian distributions, our algorithm coincides exactly with the power method for computing eigenvectors; in this sense it is a generalization of one of the oldest algorithms for PCA. Although omitted for lack of space, we can show that our procedure converges in that any limit point of the computed coefficients is a stationary point of the loss function. Moreover, a slight modification of the optimization criterion guarantees the existence of at least one limit point. Some comments on notation: All vectors in this paper are row vectors. If is a matrix, we denote its ’th row by ! and its #" ’th element by $ &% . 2 The Exponential Family, GLM’s, and Bregman Distances 2.1 The Exponential Family and Generalized Linear Models In the exponential family of distributions the conditional probability of a value ' given parameter value takes the following form: ' () +* ' -, '. 0/ (2) Here, is the “natural parameter” of the distribution, and can usually take any value in the reals. / is a function that ensures that the sum (integral) of ' over the domain of ' is 1. From this it follows that / () * ' . We use to denote the domain of ' . The sum is replaced by an integral in the continuous case, where defines a density over . * is a term that depends only on ' , and can usually be ignored as a constant during estimation. The main difference between different members of the family is the form of / . We will see that almost all of the concepts of the PCA algorithms in this paper stem directly from the definition of / . A first example is a normal distribution, with mean and unit variance, which has a density that is usually written as ' () ' . It can be verified that this is a member of the exponential family with * ' ' , , and / . Another common case is a Bernoulli distribution for the case of binary outcomes. In this case . The probability of ' is usually written ' + -
where is a parameter in . This is a member of the exponential family with * ' , () , and / , . A critical function is the derivative / , which we will denote as throughout this paper. By differentiating / * ' , it is easily verified that ' , the expectation of ' under ' . In the normal distribution, ' , and in the Bernoulli case ' . In the general case, ' is referred to as the “expectation parameter”, and defines a function from the natural parameter values to the expectation parameter values. Our generalization of PCA is analogous to the manner in which generalized linear models (GLM’s) [8] provide a unified treatment of regression for the exponential family by generalizing least-squares regression to loss functions that are more appropriate for other members of this family. The regression set-up assumes a training sample of pairs, where is a vector of attributes, and is some response variable. The parameters of the model are a vector . The dot product is taken to be an approximation of . In least squares regression the optimal parameters are set to be ! " $#&%(' *) +& . In GLM’s, , - is taken to approximate the expectation parameter of the exponential model, where , is the inverse of the “link function” [8]. A natural choice is to use the “canonical link”, where , , being the derivative /. . In this case the natural parameters are directly approximated by / , and the log-likelihood ' is simply () * , 01
/ 2
. In the case of a normal distribution with fixed variance, / and it follows easily that the maximum-likelihood criterion is equivalent to the least squares criterion. Another interesting case is logistic regression where / , , and the negative log-likelihood for parameters is 3 , )546 ! 7 8 !:9 where if , if . 2.2 Bregman Distances and the Exponential Family Let ;2<>=@? be a differentiable and strictly convex function defined on a closed, convex set =BA . The Bregman distance associated with ; is defined for
C D= to be EGF HJI C ; H+ ; C LK C M C where K ' ; ' . It can be shown that, in general, every Bregman distance is nonnegative and is equal to zero if and only if its two arguments are equal. For the exponential family the log-likelihood () ' is directly related to a Bregman normal Bernoulli Poisson
"!# !# $ % &('*) % + ! # , .-/ 0 .-/ 21 -3 0 41 -/ 0 .-/ 1 56.-/87 $39 ' .-3 ;: ' 9 : <>= ?A@"B0 C? 1 B0 ? 0ED FG H1 ?/ 0 ' 9 D ' 9 F ? 0GD FE B 1 ? < = I-A@ $ 0J .1 0
"! 9 : 6# where -LKH7 1M ! # 1 -N 0 1 Table 1: Various functions of interest for three members of the exponential family distance. Specifically, [1, 3] define a “dual” function ; through / and : ; , / (3) It can be shown under fairly general conditions that K ' ' . Application of these identities implies that the negative log-likelihood of a point can be expressed through a Bregman distance [1, 3]: () ' () +* ' ; ' , E F ' I (4) In other words, negative log-likelihood can always be written as a Bregman distance plus a term that is constant with respect to and which therefore can be ignored. Table 1 summarizes various functions of interest for examples of the exponential family. We will find it useful to extend the idea of Bregman distances to divergences between vectors and matrices. If , O are vectors, and P , Q are matrices, then we overload the notation as E F I O E F ' I and E F P I Q % E F SR % IUT % . (The notion of Bregman distance as well as our generalization of PCA can be extended to vectors in a more general manner; here, for simplicity, we restrict our attention to Bregman distances and PCA problems of this particular form.) 3 PCA for the Exponential Family We now generalize PCA to other members of the exponential family. We wish to find ’s that are “close” to the ’s and which belong to a lower dimensional subspace of parameter space. Thus, our approach is to find a basis V % VXW in and to represent each as the linear combination of these elements ZY R Y V Y that is “closest” to
. Let [ be the ]\_^ matrix whose ’th row is . Let ` be the a \b^ matrix whose c ’th row is V Y , and let P be the d\ a matrix with elements R Y . Then e PG` is an d\f^ matrix whose ’th row is as above. This is a matrix of natural parameter values which define the probability of each point in [ . Following the discussion in Section 2, we consider the loss function taking the form g ` P [ P ` % ' &% &% Zh , % ' &% % , / % where h is a constant term which will be dropped from here on. The loss function varies depending on which member of the exponential family is taken, which simply changes the form of / . For example, if [ is a matrix of real values, and the normal distribution is appropriate for the data, then / and the loss criterion is the usual squared loss for PCA. For the Bernoulli distribution, / , ) . If we define ' &% ' &% , then g ` P %- , 6 !i !i . From the relationship between log-likelihood and Bregman distances (see Eq. (4)), the loss can also be written as g ` P % E F ' &% I % E F I (where we allow to be applied to vectors and matrices in a pointwise manner). Once ` and P have been found for the data points, the ’th data point can be represented as the vector in the lower dimensional space W . Then are the coefficients which define a Bregman projection of the vector : " #&% ' E F I 3` (5) The generalized form of PCA can also be considered to be search for a low dimensional basis (matrix ` ) which defines a surface that is close to all the data points . We define the set of points ` to be ` 3` W . The optimal value for ` then minimizes the sum of projection distances: ` " # %(' #&%(' E F I
. Note that for the normal distribution and the Bregman distance is Euclidean distance so that the projection operation in Eq. (5) is a simple linear projection ( "
` ). ` is also simplified in the normal case, simply being the hyperplane whose basis is ` . To summarize, once a member of the exponential family — and by implication a convex function / — is chosen, regular PCA is generalized in the following way: The loss function is negative log-likelihood, () ' ' , / -, constant. The matrix e PG` is taken to be a matrix of natural parameter values. The derivative of / defines a matrix of expectation parameters, PG` . A function ; is derived from / and . A Bregman distance E F is derived from ; . The loss is a sum of Bregman distances from the elements ' % to values % % . PCA can also be thought of as search for a matrix ` that defines a surface ` which is “close” to all the data points. The normal distribution is a simple case because , and the divergence is Euclidean distance. The projection operation is a linear operation, and ` is the hyperplane which has ` as its basis. 4 Generic Algorithms for Minimizing the Loss Function We now describe a generic algorithm for minimization of the loss function. First, we concentrate on the simplest case where there is just a single component so that a . (We drop the c subscript from R Y and % Y .) The method is iterative, with an initial random choice for the value of ` . Let ` , P , etc. denote the values at the ’th iteration, and let ` * be the initial random choice. We propose the iterative updates P " # %(' g ` P and ` " # %(' g ` P . Thus g is alternately minimized with respective to its two arguments, each time optimizing one argument while keeping the other one fixed, reminiscent of Csisz´ar and Tusn´ady’s alternating minization procedures [2]. It is useful to write these minimization problems as follows: For , R ! " $#&%(' % E F 3 ' &% I *R % 9 For " ^ , % ! " $#&%(' E F 3 ' % I SR 9 . We can then see that there are , ^ optimization problems, and that each one is essentially identical to a GLM regression problem (a very simple one, where there is a single parameter being optimized over). These sub-problems are easily solved, as the functions are convex in the argument being optimized over, and the large literature on maximumlikelihood estimation in GLM’s can be directly applied to the problem. These updates take a simple form for the normal distribution: P [ ` ` , and ` P [ P . It follows that ` ` [ [ h , where h is a scalar value. The method is then equivalent to the power method (see Jolliffe [5]) for finding the eigenvector of [ [ with the largest eigenvalue, which is the best single component solution for ` . Thus the generic algorithm generalizes one of the oldest algorithms for solving the regular PCA problem. The loss is convex in either of its arguments with the other fixed, but in general is not convex in the two arguments together. This makes it very difficult to prove convergence to the global minimum. The normal distribution is an interesting special case in this respect — the power method is known to converge to the optimal solution, in spite of the non-convex nature of the loss surface. A simple proof of this comes from properties of eigenvectors (Jolliffe [5]). It can also be explained by analysis of the Hessian : for any stationary point which is not the global minimum, is not positive semi-definite. Thus these stationary points are saddle points rather than local minima. The Hessian for the generalized loss function is more complex; it remains an open problem whether it is also not positive semidefinite at stationary points other than the global minimum. It is also open to determine under which conditions this generic algorithm will converge to a global minimum. In preliminary numerical studies, the algorithm seems to be well behaved in this respect. Moreover, any limit point of the sequence e P ` will be a stationary point. However, it is possible for this sequence to diverge since the optimum may be at infinity. To avoid such degenerate choices of e , we can use a modified loss % E F ' &% I % , E F * I % where is a small positive constant, and +* is any value in the range of (and therefore for which * is finite). This is roughly equivalent to adding a conjugate prior and finding the maximum a posteriori solution. It can be proved, for this modified loss, that the sequence e remains in a bounded region and hence always has at least one limit point which must be a stationary point. (All proofs omitted for lack of space.) There are various ways to optimize the loss function when there is more than one component. We give one algorithm which cycles through the a components, optimizing each in turn while the others are held fixed: //Initialization Set 7 , 7 //Cycle through components times For 7 0 , 7 0
: //Now optimize the ’th component with other components fixed Initialize & + randomly, and set 7 For 7 0 convergence For ! 7
, &#" + 7$% '&)(#*,+ < = 3 @ $ &#" 9 ' + . 9 For / 7 810 , &#" + 7$% '&)(#*32 < = 3 @ $ &" + . 9 The modified Bregman projections now include a term 4 &% representing the contribution of the a fixed components. These sub-problems are again a standard optimization problem regarding Bregman distances, where the terms 4 % form a “reference prior”. 0 10 20 30 40 50 60 70 80 90 0 10 20 30 40 50 60 70 80 90 X Y data PCA exp 0 20 40 60 80 100 0 50 100 150 200 250 300 350 400 450 500 X Y data PCA exp Figure 1: Regular PCA vs. PCA for the exponential distribution. 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1 A C’ B B’ C 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1 C C’ D B B’ A D’ Figure 2: Projecting from 3- to 1-dimensional space, via Bernoulli PCA. Left: the three points E h are projected onto a one-dimensional curve. Right: point is added. 5 Illustrative examples Exponential distribution. Our generalization of PCA behaves rather differently for different members of the exponential family. One interesting example is that of the exponential distributions on nonnegative reals. For one-dimensional data, these densities are usually written as , where is the mean. In the uniform system of notation we have been using, we would instead index each distribution by a single natural parameter (basically, ), and write the density as ' , where / ' . The link function in this case is , the mean of the distribution. Suppose we are given data [ and want to find the best one-dimensional approximation: a vector V and coefficients such that the approximation *R V & has minimum loss. The alternating minimization procedure of the previous section has a simple closed form in this case, consisting of the iterative update rule V
^ [ [bV Here the shorthand denotes a componentwise reciprocal, i.e., ) . Notice the similarity to the update rule of the power method for PCA: V
[ [bV . Once V is found, we can recover the coefficients ^ The points R V lie on a line through the origin. Normally, we would not expect the points to also lie on a straight line; however, in this case they do, because any point of the form SR V R , can be written as and so must lie in the direction . Therefore, we can reasonably ask how the lines found under this exponential assumption differ from those found under a Gaussian assumption (that is, those found by regular PCA), provided all data is nonnegative. As a very simple illustration, we conducted two toy experiments with twenty data points in (Figure 1). In the first, the points all lay very close to a line, and the two versions of PCA produced similar results. In the second experiment, a few of the points were moved farther afield, and these outliers had a larger effect upon regular PCA than upon its exponential variant. Bernoulli distribution. For the Bernoulli distribution, a linear subspace of the space of parameters is typically a nonlinear surface in the space of the data. In Figure 2 (left), three points in the three-dimensional hypercube are mapped via our PCA to a onedimensional curve. The curve passes through one of the points ( ); the projections of the two other ( E ? E and h ? h ) are indicated. Notice that the curve is symmetric about the center of the hypercube, . In Figure 2 (right), another point (D) is added, and causes the approximating one-dimensional curve to swerve closer to it. 6 Relationship to Previous Work Lee and Seung [6, 7] and Hofmann [4] also describe probabilistic alternatives to PCA, tailored to data types that are not gaussian. In contrast to our method, [4, 6, 7] approximate mean parameters underlying the generation of the data points, with constraints on the matrices P and ` ensuring that the elements of PG` are in the correct domain. By instead choosing to approximate the natural parameters, in our method the matrices P and ` do not usually need to be constrained—instead, we rely on the link function to give a transformed matrix PG` which lies in the domain of the data points. More specifically, Lee and Seung [6] use the loss function % ' &% () % , &% (ignoring constant factors, and again defining &%$ Y R Y Y % ). This is optimized with the constraint that P and ` should be positive. This method has a probabilistic interpretation, where each data point ' &% is generated from a Poisson distribution with mean parameter % . For the Poisson distribution, our method uses the loss function % ' % &% , !i , but without any constraints on the matrices P and ` . The algorithm in Hofmann [4] uses a loss function % ' &% () % , where the matrices P and ` are constrained such that all the % ’s are positive, and also such that % % . Bishop and Tipping [9] describe probabilistic variants of the gaussian case. Tipping [10] discusses a model that is very similar to our case for the Bernoulli family. Acknowledgements. This work builds upon intuitions about exponential families and Bregman distances obtained largely from interactions with Manfred Warmuth, and from his papers. Thanks also to Andreas Buja for several helpful comments. References [1] Katy S. Azoury and M. K. Warmuth. Relative loss bounds for on-line density estimation with the exponential family of distributions. Machine Learning, 43:211–246, 2001. [2] I. Csisz´ar and G. Tusn´ady. Information geometry and alternating minimization procedures. Statistics and Decisions, Supplement Issue, 1:205–237, 1984. [3] J¨urgen Forster and Manfred Warmuth. Relative expected instantaneous loss bounds. Journal of Computer and System Sciences, to appear. [4] Thomas Hofmann. Probabilistic latent semantic indexing. In Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, 1999. [5] I. T. Jolliffe. Principal Component Analysis. Springer-Verlag, 1986. [6] D. D. Lee and H. S. Seung. Learning the parts of objects with nonnegative matrix factorization. Nature, 401:788, 1999. [7] Daniel D. Lee and H. Sebastian Seung. Algorithms for non-negative matrix factorization. In Advances in Neural Information Processing Systems 13, 2001. [8] P. McCullagh and J. A. Nelder. Generalized Linear Models. CRC Press, 2nd edition, 1990. [9] M. E. Tipping and C. M. Bishop. Probabilistic principal component analysis. Journal of the Royal Statistical Society, Series B, 61(3):611–622, 1999. [10] Michael E. Tipping. Probabilistic visualisation of high-dimensional binary data. In Advances in Neural Information Processing Systems 11, pages 592–598, 1999.
|
2001
|
188
|
2,003
|
Minimax Probability Machine Gert R.G. Lanckriet* Department of EECS University of California, Berkeley Berkeley, CA 94720-1770 gert@eecs.berkeley.edu Chiranjib Bhattacharyya Department of EECS University of California, Berkeley Berkeley, CA 94720-1776 chiru@eecs.berkeley.edu Laurent EI Ghaoui Department of EECS University of California, Berkeley Berkeley, CA 94720-1770 elghaoui@eecs.berkeley.edu Michael I. Jordan Computer Science and Statistics University of California, Berkeley Berkeley, CA 94720-1776 jordan@cs.berkeley.edu Abstract When constructing a classifier, the probability of correct classification of future data points should be maximized. In the current paper this desideratum is translated in a very direct way into an optimization problem, which is solved using methods from convex optimization. We also show how to exploit Mercer kernels in this setting to obtain nonlinear decision boundaries. A worst-case bound on the probability of misclassification of future data is obtained explicitly. 1 Introduction Consider the problem of choosing a linear discriminant by minimizing the probabilities that data vectors fall on the wrong side of the boundary. One way to attempt to achieve this is via a generative approach in which one makes distributional assumptions about the class-conditional densities and thereby estimates and controls the relevant probabilities. The need to make distributional assumptions, however, casts doubt on the generality and validity of such an approach, and in discriminative solutions to classification problems it is common to attempt to dispense with class-conditional densities entirely. Rather than avoiding any reference to class-conditional densities, it might be useful to attempt to control misclassification probabilities in a worst-case setting; that is, under all possible choices of class-conditional densities. Such a minimax approach could be viewed as providing an alternative justification for discriminative approaches. In this paper we show how such a minimax programme can be carried out in the setting of binary classification. Our approach involves exploiting the following powerful theorem due to Isii [6], as extended in recent work by Bertsimas • http://robotics.eecs.berkeley.edur gert/ and Sethuraman [2]: where y is a random vector, where a and b are constants, and where the supremum is taken over all distributions having mean y and covariance matrix ~y. This theorem provides us with the ability to bound the probability of misclassifying a point, without making Gaussian or other specific distributional assumptions. We will show how to exploit this ability in the design of linear classifiers. One of the appealing features of this formulation is that one obtains an explicit upper bound on the probability of misclassification of future data: 1/(1 + rP). A second appealing feature of this approach is that, as in linear discriminant analysis [7], it is possible to generalize the basic methodology, utilizing Mercer kernels and thereby forming nonlinear decision boundaries. We show how to do this in Section 3. The paper is organized as follows: in Section 2 we present the minimax formulation for linear classifiers, while in Section 3 we deal with kernelizing the method. We present empirical results in Section 4. 2 Maximum probabilistic decision hyperplane In this section we present our minimax formulation for linear decision boundaries. Let x and y denote random vectors in a binary classification problem, with mean vectors and covariance matrices given by x '" (x, ~x) and y '" (y, ~y) , respectively, where ""," means that the random variable has the specified mean and covariance matrix but that the distribution is otherwise unconstrained. Note that x, x, y, Y E JRn and ~x, ~y E JRnxn. We want to determine the hyperplane aT z = b (a, z E JRn and b E JR) that separates the two classes of points with maximal probability with respect to all distributions having these means and covariance matrices. This boils down to: or, max a s.t. inf Pr{ aT x 2: b} 2: a (2) a ,a ,b max a s.t. a,a,b 1 - a 2: sup Pr{ aT x :s b} 1- a 2: sup Pr{aT y 2: b}. (3) Consider the second constraint in (3). Recall the result of Bertsimas and Sethuraman [2]: 1 supPr{aTY2:b}=-d2' with d2 = inf (Y_Yf~y-1(y_y) (4) 1 + aTy?b We can write this as d2 = infcTw>d wT w, where w = ~y -1/2 (y_y), cT = aT~y 1/2 and d = b - aTy. To solve this,-first notice that we can assume that aTy :S b (i.e. y is classified correctly by the decision hyperplane aT z = b): indeed, otherwise we would find d2 = 0 and thus a = 0 for that particular a and b, which can never be an optimal value. So, d> o. We then form the Lagrangian: £(w, >.) = wT w + >.(d - cT w), (5) which is to be maximized with respect to A 2: 0 and minimized with respect to w . At the optimum, 2w = AC and d = cT W , so A = -!#c and w = c%c. This yields: (6) Using (4), the second constraint in (3) becomes 1-0: 2: 1/(I+d2 ) or ~ 2: 0:/(1-0:). Taking (6) into account, this boils down to: b-aTY2:,,(o:)/aT~ya where ,,(0:)=) 0: (7) V 1-0: We can handle the first constraint in (3) in a similar way (just write aT x ::::: b as _aT x 2: -b and apply the result (7) for the second constraint). The optimization problem (3) then becomes: max 0: s.t. a ,a,b -b + aTx 2: ,,(o:)JaT~xa b - aTy 2: "(o:h/aT~ya. Because "(0:) is a monotone increasing function of 0:, we can write this as: max" s.t. ""a,b b - aTy 2: "JaT~ya. From both constraints in (9), we get aTy + "JaT~ya::::: b::::: aTx - "JaT~xa, which allows us to eliminate b from (9): aTy + "JaT~ya::::: aTx - "JaT~xa. I<,a max" s.t. (8) (9) (10) (11) Because we want to maximize ", it is obvious that the inequalities in (10) will become equalities at the optimum. The optimal value of b will thus be given by where a* and "* are the optimal values of a and " respectively. constraint in (11), we get: aT(x - y) 2:" (JaT~xa+ JaT~ya). (12) Rearranging the (13) The above is positively homogeneous in a: if a satisfies (13), sa with s E 114 also does. Furthermore, (13) implies aT(x - y) 2: O. Thus, we can restrict a to be such that aT(x - y) = 1. The optimization problem (11) then becomes max" s.t. I<,a which allows us to eliminate ,,: ~ 2: JaT~xa + JaT~ya aT (x-Y)=I, m~n JaT~xa + JaT~ya s.t. aT(x - y) = 1, (14) (15) or, equivalently (16) This is a convex optimization problem, more precisely a second order cone program (SOCP) [8,5]. Furthermore, notice that we can write a = ao +Fu, where U E Il~n-l, ao = (x - y)/llx - y112, and F E IRnx (n-l) is an orthogonal matrix whose columns span the subspace of vectors orthogonal to x - y. Using this we can write (16) as an unconstrained SOCP: (17) We can solve this problem in various ways, for example using interior-point methods for SOCP [8], which yield a worst-case complexity of O(n3 ). Of course, the first and second moments of x, y must be estimated from data, using for example plug-in estimates X, y, :Ex, :Ey for respectively x, y, ~x, ~y. This brings the total complexity to O(ln3 ), where l is the number of data points. This is the same complexity as the quadratic programs one has to solve in support vector machines. In our implementations, we took an iterative least-squares approach, which is based on the following form, equivalent to (17): (18) At iteration k, we first minimize with respect to 15 and E by setting 15k = II~x 1/2(ao + Fuk- d112 and Ek = II~y 1/2(ao + Fuk - 1)112. Then we minimize with respect to U by solving a least squares problem in u for 15 = 15k and E = Ek, which gives us Uk. Because in both update steps the objective of this COP will not increase, the iteration will converge to the global minimum II~xl/2(ao + Fu*)112 + II~yl /2(ao + Fu*)lb with u* an optimal value of u. We then obtain a* as ao + Fu* and b* from (12) with "'* = l/h/ar~xa* + Jar~ya*). Classification of a new data point Znew is done by evaluating sign( a;; Znew - b*): if this is + 1, Znew is classified as from class x, otherwise Znew is classified as from class y. It is interesting to see what happens if we make distributional assumptions; in particular, let us assume that x "" N(x, ~x) and y "" N(y, ~y). This leads to the following optimization problem: max a S.t. -b + aTx ::::: <I>-l(a)JaT~xa o:,a ,b (19) where <I>(z) is the cumulative distribution function for a standard normal Gaussian distribution. This has the same form as (8), but now with ",(a) = <I>-l(a) instead of ",(a) = V l~a (d. a result by Chernoff [4]). We thus solve the same optimization problem (a disappears from the optimization problem because ",(a) is monotone increasing) and find the same decision hyperplane aT z = b. The difference lies in the value of a associated with "'*: a will be higher in this case, so the hyperplane will have a higher predicted probability of classifying future data correctly. 3 Kernelization In this section we describe the "kernelization" of the minimax approach described in the previous section. We seek to map the problem to a higher dimensional feature space ]Rf via a mapping cP : ]Rn 1-+ ]Rf, such that a linear discriminant in the feature space corresponds to a nonlinear discriminant in the original space. To carry out this programme, we need to try to reformulate the minimax problem in terms of a kernel function K(Z1' Z2) = cp(Z1)T CP(Z2) satisfying Mercer's condition. Let the data be mapped as x 1-+ cp(x) ""' (cp(X) , ~cp(x)) and Y 1-+ cp(y) ""' (cp(y) , ~cp(y)) where {Xi}~1 and {Yi}~1 are training data points in the classes corresponding to x and Y respectively. The decision hyperplane in ]Rf is then given by aT cp(Z) = b with a, cp(z) E ]Rf and b E ]R. In ]Rf, we need to solve the following optimization problem: mln Jr-aT-~-cp-(-x)-a + J aT~cp(y)a s.t. aT (cp(X) - cp(y)) = 1, (20) where, as in (12), the optimal value of b will be given by b* = a; cp(x) "'*Jar~cp(x)a* = a; cp(y) + "'*Jar~cp(y)a*, (21) where a* and "'* are the optimal values of a and '" respectively. However, we do not wish to solve the COP in this form, because we want to avoid using f or cp explicitly. If a has a component in ]Rf which is orthogonal to the subspace spanned by CP(Xi), i = 1,2, ... , N x and CP(Yi), i = 1,2, ... , Ny, then that component won't affect the objective or the constraint in (20) . This implies that we can write a as N. Ny a = LaiCP(Xi) + L;)jCP(Yj). (22) i=1 j=1 Substituting expression (22) for a and estimates ;Pw = J. 2:~1 CP(Xi) , ;p(Y) = 1 Ny A _ 1 N. .....--.. .....--.. T A _ Ny 2:i=l cp(Yi), ~cp(x) N. 2:i=1 (cp(Xi) - cp(X)) (cp(Xi) - cp(x)) and ~cp(y) N .....--.. .....--.. J 2:i~1(CP(Yi) - cp(y))(cp(Yi) - cp(y))T for the means and the covariance matriy ces in the objective and the constraint of the optimization problem (20), we see that both the objective and the constraints can be written in terms of the kernel function K(Zl' Z2) = CP(Z1)T cp(Z2) . We obtain: T "f (kx - ky) = 1, (23) T N N . where "f = [a1 a2 ... aN. ;)1 ;)2 ... ;)Nyl , kx E ]R .+ y WIth [kxli = J. 2:f;1 K(xj, Zi), ky E ]RN. +Ny with [kyli = Jy 2:f~l K(Yj, Zi), Zi = Xi for i = 1,2, ... ,Nx and Zi = Yi- N. for i = Nx + 1, Nx + 2, ... ,Nx + Ny . K is defined as: K = (Kx -IN.~~) = (*x) Ky -lNy ky Ky (24) where 1m is a column vector with ones of dimension m. Kx and Ky contain respectively the first N x rows and the last Ny rows of the Gram matrix K (defined as Kij = cp(zdTcp(zj) = K(Zi,Zj)). We can also write (23) as Kx I I Ky I T m~n II ~"f12 + I .jlV;"f 12 s.t. "f (kx - ky) = 1, (25) which is a second order cone program (SOCP) [5] that has the same form as the SOCP in (16) and can thus be solved in a similar way. Notice that, in this case, the optimizing variable is "f E ~Nz +Ny instead of a E ~n. Thus the dimension of the optimization problem increases, but the solution is more powerful because the kernelization corresponds to a more complex decision boundary in ~n . Similarly, the optimal value b* of b in (21) will then become (26) where "f* and "'* are the optimal values of "f and", respectively. Once "f* is known, we get "'* = 1/ ( J ~z "f;K~Kx"f* + J ~y "f;K~Ky "f* ) and then b* from (26). Classification of a new data point Znew is then done by evaluating sign(a; <p(znew) -b*) = sign ( (L~l+Ny b*]iK(Zi, Znew) ) - b*) (again only in terms of the kernel function): if this is + 1, Znew is classified as from class x, otherwise Znew is classified as from class y. 4 Experiments In this section we report the results of experiments that we carried out to test our algorithmic approach. The validity of 1 - a as the worst case bound on the probability of misclassification of future data is checked, and we also assess the usefulness of the kernel trick in this setting. We compare linear kernels and Gaussian kernels. Experimental results on standard benchmark problems are summarized in Table 1. The Wisconsin breast cancer dataset contained 16 missing examples which were not used. The breast cancer, pima, diabetes, ionosphere and sonar data were obtained from the VCI repository. Data for the twonorm problem data were generated as specified in [3]. Each dataset was randomly partitioned into 90% training and 10% test sets. The kernel parameter (u) for the Gaussian kernel (e-llx-yI12/,,) was tuned using cross-validation over 20 random partitions. The reported results are the averages over 50 random partitions for both the linear kernel and the Gaussian kernel with u chosen as above. The results are comparable with those in the existing literature [3] and with those obtained with Support Vector Machines. Also, we notice that a is indeed smaller Table 1: a and test-set accuracy (TSA) compared to BPB (best performance in [3]) and to the performance of an SVM with linear kernel (SVML) and an SVM with Gaussian kernel (SVMG) Dataset Linear kernel Gaussian kernel BPB SVML SVMG a TSA: a TSA: Twonorm 80.2 % 96.0 % 83.6 % 97.2 % 96.3 % 95.6 % 97.4 % Breast cancer 84.4 % 97.2 % 92.7 % 97.3 % 96.8 % 92.6 % 98.5 % Ionosphere 63.3 % 85.4 % 89.9 % 93.0 % 93.7 % 87.8 % 91.5 % Pima diabetes 31.2 % 73.8 % 33.0 % 74.6 % 76.1 % 70.1 % 75.3 % Sonar 62.4 % 75.1 % 87.1 % 89.8 % 75.9 % 86.7 % than the test-set accuracy in all cases. Furthermore, a is smaller for a linear decision boundary then for the nonlinear decision boundary obtained via the Gaussian kernel. This clearly shows that kernelizing the method leads to more powerful decision boundaries. 5 Conclusions The problem of linear discrimination has a long and distinguished history. Many results on misclassification rates have been obtained by making distributional assumptions (e.g., Anderson and Bahadur [1]). Our results, on the other hand, make use of recent work on moment problems and semidefinite optimization to obtain distribution-free results for linear discriminants. We have also shown how to exploit Mercer kernels to generalize our algorithm to nonlinear classification. The computational complexity of our method is comparable to the quadratic program that one has to solve for the support vector machine (SVM). While we have used a simple iterative least-squares approach, we believe that there is much to gain from exploiting analogies to the SVM and developing specialized, more efficient optimization procedures for our algorithm, in particular tools that break the data into subsets. The extension towards large scale applications is a current focus of our research, as is the problem of developing a variant of our algorithm for multiway classification and function regression. Also the statistical consequences of using plug-in estimates for the mean vectors and covariance matrices needs to be investigated. Acknowledgements We would like to acknowledge support from ONR MURI N00014-00-1-0637, from NSF grants IIS-9988642 and ECS-9983874 and from the Belgian American Educational Foundation. References [1] Anderson, T . W . and Bahadur, R. R. (1962) Classification into two multivariate Normal distributions with different covariance matrices. Annals of Mathematical Statistics 33(2): 420-431. [2] Bertsimas, D. and Sethuraman, J. (2000) Moment problems and semidefinite optimization. Handbook of Semidefinite Optimization 469-509, Kluwer Academic Publishers. [3] Breiman L. (1996) Arcing classifiers. Technical Report 460, Statistics Department, University of California, December 1997. [4] Chernoff H. (1972) The selection of effective attributes for deciding between hypothesis using linear discriminant functions. In Frontiers of Pattern Recognition, (S. Watanabe, ed.), 55-60. New York: Academic Press. [5] Boyd, S. and Vandenberghe, L. (2001) Convex Optimization. Course notes for EE364, Stanford University. Available at http://www . stanford. edu/ class/ee364. [6] Isii, K. (1963) On the sharpness of Chebyshev-type inequalities. Ann. Inst. Stat. Math. 14: 185-197. [7] Mika, M. Ratsch, G., Weston, J., SchOikopf, B., and Mii11er, K.-R. (1999) Fisher discriminant analysis with kernels. In Neural Networks for Signal Processing IX, 41- 48, New York: IEEE Press. [8] Nesterov, Y. and Nemirovsky, A. (1994) Interior Point Polynomial Methods in Convex Programming: Theory and Applications. Philadelphia, PA: SIAM.
|
2001
|
189
|
2,004
|
Agglomerative Multivariate Information Bottleneck Noam Sionim Nir Friedman Naftali Tishby School of Computer Science & Engineering, Hebrew University, Jerusalem 91904, Israel {noamm, nir, tishby } @cs.huji.ac.il Abstract The information bottleneck method is an unsupervised model independent data organization technique. Given a joint distribution peA, B), this method constructs a new variable T that extracts partitions, or clusters, over the values of A that are informative about B. In a recent paper, we introduced a general principled framework for multivariate extensions of the information bottleneck method that allows us to consider multiple systems of data partitions that are inter-related. In this paper, we present a new family of simple agglomerative algorithms to construct such systems of inter-related clusters. We analyze the behavior of these algorithms and apply them to several real-life datasets. 1 Introduction The information bottleneck (IB) method of Tishby et al [14] is an unsupervised nonparametric data organization technique. Given a joint distribution P(A, B), this method constructs a new variable T that represents partitions of A which are (locally) maximizing the mutual information about B. In other words, the variable T induces a sufficient partition, or informative features of the variable A with respect to B. The construction of T finds a tradeoff between the information about A that we try to minimize, J(T; A), and the information about B which we try to maximize, J(T; B). This approach is particularly useful for co-occurrence data, such as words and documents [12], where we want to capture what information one variable (e.g., use of a word) contains about the other (e.g., the document). In a recent paper, Friedman et al. [4] introduce multivariate extension of the IB principle. This extension allows us to consider cases where the data partition is relevant with respect to several variables, or where we construct several systems of clusters simultaneously. In this framework, we specify the desired interactions by a pair of Bayesian networks. One network, Gin, represents which variables are compressed versions of the observed variables - each new variable compresses its parents in the network. The second network, Gout> defines the statistical relationship between these new variables and the observed variables that should be maintained. Similar to the original IB, in Friedman et al. we formulated the general principle as a tradeoff between the (multi) information each network carries. On the one hand, we want to minimize the information maintained by G in and on the other to maximize the information maintained by Gout. We also provide a characterization of stationary points in this tradeoff as a set of self-consistent equations. Moreover, we prove that iterations of these equations converges to a (local) optimum. Then, we describe a deterministic annealing procedure that constructs a solution by tracking the bifurcation of clusters as it traverses the tradeoff curve, similar to the original IB method. In this paper, we consider an alternative approach to solving multivariate IB problems which is motivated by the success of the agglomerative IB of Slonim and Tishby [11]. As shown there, a bottom-up greedy agglomeration is a simple heuristic procedure that can yield good solutions to the original IB problem. Here we extend this idea in the context of multivariate IB problems. We start by analyzing the cost of agglomeration steps within this framework. This both elucidates the criteria that guides greedy agglomeration and provides for efficient local evaluation rules for agglomeration steps. This construction results with a novel family of information theoretic agglomerative clustering algorithms, that can be specified using the graphs Gin and G out. We demonstrate the performance of some of these algorithms for document and word clustering and gene expression analysis. 2 Multivariate Information Bottleneck A Bayesian network structure G is a DAG that specifies interactions among variables [8]. A distribution P is consistent with G (denoted P F G), if P(Xl , ... , X n) = I1 P(Xi I Pa<fJ, where Pa<fi are the parents of X i in G. Our main interest is in the information that the variables Xl " '" X n contain about each other. A quantity that captures this is the multi-information given by where V(Pllq) is the familiar Kullback-Liebler divergence [2]. Proposition 2.1 [4] Let G be a DAG over {Xl , ... , X n}, and let P F G be a distribution. Then, I G = I(Xl' ... , X n) = L i I(Xi ; Pa<fi ). That is, the multi-information is the sum of local mutual information terms between each variable and its parents (denoted I G). Friedman et al. define the multivariate IE problem as follows. Suppose we are given a set of observed variables, X = {Xl , ... , X n} and their joint distribution P (X l , ... , X n). We want to "construct" new variables T , where the relations between the observed variables and these new compression variables are specified using a DAG Gin over X U T where the variables in T are leafs. Thus, each Tj is a stochastic function of a set of variables U j = Pa~;in ~ X. Once these are set, we have a joint distribution over the combined set of variables: P(X, T) = P(X) ITj P(Tj I U j ). The "relevant" information that we want to preserve is specified by another DAG, Gout . This graph specifies, for each Tj which variables it predicts. These are simply its children in G out . More precisely, we want to predict each Xi (or Tj ) by V X i = Pa~;"t (resp. V T; = Pa~;out ), its parents in G out. Thus, we think ofIGout as a measure of how much information the variables in T maintain about their target variables. The Lagrangian can then be defined as (1) with a tradeoff parameter (Lagrange multiplier) (3. 1 The variation is done subject to the normalization constraints on the partition distributions. Thus, we balance between the information T loses about X in G in and the information it preserves in G out. Friedman et al. [4] show that stationary points of this Lagrangian satisfy a set of selfconsistent equations. Moreover, they show that iterating these equations converges to a INotice that under this formulation we would like to maximize £. An equivalent definition [4] would be to minimize £ = 'LGin - (J . 'LGO"t . stationary point of the tradeoff. Then, extending the procedure of Tishby et al [14], they propose a procedure that searches for a solution of the IB equations using a 'deterministic annealing' approach [9]. This is a top-down hierarchical algorithm that starts from a single cluster for each Tj at j3 -+ 0, and then undergoes a cascade of cluster splits as j3 is being "cooled". These determines "soft" trees of clusters (one for each Tj ) that describe solutions at different tradeoff values of j3. 3 The Agglomerative Procedure For the original IB problem, Slonim and Tishby [11] introduced a simpler procedure that performs greedy bottom-up merging of values. Several successful applications of this algorithm are already presented for a variety of real-world problems [10, 12, 13, 15]. The main focus of the current work is in extending this approach for the multivariate IB problem. As we will show, this will lead to further insights about the method, and also provide a rather simple and intuitive clustering procedures. We consider procedures that start with a set of clusters for each Tj (usually the most fine-grained solution we can consider where Tj = U j ) and then iteratively reduce the cardinality of one of the Tj 's by merging two values t~ and tj of Tj into a single value lj. To formalize this notion we must define the membership probability of a new cluster lj, resulting from merging {t~, tj} '* lj in Tj . This is done rather naturally by (2) In other words, we view the event l j as the union of the events t~ and tj. Given the membership probabilities, at each step we can also draw the connection between Tj and the other variables. This is done using the following proposition which is based on the conditional independence assumptions given in Gin. Proposition 3.1 Let Y , Z C X U T \ {Tj} then, (3) h II { } { p( t ~ I Z) p(t j IZ) }. h d· ·b· d·· d were Z = 1f1,Z, 1fr ,z = p(t; IZ)' p(t; IZ) IS t e merger 1Str1 uttOn can ltzone on z. In particular, this proposition allows us to evaluate all the predictions defined in G out and all the informations terms in £ that involve Tj . The crucial question in an agglomerative process is of course which pair to merge at each step. We know that the merger "cost" in our terms is exactly the difference in the values of £ , before and after the merger. Let T}ef and T/ft denote the random variables that correspond to Tj , before and after the merger, respectively. Thus, the values of £ before and after the merger are calculated based on Trf and Ttt. The merger cost is then simply given by, (4) The greedy procedure evaluates all the potential mergers (for all Tj ) and then applies the best one (i.e., the one that minimizes 6.£( t~ , tj). This is repeated until all the variables in T degenerate into trivial clusters. The resulting set of trees describes a range of solutions at different resolutions. This agglomerative approach is different in several important aspects from the deterministic annealing approach described above. In that approach, by "cooling" (i.e., increasing) j3, we move along a tradeoff curve from the trivial - single cluster - solution toward solutions with higher resolutions that preserve more information in G out. In contrast, in the agglomerative approach we progress in the opposite direction. We start with a high resolution clustering and as the merging process continues we move toward more and more compact solutions. During this process (3 is kept constant and the driving force is the reduction in the cardinality of the T/s. Therefore, we are able to look for good solutions in different resolutions for ajixed tradeoff parameter (3. Since the merging does not attempt directly to maintain the (stationary) self-consistent "soft" membership probabilities, we do not expect the self-consistent equations to hold at solutions found by the agglomerative procedure. On the other hand, the agglomerative process is much simpler to implement and fully deterministic. As we will show, it provides sufficiently good solutions for the IB problem in many situations. 4 Local Merging Criteria In the procedure we outline above, at every step there are O(ITj 12) possible mergers of values of Tj (for every j). A direct calculation of the costs of all these potential mergers is typically infeasible. However, it turns out that one may calculate t:...c (t; , tj) while examining only the probability distributions that involve t; and tj directly. Generalizing the results of [11] for the original IB, we now develop a closed-form formula for t:...c(t;, tj). To describe this result we need the following definition. The Jensen-Shannon ( J S) divergence [7, 3] between two probabilities PI , P2 is given by where II = {7rl' 7r2} is a normalized probability and p = 7rlPl + 7r2P 2 . The J S divergence is equal zero if and only if both its arguments are identical. It is upper bounded and symmetric, though it is not a metric. One interpretation of the J S -divergence relates it to the (logarithmic) measure of the likelihood that the two sample distributions originate by the most likely common source, denoted by p. In addition, we need the notation V-X~ = V X i - {Tj} (similarly for VT!). Theorem 4.1 Let t; , tj E Tj be two clusters. Then, t:...c( t; , tj) = p(tj) . d(t;, tj) where d(t; , tj) L Ep(' lt;) [JSrrv _; (P(Xi 1 t;, V-X~ ),p(Xi 1 tj, V -X{))] i:T; EVXi Xi + L Ep(' lt; ) [JSrrv _; (p(Te 1 t;, V Tj) ,p(Te 1 tj , VT! ))] e:T;EvT£ T£ + JSrr(p(VT; 1 t;) ,p(VT; 1 tj)) - (3-1. JSrr(p(Uj 1 t;) ,p(Uj 1 tj)) A detailed proof of this theorem will be given elsewhere. Thus, the merger cost is a multiplication of the weight of the merger components (P(tj)) with their "distance" given by d(t; , tj). Notice that due to the properties of the JS-divergence, this distance is symmetric. In addition, the last term in this distance has the opposite sign to the first three terms. Thus, the distance between two clusters is a tradeoff between these two factors. Roughly speaking, we may say that the distance is minimized for pairs that give similar predictions about the variables connected with Tj in Gout and have different predictions (minimum overlap) about the variables connected with Tj in Gin. We notice also the analogy between this result and the main theorem in [4]. In [4] the optimization is governed by the KL divergences between data and cluster's centroids, or by the likelihood that the data was generated by the centroid distribution. Here the optimization is controlled through the J S divergences, i.e. the likelihood that the two clusters have a common source. Next, we notice that after applying a merger, only a small portion of the other mergers costs change. The following proposition characterizes these costs. r 0~ n T} Tz ~ o 8 n T,( TIJ n A 8 Gin G out Gin G out Gin G out I(T; B) - (3-1 I(T; A) I(T1, T2; B) I(TA;TB ) _(3-" (I(T1 ; A) + I(T2; A)) _((3-" - l)(I(TA; A) + I(TB; B)) (a) Original Bottleneck (b) Parallel Bottleneck (c) Symmetric Bottleneck Figure 1: The source and target networks and the corresponding Lagrangian for the three examples we consider. Proposition4.2 The merger {t; ,tj} :::} tj in Tj can change the cost 6..c(t~ , tc ) only if p(tj , te) > 0 and Tj , Te co-appear in some information term in r Gout • This proposition is particularly useful, when we consider "hard" clustering where Tj is a (deterministic) function ofUj . In this case, p(tj,te) is often zero (especially when Tj and Te compressing similar variables, i.e., U j n U e =I- 0). In particular, after the merger {t;, tj} :::} tj, we do not have to reevaluate merger costs of other values of Tj, except for mergers of tj with each of these values. In the case of hard clustering we also find thatI(Tj; U j ) = H(Tj) (where H(P) is Shannon's entropy). Roughly speaking, we may say that H(P) is decreasing for less balanced probability distributions p. Therefore, increasing (3-1 will result with a tendency to look for less balanced "hard" partitions and vice verse. This is reflected by the fact that the last term in d( t; , tj) is then simplified through J Sn (p(U j I t;), p(U j I tj)) = H (II). 5 Examples We now briefly consider three examples of the general methodology. For brevity we focus on the simpler case of hard clustering. We first consider the example shown in figure I(a). This choice of graphs results in the original IB problem. The merger cost in this case is given by, 6..c(tl, n = p(t) . (JSn(p(B I tl),p(B I n) - (3-1 H(II)) . (5) Note that for (3-1 -+ 0 we get exactly the algorithm presented in [11]. One simple extension of the original IB is the parallel bottleneck [4]. In this case we introduce two variables T1 and T2 as in Figure I(b), both of them are functions of A. Similarly to the original IB, Gout specifies that T1 and T2 should predict B. We can think of this requirement as an attempt to decompose the information A contains about B into two "orthogonal" components. In this case, the merger cost for T1 is given by, 6..c(ti, tD = p(t1) . (Ep(.lld[JSnT2 (P(B I ti,T2),p(B I tLT2))]- (3-1 H(II)) . (6) Finally, we consider the symmetric bottleneck [4, 12]. In this case, we want to compress A into T A and B into T B so that T A extracts the information A contains about B, and at the same time TB extracts the information B contains about A. The DAG Gin of figure I(c) captures the form of the compression. The choice of G out is less obvious and several alternatives are described in [4]. Here, we concentrate only in one option, shown in figure I(c). In this case we attempt to make each ofTA and TB sufficient to separate A from B. Thus, on one hand we attempt to compress, and on the other hand we attempt to make T A and T B as informative about each other as possible. The merger cost in T A is given by 6..c(t~, tA) = P(tA) . JSn(p(TB I t~) , p(TB ItA)) - ((3-1 - l)H(II)), (7) while for merging in TB we will get an analogous expression. 6 Applications We examine a few applications of the examples presented above. As one data set we used a subset ofthe 20 newsgroups corpus [6] where we randomly choose 2000 documents evenly distributed among the 4 science discussion groups (sci. crypt, sci. electronics, sci.med and sci.space).2 Our pre-processing included ignoring file headers (and the subject lines), lowering upper case and ignoring words that contained non 'a .. z' characters. Given this document set we can evaluate the joint probability p(W, D), which is the probability that a random word position is equal to w E Wand at the same time the document is dE D . We sort all words by their contribution to I(W; D) and used only the 2000 'most informative' ones, ending up with a joint probability with I W I = ID I = 2000. We first used the original IB to cluster W , while trying to preserve the information about D. This was already done in [12] with (3-1 = 0, but in this new experiment we took (3-1 = 0.15. Recall that increasing (3-1 results in a tendency for finding less balanced clusters. Indeed, while for (3- 1 = 0 we got relatively balanced word clusters (high H(Tw )), for (3-1 = 0.15 the probability p(Tw) is much less smooth. For 50 word clusters, one cluster contained almost half of the words, while the other clusters were typically much smaller. Since the algorithm also tries to maximize I(Tw; D), the words merged into the big cluster are usually the less informative words about D. Thus, a word must be highly informative to stay out of this cluster. In this sense, increasing (3-1 is equivalent for inducing a "noise filter", that leave only the most informative features in specific clusters. In figure 2 we present p( D I tw) for several clusters tw E Tw. Clearly, words that passed the "filter" form much more informative clusters about the real structure of D. A more formal demonstration of this effect is given in the right panel of figure 2. For a given compression level (i.e. a given I(Tw; W)), we see that taking (3-1 = 0.15 preserve much more information aboutD. While an exact implementation of the symmetric IB will require alternating mergers in Tw and TD, an approximated approach require only two steps. First we find Tw. Second, we project each d E D into the low dimensional space defined by Tw , and use this more robust representation to extract document clusters TD. Approximately, we are trying to find Tw and TD that will maximize I(Tw; TD)' This two-phase IB algorithm was shown in [12] to be significantly superior to six other document clustering methods, when the performance are measured by the correlation of the obtained document clusters with the real newsgroup categories. Here we use the same procedure, but for finding Tw we take (3-1 = 0.15 (instead of zero). Using the above intuition we predict this will induce a cleaner representation for the document set. Indeed, the averaged correlation of TD (for lTD I = 4) with the original categories was 0.65, while for (3-1 = 0 it was 0.58 (the average is taken over different number of word clusters, ITw I = 10, 11...50). Similar results were obtained for all the 9 other subsets of the 20 newsgroups corpus described in [12]. As a second data set we used the gene expression measurements of rv 6800 genes in 72 samples of Leukemia [5]. The sample annotations included type of leukemia (ALL vs. AML), type of cells, source of sample, gender and donating hospital. We removed genes that were not expressed in the data and normalized the measurements of each sample to get a joint probability P(G, A) over genes and samples (with uniform prior on samples). We sorted all genes by their contribution to I(G; A) and chose the 500 most informative ones, which capture 47% of the original information, ending up with a joint probability with IAI = 72 and IGI = 500. We first used an exact implementation of the symmetric IB with alternating mergers be2We used the same subset already used in [12]. aotthetoand 0·04'---~----r~_~CC= , ,~ 905;O=w~ ord~ S 0.03 0.02 aCldvltammGaIClumlntakekldDey ... 0.04 c4,35words 0.03 0.02 algonthm secure secunty enayptlon ClaSSlIJed ... analog mOde signaimput output ... 0.04 c2, 20 words 0.04,---~-~~ c3~, 19;O=w~ ord~ S 0.03 0.02 ames planetary nasa spaceanane ... 0·04'---~-~~ C5o=" 35;O=w~ ord~ s 0.03 0.02 0.03 0.02 0.01 00 500 1000 1500 2000 sCience dataset, lnl0rmallon curves 1 .5~~~------, I ~-:=O I ~ 1 o ~ 0.5 IfTw·W\ Figure 2: P(D I tw) for 5 word clusters, tw E Tw. Documents 1 - 500 belong to sci. crypt category, 501 - 1000 to sci. electronics, 1001 - 1500 to sci.med and 1501 - 2000 to sci. space. In the title of each panel we see the 5 most frequent words in the cluster. The 'big' cluster (upper left panel) is clearly less informative about the structure of D. In the lower right panel we see the two information curves. Given some compression level, for (3- 1 = 0.15 we preserve much more information about D than for (3-1 = O. tween both clustering hierarchies (and /3-1 = 1). For ITA I = 2 we found an almost perfect correlation with the ALL vs. AML annotations (with only 4 exceptions). For ITA I = 8 and ITGI = 10 we found again high correlation between our sample clusters and the different sample annotations. For example, one cluster contained 10 samples that were all annotated as ALL type, taken from male patients in the same hospital. Almost all of these 10 were also annotated as T-cells, taken from bone marrow. Looking at p(TA I TG) we see that given the third genes cluster (which contained 17 genes) the probability of the above specific samples cluster is especially high. Further such analysis might yield additional insights about the structure of this data and will be presented elsewhere. Finally, to demonstrate the performance of the parallel IB we apply it to the same data. Using the parallel IB algorithm (with /3-1 = 0) we clustered the arrays A into two clustering hierarchies, T1 and T2 , that try together to capture the information about G. For ITj I = 4 we find that each I(Tj; G) preserve about 15% of the original information. However, taking ITj I = 2 (i.e. again, just 4 clusters) we see that the combination of the hierarchies, I(T1, T2 ; G), preserve 21 % of the original information. We then compared the two partitions we found against sample annotations. We found that the first hierarchy with IT11 = 2 almost perfectly match the split between B-cells and T-cells (among the 47 samples for which we had this annotation). The second hierarchy, with IT21 = 2 separates a cluster of 18 samples, almost all of which are ALL samples taken from the bone marrow of patients from the same hospital. These results demonstrate the ability of the algorithm to extract in parallel different meaningful independent partitions of the data. 7 Discussion The analysis presented by this work enables to implement a family of novel agglomerative clustering algorithms. All of these algorithms are motivated by one variational framework given by the multivariate IB method. Unlike most other clustering techniques, this is a principled model independent approach, which aims directly at the extraction of informative structures about given observed variables. It is thus very different from maximumlikelihood estimation of some mixture model and relies on fundamental information theoretic notions, similar to rate distortion theory and channel coding. In fact the multivariate IB can be considered as a multivariate coding result. The fundamental tradeoff between the compressed multi-information rGin and the preserved multi-information r G ou, provides a generalized coding limiting function, similar to the information curve in the original IB and to the rate distortion function in lossy compression. Despite the only local-optimality of the resulting solutions this information theoretic quantity - the fraction of the multiinformation that is extracted by the clusters - provides an objective figure of merit for the obtained clustering schemes. The suggested approach of this paper has several practical advantages over the 'deterministic annealing' algorithms suggested in [4], as it is simpler, fully deterministic and non-parametric. There is no need to identify cluster splits which is usually rather tricky. Though agglomeration procedures do not scale linearly with the sample size as top down methods do, there exist several heuristics to improve the complexity of these algorithms (e.g. [1]). While a typical initialization of an agglomerative procedure induces "hard" clustering solutions, all of the above analysis holds for "soft" clustering as well. Moreover, as already noted in [11], the obtained "hard" partitions can be used as a platform to find also "soft" solutions through a process of "reverse annealing". This raises the possibility for using an agglomerative procedure over "soft" clustering solutions, which we leave for future work. We could describe here only a few relatively simple examples. These examples show promising results on non trivial real life data. Moreover, other choices of Gin and Gout can yield additional novel algorithms with applications over a variety of data types. Acknowledgements This work was supported in part by the Israel Science Foundation (ISF), the Israeli Ministry of Science, and by the US-Israel Bi-national Science Foundation (BSF). N. Slonim was also supported by an Eshkol fellowship. N. Friedman was also supported by an Alon fellowship and the Harry & Abe Sherman Senior Lectureship in Computer Science. References [I] L. D. Baker and A. K. McCallum. Distributional clustering of words for text classification. In ACM SIGIR 98. [2] T. M. Cover and J. A. Thomas. Elements of Information Theory. 1991. [3] R. EI-Yaniv, S. Fine, and N. Tishby. Agnostic classification of Markovian sequences. In NIPS'97. [4] N. Friedman, O. Mosenzon, N. Sionim and N. Tishby Multivariate Infonnation Bottleneck UAI,2001. [5] T. Golub, D. Slonim, P. Tamayo, C.M. Huard, J.M. Caasenbeek, H. Coller, M. Loh, J. Downing, M. Caligiuri, C. Bloomfield, and E. Lander. Molecular classification of cancer: class discovery and class prediction by gene expression monitoring Science 286, 531- 537,1999. [6] K. Lang. Learning to filter netnews. In ICML'95. [7] J. Lin. Divergence Measures Based on the Shannon Entropy. IEEE Trans. Info. Theory, 37(1):145-151 , 1991. [8] J. Pearl. Probabilistic Reasoning in Intelligent Systems. 1988. [9] K. Rose. Detenninistic annealing for clustering, compression, classification, regression, and related optimization problems. Proc. IEEE, 86:2210--2239,1998. [10] N. Sionim, R. Somerville, N. Tishby, and O. Lahav. Objective spectral classification of galaxies using the infonnation bottleneck method. in "Monthly Notices of the Royal Astronomical Society", MNRAS, 323, 270, 2001. [II] N. Slonim and N. Tishby. Agglomerative Infonnation Bottleneck. In NIPS'99. [12] N. Sionim and N. Tishby. Document clustering using word clusters via the infonnation bottleneck method. InACM SIGIR 2000. [13] N. Slonim and N. Tishby. The power of word clusters for text classification. In ECIR, 2001. [14] N. Tishby, F. Pereira, and W. Bialek. The Infonnation Bottleneck method. In Proc. 37th Allerton Conference on Communication and Computation. 1999. [15] N. Tishby and N. Slonim. Data clustering by markovian relaxation and the infonnation bottleneck method. In NIPS'OO.
|
2001
|
19
|
2,005
|
Duality, Geometry, and Support Vector Regression Jinbo Bi and Kristin P. Bennett Department of Mathematical Sciences Rensselaer Polytechnic Institute Troy, NY 12180 bij2@rpi.edu, bennek@rpi.edu Abstract We develop an intuitive geometric framework for support vector regression (SVR). By examining when ϵ-tubes exist, we show that SVR can be regarded as a classification problem in the dual space. Hard and soft ϵ-tubes are constructed by separating the convex or reduced convex hulls respectively of the training data with the response variable shifted up and down by ϵ. A novel SVR model is proposed based on choosing the max-margin plane between the two shifted datasets. Maximizing the margin corresponds to shrinking the effective ϵ-tube. In the proposed approach the effects of the choices of all parameters become clear geometrically. 1 Introduction Support Vector Machines (SVMs) [6] are a very robust methodology for inference with minimal parameter choices. Intuitive geometric formulations exist for the classification case addressing both the error metric and capacity control [1, 2]. For linearly separable classification, the primal SVM finds the separating plane with maximum hard margin between two sets. The equivalent dual SVM computes the closest points in the convex hulls of the data from each class. For the inseparable case, the primal SVM optimizes the soft margin of separation between the two classes. The corresponding dual SVM finds the closest points in the reduced convex hulls. In this paper, we derive analogous arguments for SVM regression (SVR). We provide a geometric explanation for SVR with the ϵ-insensitive loss function. From the primal perspective, a linear function with no residuals greater than ϵ corresponds to an ϵ-tube constructed about the data in the space of the data attributes and the response variable [6] (see e.g. Figure 1(a)). The primary contribution of this work is a novel geometric interpretation of SVR from the dual perspective along with a mathematically rigorous derivation of the geometric concepts. In Section 2, for a fixed ϵ > 0 we examine the question “When does a “perfect” or “hard” ϵ-tube exist?”. With duality analysis, the existence of a hard ϵ-tube depends on the separability of two sets. The two sets consist of the training data augmented with the response variable shifted up and down by ϵ. In the dual space, regression becomes the classification problem of distinguishing between these two sets. The geometric formulations developed for the classification case [1] become applicable to the regression case. We call the resulting formulation convex SVR (C-SVR) since it is based on convex hulls of the augmented training data. Much like in SVM classification, to compute a hard ϵ-tube, C-SVR computes the nearest points in the convex hulls of the augmented classes. The corresponding maximum margin (max-margin) planes define the effective ϵ-tube. The size of margin determines how much the effective ϵ-tube shrinks. Similarly, to compute a soft ϵ-tube, reduced-convex SVR (RC-SVR) finds the closest points in the reduced convex hulls of the two augmented sets. This paper introduces the geometrically intuitive RC-SVR formulation which is a variation of the classic ϵ-SVR [6] and ν-SVR models [5]. If parameters are properly tuned, the methods perform similarly although not necessarily identically. RCSVR eliminates the pesky parameter C used in ϵ-SVR and ν-SVR. The geometric role or interpretation of C is not known for these formulations. The geometric roles of the two parameters of RC-SVR, ν and ϵ, are very clear, facilitating model selection, especially for nonexperts. Like ν-SVR, RC-SVR shrinks the ϵ-tube and has a parameter ν controlling the robustness of the solution. The parameter ϵ acts as an upper bound on the size of the allowable ϵ-insensitive error function. In addition, RC-SVR can be solved by fast and scalable nearest-point algorithms such as those used in [3] for SVM classification. 2 When does a hard ϵ-tube exist? y x ε ε (a) x y D D + (b) x y (c) D D + x y D D + (d) Figure 1: The (a) primal hard ϵ0-tube, and dual cases: (b) dual strictly separable ϵ > ϵ0, (c) dual separable ϵ = ϵ0, and (d) dual inseparable ϵ < ϵ0. SVR constructs a regression model that minimizes some empirical risk measure regularized to control capacity. Let x be the n predictor variables and y the dependent response variable. In [6], Vapnik proposed using the ϵ-insensitive loss function Lϵ(x, y, f) = |y −f(x)|ϵ = max(0, |y −f(x)| −ϵ), in which an example is in error if its residual |y −f(x)| is greater than ϵ. Plotting the points in (x, y) space as in Figure 1(a), we see that for a “perfect” regression model the data fall in a hard ϵ-tube about the regression line. Let (Xi, yi) be an example where i = 1, 2, · · · , m, Xi is the ith predictor vector, and yi is its response. The training data are then (X, y) where Xi is a row of the matrix X ∈Rm×n and y ∈Rm is the response. A hard ϵ-tube for a fixed ϵ > 0 is defined as a plane y = w′x+b satisfying −ϵe ≤y −Xw −be ≤ϵe where e is an m-dimensional vector of ones. When does a hard ϵ-tube exist? Clearly, for ϵ large enough such a tube always exists for finite data. The smallest tube, the ϵ0-tube, can be found by optimizing: min w,b,ϵϵ s.t. −ϵe ≤y −Xw −be ≤ϵe (1) Note that the smallest tube is typically not the ϵ-SVR solution. Let D+ and D−be formed by augmenting the data with the response variable respectively increased and decreased by ϵ, i.e. D+ = {(Xi, yi + ϵ), i = 1, · · · , m} and D−= {(Xi, yi − ϵ), i = 1, · · · , m}. Consider the simple problem in Figure 1(a). For any fixed ϵ > 0, there are three possible cases: ϵ > ϵ0 in which strict hard ϵ-tubes exist, ϵ = ϵ0 in which only ϵ0-tubes exist, and ϵ < ϵ0 in which no hard ϵ-tubes exist. A strict hard ϵ-tube with no points on the edges of the tube only exists for ϵ > ϵ0. Figure 1(b-d) illustrates what happens in the dual space for each case. The convex hulls of D+ and D−are drawn along with the max-margin plane in (b) and the supporting plane in (c) for separating the convex hulls. Clearly, the existence of the tube is directly related to the separability of D+ and D−. If ϵ > ϵ0 then a strict tube exists and the convex hulls of D+ and D−are strictly separable1. There are infinitely many possible ϵ-tubes when ϵ > ϵ0. One can see that the max-margin plane separating D+ and D−corresponds to one such ϵ. In fact this plane forms an ˆϵ tube where ϵ > ˆϵ ≥ϵ0. If ϵ = ϵ0, then the convex hulls of D+ and D−are separable but not strictly separable. The plane that separates the two convex hulls forms the ϵ0 tube. In the last case, where ϵ < ϵ0, the two sets D+ and D−intersect. No ϵ-tubes or max-margin planes exist. It is easy to show by construction that if a hard ϵ-tube exists for a given ϵ > 0 then the convex hulls of D+ and D−will be separable. If a hard ϵ-tube exists, then there exists (w, b) such that (y + ϵe) −Xw −be ≥0, (y −ϵe) −Xw −be ≤0. (2) For any convex combination of D+, X′ (y+ϵe)′ u where e′u = 1, u ≥0 of points (Xi, yi + ϵ), i = 1, 2, · · · , m, we have (y + ϵe)′u −w′(X′u) −b ≥0. Similarly for D−, X′ (y−ϵe)′ v where e′v = 1, v ≥0 of points (Xi, yi −ϵ), i = 1, 2, · · · , m, we have (y−ϵe)′v−w′(X′v)−b ≤0. Then the plane y = w′x+b in the ϵ-tube separates the two convex hulls. Note the separating plane and the ϵ-tube plane are the same. If no separating plane exists, then there is no tube. Gale’s Theorem2 of the alternative can be used to precisely characterize the ϵ-tube. Theorem 2.1 (Conditions for existence of hard ϵ-tube) A hard ϵ-tube exists for a given ϵ > 0 if and only if the following system in (u, v) has no solution: X′u = X′v, e′u = e′v = 1, (y + ϵe)′u −(y −ϵe)′v < 0, u ≥0, v ≥0. (3) Proof A hard ϵ-tube exists if and only if System (2) has a solution. By Gale’s Theorem of the alternative [4], system (2) has a solution if and only if the following alternative system has no solution: X′u = X′v, e′u = e′v, (y+ϵe)′u−(y −ϵe)′v = −1, u ≥0, v ≥0. Rescaling by 1 σ where σ = e′u = e′v > 0 yields the result. 1We use the following definitions of separation of convex sets. Let D+ and D−be nonempty convex sets. A plane H = {x : w′x = α} is said to separate D+ and D−if w′x ≥α, ∀x ∈D+ and w′x ≤α, ∀x ∈D−. H is said to strictly separate D+ and D−if w′x ≥α + ∆for x ∈D+, and w′x ≤α −∆for each x ∈D−where ∆is a positive scalar. 2The system Ax ≤c has a (or has no) solution if and only if the alternative system A′y = 0, c′y = −1, y ≥0 has no (or has a) solution. Note that if ϵ ≥ϵ0 then (y + ϵe)′u −(y −ϵe)′v ≥0. for any (u, v) such that X′u = X′v, e′u = e′v = 1, u, v ≥0. So as a consequence of this theorem, if D+ and D−are separable, then a hard ϵ-tube exists. 3 Constructing the ϵ-tube For any ϵ > ϵ0 infinitely many possible ϵ-tubes exist. Which ϵ-tube should be used? The linear program (1) can be solved to find the smallest ϵ0-tube. But this corresponds to just doing empirical risk minimization and may result in poor generalization due to overfitting. We know capacity control or structural risk minimization is fundamental to the success of SVM classification and regression. We take our inspiration from SVM classification. In hard-margin SVM classification, the dual SVM formulation constructs the max-margin plane by finding the two nearest points in the convex hulls of the two classes. The max-margin plane is the plane bisecting these two points. We know that the existence of the tube is linked to the separability of the shifted sets, D+ and D−. The key insight is that the regression problem can be regarded as a classification problem between D+ and D−. The two sets D+ and D−defined as in Section 2 both contain the same number of data points. The only significant difference occurs along the y dimension as the response variable y is shifted up by ϵ in D+ and down by ϵ in D−. For ϵ > ϵ0, the max-margin separating plane corresponds to a hard ˆϵ-tube where ϵ > ˆϵ ≥ϵ0. The resulting tube is smaller than ϵ but not necessarily the smallest tube. Figure 1(b) shows the max-margin plane found for ϵ > ϵ0. Figure 1(a) shows that the corresponding linear regression function for this simple example turns out to be the ϵ0 tube. As in classification, we will have a hard and soft ϵ-tube case. The soft ϵ-tube with ϵ ≤ϵ0 is used to obtain good generalization when there are outliers. 3.1 The hard ϵ-tube case We now apply the dual convex hull method to constructing the max-margin plane for our augmented sets D+ and D−assuming they are strictly separable, i.e. ϵ > ϵ0. The problem is illustrated in detail in Figure 2. The closest points of D+ and D−can be found by solving the following dual C-SVR quadratic program: min u,v 1 2
X′ (y+ϵe)′ u − X′ (y−ϵe)′ v
2 s.t. e′u = 1, e′v = 1, u ≥0, v ≥0. (4) Let the closest points in the convex hulls of D+ and D−be c = X′ (y+ϵe)′ ˆu and d = X′ (y−ϵe)′ ˆv respectively. The max-margin separating plane bisects these two points. The normal (ˆw, ˆδ) of the plane is the difference between them, i.e., ˆw = X′ˆu −X′ˆv, ˆδ = (y + ϵe)′ˆu −(y −ϵe)′ˆv. The threshold, ˆb, is the distance from the origin to the point halfway between the two closest points along the normal: ˆb = ˆw′ X′ˆu+X′ˆv 2 +ˆδ y′ˆu+y′ˆv 2 . The separating plane has the equation ˆw′x+ˆδy−ˆb = 0. Rescaling this plane yields the regression function. Dual C-SVR (4) is in the dual space. The corresponding Primal C-SVR is:
Figure 2: The solution ˆϵ-tube found by C-SVR can have ˆϵ < ϵ. Squares are original data. Dots are in D+. Triangles are in D−. Support Vectors are circled. min w,δ,α,β 1 2 ∥w∥2 + 1 2δ2 −(α −β) s.t. Xw + δ(y + ϵe) −αe ≥0 Xw + δ(y −ϵe) −βe ≤0. (5) Dual C-SVR (4) can be derived by taking the Wolfe or Lagrangian dual [4] of primal C-SVR (5) and simplifying. We prove that the optimal plane from C-SVR bisects the ˆϵ tube. The supporting planes for class D+ and class D−determines the lower and upper edges of the ˆϵ-tube respectively. The support vectors from D+ and D−correspond to the points along the lower and upper edges of the ˆϵ-tube. See Figure 2. Theorem 3.1 (C-SVR constructs ˆϵ-tube) Let the max-margin plane obtained by C-SVR (4) be ˆw′x+ˆδy−ˆb = 0 where ˆw = X′ˆu−X′ˆv, ˆδ = (y+ϵe)′ˆu−(y−ϵe)′ˆv, and ˆb = ˆw′ X′ˆu+X′ˆv 2 + ˆδ y′ˆu+y′ˆv 2 . If ϵ > ϵ0, then the plane y = w′x + b corresponds to an ˆϵ-tube of training data (Xi, yi), i = 1, 2, · · · , m where w = −ˆw ˆδ , b = ˆb ˆδ and ˆϵ = ϵ −ˆα−ˆβ 2ˆδ < ϵ. Proof First, we show ˆδ > 0. By the Wolfe duality theorem [4], ˆα −ˆβ > 0, since the objective values of (5) and the negative objective value of (4) are equal at optimality. By complementarity, the closest points are right on the margin planes ˆw′x + ˆδy −ˆα = 0 and ˆw′x + ˆδy −ˆβ = 0 respectively, so ˆα = ˆw′X′ˆu + ˆδ(y + ϵe)′ˆu and ˆβ = ˆw′X′ˆv+ˆδ(y−ϵe)′ˆv. Hence ˆb = ˆα+ ˆβ 2 , and ˆw, ˆδ, ˆα, and ˆβ satisfy the constraints of problem (5), i.e., Xˆw+ˆδ(y+ϵe)−ˆαe ≥0, Xˆw+ˆδ(y−ϵe)−ˆβe ≤0. Then subtract the second inequality from the first inequality: 2ˆδϵ −ˆα + ˆβ ≥0, that is, ˆδ ≥ˆα−ˆβ 2ϵ > 0 because ϵ > ϵ0 ≥0. Rescale constraints by −ˆδ < 0, and reverse the signs. Let w = −ˆw ˆδ , then the inequalities become Xw −y ≤ϵe −ˆα ˆδ e, Xw −y ≥−ϵe − ˆβ ˆδ e. Let b = ˆb ˆδ, then ˆα ˆδ = b + ˆα−ˆβ 2ˆδ and ˆβ ˆδ = b −ˆα−ˆβ 2ˆδ . Substituting into the previous inequalities yields Xw−y ≤ ϵ −ˆα−ˆβ 2ˆδ e−be, Xw−y ≥− ϵ −ˆα−ˆβ 2ˆδ e−be. Denote ˆϵ = ϵ−ˆα−ˆβ 2ˆδ < ϵ. These inequalities become Xw+be−y ≤ˆϵe, Xw+be−y ≥−ˆϵe. Hence the plane y = w′x + b is in the middle of the ˆϵ < ϵ tube. 3.2 The soft ϵ-tube case For ϵ < ϵ0, a hard ϵ-tube does not exist. Making ϵ large to fit outliers may result in poor overall accuracy. In soft-margin classification, outliers were handled in the
2ε^ y x Figure 3: Soft ˆϵ-tube found by RC-SVR: left: dual, right: primal space. dual space by using reduced convex hulls. The same strategy works for soft ϵ-tubes, see Figure 3. Instead of taking the full convex hulls of D+ and D−, we reduce the convex hulls away from the difficult boundary cases. RC-SVR computes the closest points in the reduced convex hulls min u,v 1 2
X′ (y+ϵe)′ u − X′ (y−ϵe)′ v
2 s.t. e′u = 1, e′v = 1, 0 ≤u ≤De, 0 ≤v ≤De. (6) Parameter D determines the robustness of the solution by reducing the convex hull. D limits the influence of any single point. As in ν-SVM, we can parameterize D by ν. Let D = 1 νm where m is the number of points. Figure 3 illustrates the case for m = 6 points, ν = 2/6, and D = 1/2. In this example, every point in the reduced convex hull must depend on at least two data points since Pm i=1 ui = 1 and 0 ≤ui ≤1/2. In general, every point in the reduced convex hull can be written as the convex combination of at least ⌈1/D⌉= ⌈ν ∗m⌉. Since these points are exactly the support vectors and there are two reduced convex hulls, 2 ∗⌈νm⌉is a lower bound on the number of support vectors in RC-SVR. By choosing ν sufficiently large, the inseparable case with ϵ ≤ϵ0 is transformed into a separable case where once again our nearest-points-in-the-convex-hull-problem is well defined. As in classification, the dual reduced convex hull problem corresponds to computing a soft ϵ-tube in the primal space. Consider the following soft tube version of the primal C-SVR (7) which has its Wolfe Dual RC-SVR (6): min w,δ,α,β,ξ,η 1 2 ∥w∥2 + 1 2δ2 −(α −β) + C(e′ξ + e′η) s.t. Xw + δ(y + ϵe) −αe + ξ ≥0, ξ ≥0 Xw + δ(y −ϵe) −βe −η ≤0, η ≥0 (7) The results of Theorem 3.1 can be easily extended to soft ϵ-tubes. Theorem 3.2 (RC-SVR constructs soft ˆϵ-tube) Let the soft max-margin plane obtained by RC-SVR (6) be ˆw′x + ˆδy −ˆb = 0 where ˆw = X′ˆu −X′ˆv, ˆδ = (y+ϵe)′ˆu−(y −ϵe)′ˆv, and ˆb = X′ˆu+X′ˆv 2 ′ ˆw+ y′ˆu+y′ ˆv 2 ˆδ. If 0 < ϵ ≤ϵ0, then the plane y = w′x + b corresponds to a soft ˆϵ = ϵ −˜α−˜β 2ˆδ < ϵ-tube of training data (Xi, yi), i = 1, 2, · · · , m, i.e., a ˆϵ-tube of reduced convex hull of training data where w = −ˆw ˆδ , b = ˆb ˆδ and ˜α = ˆw′X′ˆu + ˆδ(y + ϵe)′ˆu, ˜β = ˆw′X′ˆv + ˆδ(y −ϵe)′ˆv. Notice that the ˜α and ˜β determine the planes parallel to the regression plane and through the closest points in each reduced convex hull of shifted data. In the inseparable case, these planes are parallel but not necessarily identical to the planes obtained by the primal RC-SVR (7). Nonlinear C-SVR and RC-SVR can be achieved by using the usual kernel trick. Let Φ by a nonlinear mapping of x such that k(Xi, Xj) = Φ(Xi) · Φ(Xj). The objective function of C-SVR (4) and RC-SVR (6) applied to the mapped data becomes 1 2 Pm i=1 Pm j=1 ((ui −vi)(uj −vj)(Φ(Xi) · Φ(Xj) + yiyj)) + 2ϵ Pm i=1 (yi(ui −vi)) = 1 2 Pm i=1 Pm j=1 ((ui −vi)(uj −vj)(k(Xi, Xj) + yiyj)) + 2ϵ Pm i=1 (yi(ui −vi)) (8) The final regression model after optimizing C-SVR or RC-SVR with kernels takes the form of f(x) = Pm i=1 (¯ui −¯vi) k(Xi, x) + ¯b, where ¯ui = ˆui ˆδ , ¯vi = ˆvi ˆδ , ˆδ = (ˆu − ˆv)′y+2ϵ, and the intercept term ¯b = (ˆu+ˆv)′K(ˆu−ˆv) 2ˆδ + (ˆu+ˆv)′y 2 where Kij = k(Xi, Xj). 4 Computational Results We illustrate the difference between RC-SVR and ϵ-SVR on a toy linear problem3. Figure 4 depicts the functions constructed by RC-SVR and ϵ-SVR for different values of ϵ. For large ϵ, ϵ-SVR produces undesirable results. RC-SVR constructs the same function for ϵ sufficiently large. Too small ϵ can result in poor generalization. −1 0 1 2 3 4 5 6 −0.5 0 0.5 1 1.5 2 2.5 ε = 0.75 ε = 0.45 ε = 0.25 ε = 0.15 (a) −1 0 1 2 3 4 5 6 −0.5 0 0.5 1 1.5 2 2.5 ε = 0.75, 0.45, 0.25 (b) Figure 4: Regression lines from (a) ϵ-SVR and (b) RC-SVR with distinct ϵ. In Table 1, we compare RC-SVR, ϵ-SVR and ν-SVR on the Boston Housing problem. Following the experimental design in [5] we used RBF kernel with 2σ2 = 3.9, C = 500·m for ϵ-SVR and ν-SVR, and ϵ = 3.0 for RC-SVR. RC-SVR, ϵ-SVR, and ν-SVR are computationally similar for good parameter choices. In ϵ-SVR, ϵ is fixed. In RC-SVR, ϵ is the maximum allowable tube width. Choosing ϵ is critical for ϵ-SVR but less so for RC-SVR. Both RC-SVR and ν-SVR can shrink or grow the tube according to desired robustness. But ν-SVR has no upper ϵ bound. 5 Conclusion and Discussion By examining when ϵ-tubes exist, we showed that in the dual space SVR can be regarded as a classification problem. Hard and soft ϵ-tubes are constructed by separating the convex or reduced convex hulls respectively of the training data with the response variable shifted up and down by ϵ. We proposed RC-SVR based on choosing the soft max-margin plane between the two shifted datasets. Like ν-SVM, RC-SVR shrinks the ϵ-tube. The max-margin determines how much the tube can shrink. Domain knowledge can be incorporated into the RC-SVR parameters ϵ 3The data consist of (x,y): (0 0), (1 0.1), (2 0.7), (2.5 0.9), (3 1.1) and (5 2). The CPLEX 6.6 optimization package was used. Table 1: Testing Results for Boston Housing, MSE= average of mean squared errors of 25 testing points over 100 trials, STD: standard deviation 2ν 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 RC-SVR MSE 37.3 11.2 10.7 9.6 8.9 10.6 11.5 12.5 STD 72.3 7.6 7.3 7.4 8.4 9.1 9.3 9.8 ϵ 0 1 2 3 4 5 6 7 ϵ-SVR MSE 11.2 10.8 9.5 10.3 11.6 13.6 15.6 17.2 STD 8.3 8.2 8.2 7.3 5.8 5.8 5.9 5.8 ν 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 ν-SVR MSE 9.6 8.9 9.5 10.8 10.9 11.0 11.2 11.1 STD 5.8 7.9 8.3 8.2 8.3 8.4 8.5 8.4 and ν. The parameter C in ν-SVM and ϵ-SVR has been eliminated. Computationally, no one method is superior for good parameter choices. RC-SVR alone has a geometrically intuitive framework that allows users to easily grasp the model and its parameters. Also, RC-SVR can be solved by fast nearest point algorithms. Considering regression as a classification problem suggests other interesting SVR formulations. We can show ϵ-SVR is equivalent to finding closest points in a reduced convex hull problem for certain C, but the equivalent problem utilizes a different metric in the objective function than RC-SVR. Perhaps other variations would yield even better formulations. Acknowledgments Thanks to referees and Bernhard Sch¨olkopf for suggestions to improve this work. This work was supported by NSF IRI-9702306, NSF IIS-9979860. References [1] K. Bennett and E. Bredensteiner. Duality and Geometry in SVM Classifiers. In P. Langley, eds., Proc. of Seventeenth Intl. Conf. on Machine Learning, p 57–64, Morgan Kaufmann, San Francisco, 2000. [2] D. Crisp and C. Burges. A Geometric Interpretation of ν-SVM Classifiers. In S. Solla, T. Leen, and K. Muller, eds., Advances in Neural Info. Proc. Sys., Vol 12. p 244–251, MIT Press, Cambridge, MA, 1999. [3] S.S. Keerthi, S.K. Shevade, C. Bhattacharyya and K.R.K. Murthy, A Fast Iterative Nearest Point Algorithm for Support Vector Machine Classifier Design, IEEE Transactions on Neural Networks, Vol. 11, pp.124-136, 2000. [4] O. Mangasarian. Nonlinear Programming. SIAM, Philadelphia, 1994. [5] B. Sch¨olkopf, P. Bartlett, A. Smola and R. Williamson. Shrinking the Tube: A New Support Vector Regression Algorithm. In M. Kearns, S. Solla, and D. Cohn eds., Advances in Neural Info. Proc. Sys., Vol 12, MIT Press, Cambridge, MA, 1999. [6] V. Vapnik. The Nature of Statistical Learning Theory. Wiley, New York, 1995.
|
2001
|
190
|
2,006
|
BLIND SOURCE SEPARATION VIA MULTINODE SPARSE REPRESENTATION Michael Zibulevsky Department of Electrical Engineering Technion, Haifa 32000, Israel mzib@ee.technion.ac. if Yehoshua Y. Zeevi Department of Electrical Engineering Technion, Haifa 32000, Israel zeevi@ee.technion.ac. if Pavel Kisilev Department of Electrical Engineering Technion, Haifa 32000, Israel paufk@tx.technion.ac. if Barak Pearlmutter Department of Computer Science University of New Mexico Albuquerque, NM 87131 USA bap@cs. unm. edu Abstract We consider a problem of blind source separation from a set of instantaneous linear mixtures, where the mixing matrix is unknown. It was discovered recently, that exploiting the sparsity of sources in an appropriate representation according to some signal dictionary, dramatically improves the quality of separation. In this work we use the property of multi scale transforms, such as wavelet or wavelet packets, to decompose signals into sets of local features with various degrees of sparsity. We use this intrinsic property for selecting the best (most sparse) subsets of features for further separation. The performance of the algorithm is verified on noise-free and noisy data. Experiments with simulated signals, musical sounds and images demonstrate significant improvement of separation quality over previously reported results. 1 Introduction In the blind source separation problem an N-channel sensor signal x(~ ) is generated by M unknown scalar source signals s rn(~) , linearly mixed together by an unknown N x M mixing, or crosstalk, matrix A, and possibly corrupted by additive noise n(~): x(~) = As(~) + n(~ ). (1) The independent variable ~ is either time or spatial coordinates in the case of images. We wish to estimate the mixing matrix A and the M-dimensional source signal s(~). The assumption of statistical independence of the source components Srn(~) , m = 1, ... , M leads to the Independent Component Analysis (lCA) [1], [2]. A stronger assumption is the °Supported in part by the Ollendorff Minerva Center, by the Israeli Ministry of Science, by NSF CAREER award 97-02-311 and by the National Foundation for Functional Brain Imaging sparsity of decomposition coefficients, when the sources are properly represented [3]. In particular, let each 8 m (~ ) have a sparse representation obtained by means of its decomposition coefficients Cmk according to a signal dictionary offunctions Y k (~ ): 8m (~ ) = L Cmk Yk(~ )' (2) k The functions Yk (~ ) are called atoms or elements of the dictionary. These elements do not have to be linearly independent, and instead may form an overcomplete dictionary, e.g. wavelet-related dictionaries (wavelet packets, stationary wavelets, etc., see for example [9]). Sparsity means that only a small number of coefficients Cmk differ significantly from zero. Then, unmixing of the sources is performed in the transform domain, i.e. in the domain of these coefficients Cmk. The property of sparsity often yields much better source separation than standard ICA, and can work well even with more sources than mixtures. In many cases there are distinct groups of coefficients, wherein sources have different sparsity properties. The key idea in this study is to select only a subset of features (coefficients) which is best suited for separation, with respect to the following criteria: (1) sparsity of coefficients (2) separability of sources' features. After this subset is formed, one uses it in the separation process, which can be accomplished by standard ICA algorithms or by clustering. The performance of our approach is verified on noise-free and noisy data. Our experiments with ID signals and images demonstrate that the proposed method further improves separation quality, as compared with result obtained by using sparsity of all decomposition coefficients. 2 Two approaches to sparse source separation: InfoMax and Clustering Sparse sources can be separated by each one of several techniques, e.g. the Bell-Sejnowski Information Maximization (BS InfoMax) approach [1], or by approaches based on geometric considerations (see for example [8]). In the former case, the algorithm estimates the unmixing matrix W = A - I, while in the later case the output is the estimated mixing matrix. In both cases, these matrices can be estimated only up to a column permutation and a scaling factor [4]. InfoMax. Under the assumption of a noiseless system and a square mixing matrix in (1), the BS InfoMax is equivalent to the maximum likelihood (ML) formulation of the problem [4], which is used in this section. For the sake of simplicity of the presentation, let us consider the case where the dictionary of functions used in a source decomposition (2) is an orthonormal basis. (In this case, the corresponding coefficients Cmk =< 8m, 'Pk >, where < ',' > denotes the inner product). From (1) and (2) the decomposition coefficients of the noiseless mixtures, according to the same signal dictionary of functions Y k (~ ) ' are: Ak= ACk, (3) where M -dimensional vector Ck forms the k-th column of the matrix C = { Cmk}. Let Y be thefeatures, or (new) data, matrix of dimension M x K , where K is the number of features. Its rows are either the samples of sensor signals (mixtures), or their decomposition coefficients. In the later case, the coefficients Ak'S form the columns ofY. (In the following discussion we assume this setting for Y , if not stated other). We are interested in the maximum likelihood estimate of A given the data Y. Let the corresponding coefficients Cmk be independent random variables with a probability density function (pdf) of an exponential type (4) where the scalar function v(·) is a smooth approximation of an absolute value function. Such kind of distribution is widely used for modeling sparsity [5]. In view of the independence of Cmk, and (4), the prior pdf of C is p(C) ex II exp{ - V(Cmk)}. (5) m,k Taking into account that Y = AC, the parametric model for the pdf of Y with respect to parameters A is (6) Let W = A -I be the unmixing matrix, to be estimated. Then, substituting C = WY, combining (6) with (5) and taking the logarithm we arrive at the log-likelihood function: M K Lw(Y) = Klog ldetWI- L LV((WY)mk). (7) m=l k = l Maximization of Lw(Y) with respect to W is equivalent to the BS InfoMax, and can be solved efficiently by the Natural Gradient algorithm [6]. We used this algorithm as implemented in the ICAlEEG Matlab toolbox [7]. Clustering. In the case of geometry based methods, separation of sparse sources can be achieved by clustering along orientations of data concentration in the N-dimensional space wherein each column Yk of the matrix Y represents a data point (N is the number of mixtures). Let us consider a two-dimensional noiseless case, wherein two source signals, Sl(t) and S2(t), are mixed by a 2x2 matrix A, arriving at two mixtures Xl(t) and X2(t). (Here, the data matrix is constructed from these mixtures Xl (t) and xd t)). Typically, a scatter plot of two sparse mixtures X1(t) versus X2(t), looks like the rightmost plot in Figure 2. If only one source, say Sl (t), was present, the sensor signals would be Xl (t) = all Sl (t) and X2(t) = a21s1 (t) and the data points at the scatter diagram of Xl (t) versus X2(t) would belong to the straight line placed along the vector [ana21]T. The same thing happens, when two sparse sources are present. In this sparse case, at each particular index where a sample of the first source is large, there is a high probability, that the corresponding sample of the second source is small, and the point at the scatter diagram still lies close to the mentioned straight line. The same arguments are valid for the second source. As a result, data points are concentrated around two dominant orientations, which are directly related to the columns of A. Source signals are rarely sparse in their original domain. In contrast, their decomposition coefficients (2) usually show high sparsity. Therefore, we construct the data matrix Y from the decomposition coefficients of mixtures (3), rather than from the mixtures themselves. In order to determine orientations of scattered data, we project the data points onto the surface of a unit sphere by normalizing corresponding vectors, and then apply a standard clustering algorithm. This clustering approach works efficiently even if the number of sources is greater than the number of sensors. Our clustering procedure can be summarized as follows: 1. Form the feature matrix Y , by putting samples of the sensor signals or (subset of) their decomposition coefficients into the corresponding rows ofthe matrix; 2. Normalize feature vectors (columns ofY): Yk = Yk/II Yk I12' in order to project data points onto the surface of a unit sphere, where 11 · 11 2 denotes the l2 norm. Before nonnalization, it is reasonable to remove data points with a very small norm, since these very likely to be crosstalk-corrupted by small coefficients from others' sources. 3. Move data points to a half-sphere, e.g. by forcing the sign of the first coordinate yk to be positive: IF yk < 0 THEN Yk = - Yk. Without this operation each set oflineariy (i.e., along a line) clustered data points would yield two clusters on opposite sides of the sphere. · : -s tOO 200 300 ~oo »:I eoo 700 ~ 900 tOC>:l ,: 5 100 200 300 ~OO 500 600 700 1\00 900 1000 , : -s tOO 200 300 ~OO 500 eoo 700 1\00 900 tOC>:l , : -5 100 200 300 ~OO 500 600 700 1\00 900 1():Xl Figure 1: Random block signals (two upper) and their mixtures (two lower) 4. Estimate cluster centers by using a clustering algorithm. The coordinates of these centers will form the columns of the estimated mixing matrix A. We used Fuzzy C-means (FCM) clustering algorithm as implemented in Matlab Fuzzy Logic Toolbox. Sources recovery. The estimated unmixing matrix A-I is obtained by either the BS InfoMax or the above clustering procedure, applied to either complete data set, or to some subsets of data (to be explained in the next section). Then, the sources are recovered in their original domain by s(t) = A - lX(t). We should stress here that if the clustering approach is used, the estimation of sources is not restricted to the case of square mixing matrices, although the sources recovery is more complicated in the rectangular cases (this topic is out of scope of this paper). 3 Multinode based source separation Motivating example: sparsity of random blocks in the Haar basis. To provide intuitive insight into the practical implications of our main idea, we first use ID block functions, that are piecewise constant, with random amplitude and duration of each constant piece (Figure 1). It is known, that the Haar wavelet basis provides compact representation of such functions. Let us take a close look at the Haar wavelet coefficients at different resolution levels j =O, 1, ... ,1. Wavelet basis functions at the finest resolution level j =J are obtained by translation of the Haar mother wavelet: <p(t) = {I, ift E [0, 1); - I , ift E [1, 2); 0 otherwise}. Taking the scalar product ofa function s(t) with the wavelet <PJ(t - T), we produce a finite differentiation of the function s(t) at the point t = T. This means that the number of non-zero coefficients at the finest resolution for a block function will correspond roughly to the number of jumps ofthis function. Proceeding to the next, coarser resolution level, we have <P J - l (t) = {I , ift E [0, 2); - 1, if t E [2,4); ° otherwise}. At this level, the number of non-zero coefficients still corresponds to the number of jumps, but the total number of coefficients at this level is halved, and so is the sparsity. If we further proceed to coarser resolutions, we will encounter levels where the support of a wavelet <Pj(t) is comparable to the typical distance between jumps in the function s(t). In this case, most of the coefficients are expected to be nonzero, and, therefore, sparsity will fade away. To demonstrate how this influences accuracy of a blind source separation, we randomly generated two block-signal sources (Figure 1, two upper plots.), and mixed them by the crosstalk matrix A with colwnns [0.83 -0.55] and [0.62 0.78]. Resulting sensor signals, or mixtures, X l (t) and X2 (t) are shown in the two lower plots of Figure l. The scatter plot of X l (t) versus X2( t) does not exhibit any visible distinct orientations (Figure 2, left). Similarly, in the scatter plot of the wavelet coefficients at the lowest resolution distinct orientations are hardly detectable (Figure 2, middle). In contrast, the scatter plot of the wavelet coefficients at the highest resolution (Figure 2, right) depicts two distinct orientations, which correspond to the columns of the mixing matrix. InfoMax FCM Raw signals : :"~~:·;K·;\. " , " "1 .::;; Of :~~.: • •••• l.93 1.78 All wavelet coefficients 0.183 0.058 High resolution WT coefficients 1·. ~ ."'~'>!~ . : >/_~: .~ -I ; /" , .,/', " "" 0.005 0.002 Figure 2: Separation of block signals: scatter plots of sensor signals (left), and of their wavelet coefficients (middle and right). Lower colwnns present the normalized meansquared separation error (%) corresponding to the Bell-Sejnowski InfoMax, and to the Fuzzy C-Means clustering, respectively. Since a crosstalk matrix A is estimated only up to a column permutation and a scaling factor, in order to measure the separation accuracy, we normalize the original sources sm(t) and their corresponding estimated sources sm(t). The averaged (over sources) normalized squared error (NSE) is then computed as: NSE = it 2:~= 1 (ilsm - sm ll§/llsmll§)· Resulting separation errors for block sources are presented in the lower part of Figure 2. The largest error (l.93%) is obtained on the raw data, and the smallest «0.005%) - on the wavelet coefficients at the highest resolution, which have the best sparsity. Using all wavelet coefficients yields intermediate sparsity and performance. Multinode representation. Our choice of a particular wavelet basis and of the sparsest subset of coefficients was obvious in the above example: it was based on knowledge of the structure of piecewise constant signals. For sources having oscillatory components (like sounds or images with textures), other systems of basis functions, such as wavelet packets and trigonometric function libraries [9], might be more appropriate. The wavelet packet library consists of the triple-indexed family of functions: i.f!j ,i,q(t) = 2j / 2i.f!q(2j t - i), j , i E Z , q E N,where j , i are the scale and shift parameters, respectively, and q is the frequency parameter. [Roughly speaking, q is proportional to the nwnber of oscillations of a mother wavelet i.f!q(t)]. These functions form a binary tree whose nodes are indexed by the depth of the level j and the node number q = 0, 1, 2, 3, ... , 2j - l at the specified level j. This same indexing is used for corresponding subsets of wavelet packet coefficients (as well as in scatter diagrams in the section on experimental results). Adaptive selection of sparse subsets. When signals have a complex nature, it is difficult to decide in advance which nodes contain the sparsest sets of coefficients. That is why we use the following simple adaptive approach. First, for every node of the tree, we apply our clustering algorithm, and compute a measure of clusters' distortion. In our experiments we used a standard global distortion, the mean squared distance of data points to the centers of their own (closest) clusters (here again, the weights of the data points can be incorporated): d=2:f=l min II U m - Yk II ,where K is the nwnber of data points, U m is the m-th centroid m coordinates, Yk is the k-th data point coordinates, and 11 . 11 is the sum-of-squares distance. Second, we choose a few best nodes with the minimal distortion, combine their coefficients into one data set, and apply a separation algorithm (clustering or Infomax) to these data. 4 Experimental results The proposed blind separation method based on the wavelet-packet representation, was evaluated by using several types of signals. We have already discussed the relatively simple example of a random block signal. The second type of signal is a frequency modulated (FM) sinusoidal signal. The carrier frequency is modulated by either a sinusoidal function (FM signal) or by random blocks (BFM signal). The third type is a musical recording of flute sounds. Finally, we apply our algorithm to images. An example of such images is presented in the left part of Figure 3. 111 '10 ' 11 '~ • • t: , ' ' 12 ' 13 'JJ SI •• .. '~' 00 0° • . '. , 8 . , foo 0 , . , • 8 ' 22 11 S. Ss ~ :. , :Y6~ , ",t, " ' ' 26 \; '21 'lI '11 t , "*, ' , :, '8 ' " Figure 3: Left: two source images (upper pair), their mixtures (middle pair) and estimated images (lower pair). Right: scatter plots ofthe wavelet packet (WP) coefficients of mixtures of images; subsets are indexed on the WP tree. In order to compare accuracy of our adaptive best nodes method with that attainable by standard methods, we form the following feature sets: (1) raw data, (2) Short Time Fourier Transform (STFT) coefficients (in the case of ID signals), (3) Wavelet Transform coefficients (4) Wavelet packet coefficients at the best nodes found by our method, while using various wavelet families with different smoothness (haar, db-4, db-S). In the case of image separation, we used the Discrete Cosine Transform (DCT) instead of the STFT, and the sym4 and symS mother wavelet instead of db-4 and db-S, when using wavelet transform and wavelet packets. The right part of Figure 3 presents an example of scatter plots of the wavelet packet coefficients obtained at various nodes of the wavelet packet tree. The upper left scatter plot, marked with 'C', corresponds to the complete set of coefficients at all nodes. The rest are the scatter plots of sets of coefficients indexed on a wavelet packet tree. Generally speaking, the more distinct the two dominant orientations appear on these plots, the more precise is the estimation of the mixing matrix, and, therefore, the better is the quality of separation. Note, that only two nodes, C22 and C23, show clear orientations. These nodes will most likely be selected by the algorithm for further estimation process. Signals raw STFT WT WT WP WP data db8 haar db8 haar Blocks 10.16 2.669 0.174 0.037 0.073 0.002 BFM sine 24.51 0.667 0.665 2.34 0.2 0.442 FM sine 25.57 0.32 1.032 6.105 0.176 0.284 Flutes 1.48 0.287 0.355 0.852 0.154 0.648 raw OCT WT WT WP WP Images data sym8 haar sym8 haar 4.88 3.651 l.l64 l.l14 0.365 0.687 Table 1: Experimental results: normalized mean-squared separation error (%) for noisefree signals and images, applying the FCM separation to raw data and decomposition coefficients in various domains. In the case of wavelet packets (WP) the best nodes selected by our algorithm were used. Table 1 summarizes results of experiments in which we applied our approach of the best features selection along with the FCM separation to each noise-free feature set. In these experiments, we compared the quality of separation of deterministic signals by calculating N SE's (i.e., residual crosstalk errors). In the case of random block and BFM signals, we performed 100 Monte-Carlo simulations and calculated the normalized mean-squared errors (N M SE) for the above feature sets. From Table 1 it is clear that using our adaptive best nodes method outperforms all other feature sets (including complete set of wavelet coefficients), for each type of signals. Similar improvement was achieved by using our method along with the BS InfoMax separation, which provided even better results for images. In the case of the random block signals, using the Haar wavelet function for the wavelet packet representation yields a better separation than using some smooth wavelet, e.g. db-S. The reason is that these block signals, that are not natural signals, have a sparser representation in the case of the Haar wavelets. In contrast, as expected, natural signals such as the Flute's signals are better represented by smooth wavelets, that in turn provide a better separation. This is another advantage of using sets of features at multiple nodes along with various families of 'mother' functions: one can choose best nodes from several decomposition trees simultaneously. In order to verify the performance of our method in presence of noise, we added various types of noise (white gaussian and salt&pepper) to three mixtures of three images at various signal-to-noise energy ratios (SNR). Table 2 summarizes these experiments in which we applied our approach along with the BS InfoMax separation. It turns out that the ideas used in wavelet based signal denoising (see for example [10] and references therein), are applied to signal separation from noisy mixtures. In particular, in case of white gaussian noise, the noise energy is uniformly distributed over all wavelet coefficients at various scales. Therefore, at sufficiently high SNR's, the large coefficients of the signals are only slightly distorted by the noise coefficients, and the estimation of the unmixing matrix is almost not affected by the presence of noise. (In contrast, the BS InfoMax applied to three noisy mixtures themselves, failed completely, arriving at N S E of 19% even in the case of SNR=12dB). We should stress here that, although our adaptive best nodes method performs reasonably well in the presence of noise, it is not supposed to further denoise the reconstructed images (this can be achieved by some denoising method, after source signals are separated). More experimental results, as well as parameters of simulations, can be found in [11]. SNR [dB] Mixtures w. white gaussian noise Mixtures w. salt&pepper noise Table 2: Perfonnance of the algorithm in presence of various sources of noise in mixtures of images: nonnalized mean-squared separation error (%), applying our adaptive approach along with the BS InfoMax separation. 5 Conclusions Experiments with both one- and two-dimensional simulated and natural signals demonstrate that multinode sparse representations improve the efficiency of blind source separation. The proposed method improves the separation quality by utilizing the structure of signals, wherein several subsets of the wavelet packet coefficients have significantly better sparsity and separability than others. In this case, scatter plots of these coefficients show distinct orientations each of which specifies a column of the mixing matrix. We choose the 'good subsets' according to the global distortion adopted as a measure of cluster quality. Finally, we combine together coefficients from the best chosen subsets and restore the mixing matrix using only this new subset of coefficients by the Infomax algorithm or clustering. This yields significantly better results than those obtained by applying standard Infomax and clustering approaches directly to the raw data. The advantage of our method is in particular noticeable in the case of noisy mixtures. References [1] A. 1. Bell and T. 1. Sejnowski, "An information-maximization approach to blind separation and blind deconvolution," Neural Computation, vol. 7, no. 6, pp. 1129- 1159, 1995. [2] A. Hyvarinen, "Survey on independent component analysis," Neural Computing Surveys, no. 2, pp. 94- 128, 1999. [3] M. Zibulevsky and B. A. Pearlmutter, "Blind separation of sources with sparse representations in a given signal dictionary," Neural Computation, vol. l3, no. 4, pp. 863882,2001. [4] 1.-F. Cardoso. "Infomax and maximum likelihood for blind separation," IEEE Signal Processing Letters 4 112-114, 1997. [5] M. S. Lewicki and T. 1. Sejnowski, "Learning overcomplete representations," Neural Computation, 12(2): 337-365, 2000. [6] S. Amari, A. Cichocki, and H. H. Yang, "A new learning algorithm for blind signal separation," In Advances in Neural Information Processing Systems 8. MIT Press. 1996. [7] S. Makeig, ICAlEEG toolbox. Computational Neurobiology Laboratory, the Salk Institute. http://www.cnl.salk.edurtewonlica _ cnl.html, 1999. [8] A. Prieto, C. G. Puntonet, and B. Prieto, "A neural algorithm for blind separation of sources based on geometric prperties.," Signal Processing, vol. 64, no. 3, pp. 315- 331, 1998. [9] S. Mallat, A Wavelet Tour of Signal Processing. Academic Press, 1998. [10] D. L. Donoho, "De-Noising by Soft Thresholding," IEEE Trans. Inf. Theory, vol. 41, 3, 1995, pp.613-627. [11] P. Kisilev, M. Zibulevsky, Y. Y. Zeevi, and B. A. Pearlmutter, Multiresolution frameworkfor sparse blind source separation, CCIT Report no.317, June 2000
|
2001
|
191
|
2,007
|
Estimating the Reliability of leA Projections F. Meinecke l ,2, A. Ziehel , M. Kawanabel and K.-R. Miillerl ,2* 1 Fraunhofer FIRST.IDA, Kekuh~str. 7, 12489 Berlin, Germany 2University of Potsdam, Am Neuen Palais 10, 14469 Potsdam, Germany {meinecke,ziehe,nabe,klaus}©first.fhg.de Abstract When applying unsupervised learning techniques like ICA or temporal decorrelation, a key question is whether the discovered projections are reliable. In other words: can we give error bars or can we assess the quality of our separation? We use resampling methods to tackle these questions and show experimentally that our proposed variance estimations are strongly correlated to the separation error. We demonstrate that this reliability estimation can be used to choose the appropriate ICA-model, to enhance significantly the separation performance, and, most important, to mark the components that have a actual physical meaning. Application to 49-channel-data from an magneto encephalography (MEG) experiment underlines the usefulness of our approach. 1 Introduction Blind source separation (BSS) techniques have found wide-spread use in various application domains, e.g. acoustics, telecommunication or biomedical signal processing. (see e.g. [9, 5, 6, 1, 2, 4, 14, 8]). BSS is a statistical technique to reveal unknown source signals when only mixtures of them can be observed. In the following we will only consider linear mixtures; the goal is then to estimate those projection directions, that recover the source signals. Many different BSS algorithms have been proposed, but to our knowledge, so far, no principled attempts have been made to assess the reliability of BSS algorithms, such that error bars are given along with the resulting projection estimates. This lack of error bars or means for selecting between competing models is of course a basic dilemma for most unsupervised learning algorithms. The sources of potential unreliability of unsupervised algorithms are ubiquous, i.e. noise, non-stationarities, small sample size or inadequate modeling (e.g. sources are simply dependent instead of independent). Unsupervised projection techniques like PCA or BSS will always give an answer that is found within their model class, e.g. PCA will supply an orthogonal basis even if the correct modeling might be non-orthogonal. But how can we assess such a miss-specification or a large statistical error? Our approach to this problem is inspired by the large body of statistics literature on • To whom correspondence should be addressed. resampling methods (see [12] or [7] for references), where algorithms for assessing the stability of the solution have been analyzed e.g. for peA [3]. We propose reliability estimates based on bootstrap resampling. This will enable us to select a good BSS model, in order to improve the separation performance and to find potentially meaningful projection directions. In the following we will give an algorithmic description of the resampling methods, accompanied by some theoretical remarks (section 2) and show excellent experimental results (sections 3 and 4). We conclude with a brief discussion. 2 Resampling Techniques for BSS 2.1 The leA Model In blind source separation we assume that at time instant t each component Xi(t) of the observed n-dimensional data vector, x(t) is a linear superposition of m ::::: n statistically independent signals: m Xi(t) = LAijSj(t) j=l (e.g. [8]). The source signals Sj(t) are unknown, as are the coefficients Aij of the mixing matrix A. The goal is therefore to estimate both unknowns from a sample of the x(t), i.e. y(t) = s(t) = Wx(t), where W is called the separating matrix. Since both A and s(t) are unknown, it is impossible to recover the scaling or the order of the columns of the mixing matrix A. All that one can get are the projection directions. The mixing/ demixing process can be described as a change of coordinates. From this point of view the data vector stays the same, but is expressed in different coordinate systems (passive transformation). Let {ed be the canonical basis of the true sources s = 'E eiSi. Analogous, let {fj} be the basis of the estimated leA channels: y = 'E fjYj. Using this, we can define a component-wise separation error Ei as the angle difference between the true direction of the source and the direction of the respective leA channel: Ei = arccos ("e~i: ~ifill) . To calculate this angle difference, remember that component-wise we have Yj 'E WjkAkisi. With Y = s, this leads to: fj = 'E ei(WA)ij1, i.e. fj is the j-th column of (WA)- l. In the following, we will illustrate our approach for two different source separation algorithms (JADE, TDSEP). JADE [4] using higher order statistics is based on the joint diagonalization of matrices obtained from 'parallel slices' of the fourth order cumulant tensor. TDSEP [14] relies on second order statistics only, enforcing temporal decorrelation between channels. 2.2 About Resampling The objective of resampling techniques is to produce surrogate data sets that eventually allow to approximate the 'separation error' by a repeated estimation of the parameters of interest. The underlying mixing should of course be independent of the generation process of the surrogate data and therefore remain invariant under resampling. Bootstrap Resampling The most popular res amp ling methods are the Jackknife and the Bootstrap (see e.g. [12, 7]) The Jackknife produces surrogate data sets by just deleting one datum each time from the original data. There are generalizations of this approach like k-fold cross-validation which delete more than one datum at a time. A more general approach is the Bootstrap. Consider a block of, say, N data points. For obtaining one bootstrap sample, we draw randomly N elements from the original data, i.e. some data points might occur several times, others don't occur at all in the bootstrap sample. This defines a series {at} with each at telling how often the data point x(t) has been drawn. Then, the separating matrix is computed on the full block and repeatedly on each of the N -element bootstrap samples. The variance is computed as the squared average difference between the estimate on the full block and the respective bootstrap unmixings. (These resampling methods have some desirable properties, which make them very attractive; for example, it can be shown that for iid data the bootstrap estimators of the distributions of many commonly used statistics are consistent.) It is straight forward to apply this procedure to BSS algorithms that do not use time structure; however, only a small modification is needed to take time structure into account. For example, the time lagged correlation matrices needed for TDSEP, can be obtained from {ad by 1 N Cij(T) = N 2: at 'Xi(t)Xj(t+T) t= l with L at = N and at E {O, 1, 2, ... }. Other resampling methods Besides the Bootstrap, there are other res amp ling methods like the Jackknife or cross-validation which can be understood as special cases of Bootstrap. We have tried k-fold cross-validation, which yielded very similar results to the ones reported here. 2.3 The Resampling Algorithm After performing BSS, the estimated ICA-projections are used to generate surrogate data by resampling. On the whitenedl surrogate data, the source separation algorithm is used again to estimate a rotation that separates this surrogate data. In order to compare different rotation matrices, we use the fact that the matrix representation of the rotation group SO(N) can be parameterized by with (Mab)ij r5~r5t r5~r5b , where the matrices Mij are generators of the group and the aij are the rotation parameters (angles) of the rotation matrix R. Using this parameterization we can easily compare different N-dimensional rotations by comparing the rotation parameters aij. Since the sources are already separated, the estimated rotation matrices will be in the vicinity of the identity matrix.2 . IThe whitening transformation is defined as x' = Vx with V = E[xxTtl/2. 21t is important to perform the resampling when the sources are already separated, so that the aij are distributed around zero, because SO(N) is a non-Abelian group; that means that in general R(a)R«(3) of- R«(3)R(a). Var(aij) measures the instability of the separation with respect to a rotation in the (i, j)-plane. Since the reliability of a projection is bounded by the maximum angle variance of all rotations that affect this direction, we define the uncertainty of the i-th ICA-Projection as Ui := maxj Var(aij). Let us summarize the resampling algorithm: 1. Estimate the separating matrix W with some ICA algorithm. Calculate the ICA-Projections y = Wx 2. Produce k surrogate data sets from y and whiten these data sets 3. For each surrogate data set: do BSS, producing a set of rotation matrices 4. Calculate variances of rotation parameters (angles) aij 5. For each ICA component calculate the uncertainty Ui = maxVar(aij). J 2.4 Asymptotic Considerations for Resampling Properties of res amp ling methods are typically studied in the limit when the number of bootstrap samples B -+ 00 and the length of signal T -+ 00 [12]. In our case, as B -+ 00, the bootstrap variance estimator Ut(B) computed from the aiJ's converge to Ut(oo) := maxj Varp[aij] where aij denotes the res amp led deviation and F denotes the distribution generating it. Furthermore, if F -+ F, Ut (00) converges to the true variance Ui = maxj VarF[aij ] as T -+ 00. This is the case, for example, if the original signal is i.i.d. in time. When the data has time structure, F does not necessarily converge to the generating distribution F of the original signal anymore. Although we cannot neglect this difference completely, it is small enough to use our scheme for the purposes considered in this paper, e.g. in TDSEP, where the aij depend on the variation of the time-lagged covariances Cij(T) of the signals, we can show that their estimators Ctj (T) are unbiased: Furthermore, we can bound the difference t:.ijkl(T,V) = COVF [Cij(T),Ckl(V)] COV p [Ctj ( T), Ckl (v)] between the covariance of the real matrices and their bootstrap estimators as if :3a < 1, M ;::: 1, Vi: ICii (T) I :S M aJLICii(O) I. In our experiments, however, the bias is usually found to be much smaller than this upper bound. 3 Experiments 3.1 Comparing the separation error with the uncertainty estimate To show the practical applicability of the resampling idea to ICA, the separation error Ei was compared with the uncertainty Ui. The separation was performed on different artificial 2D mixtures of speech and music signals and different iid data sets of the same variance. To achieve different separation qualities, white gaussian noise of different intensity has been added to the mixtures. Uj = 0.015 U. = 0.177 ' - - -' j o L---~~~~~~~~--~ o 0.2 0.4 0.6 0.8 separation error Ej 0.7,-----------------------_ 0.6 ur ~ 0.5 § 0.4 ~ ~0.3 c . ~ 0.2 0.1 o L-----~----~----~--~ 0.05 0.15 0.25 0.35 0.45 Figure 1: (a) The probability distribution for the separation error for a small uncertainty is close to zero, for higher uncertainty it spreads over a larger range. (b) The expected error increases with the uncertainty. Figure 1 relates the uncertainty to the separation error for JADE (TDSEP results look qualitatively the same) . In Fig.1 (left) we see the separation error distribution which has a strong peak for small values of our uncertainty measure, whereas for large uncertainties it tends to become flat, i.e. - as also seen from Fig.1 (right) the uncertainty reflects very well the true separation error. 3.2 Selecting the appropriate BSS algorithm As our variance estimation gives a high correlation to the (true) separation error, the next logical step is to use it as a model selection criterion for: (a) selecting some hyperparameter of the BSS algorithm, e.g. choosing the lag values for TDSEP or (b) choosing between a set of different algorithms that rely on different assumptions about the data, i.e. higher order statistics (e.g. JADE, INFO MAX, FastICA, ... ) or second order statistics (e.g. TDSEP). It could, in principle, be much better to extract the first component with one and the next with another assumption/ algorithm. To illustrate the usefulness of our reliability measure, we study a five-channel mixture of two channels of pure white gaussian noise, two audio signals and one channel of uniformly distributed noise. The reliability analysis for higher order statistics (JADE) temporal decorrelation (TDSEP) 0.3 0.3 0.25 0.25 TDSEP 3 9.17.10-5 ~- 0.2 ~- 0.2 TDSEP 4 E E 1.29.10-5 :rg 0.15 :rg 0.15 ,---,---g g ,---::J 0.1 ::J 0.1 0.05 0.05 3 3 ICA Channel i ICA Channel i Figure 2: Uncertainty of leA projections of an artificial mixture using JADE and TDSEP. Resampling displays the strengths and weaknesses of the different models JADE gives the advice to rely only on channels 3,4,5 (d. Fig.2 left). In fact, these are the channels that contain the audio signals and the uniformly distributed noise. The same analysis applied to the TDSEP-projections (time lag = 0, ... ,20) shows, that TDSEP can give reliable estimates only for the two audio sources (which is to be expected; d. Fig.2 right). According to our measure, the estimation for the audio sources is more reliable in the TDSEP-case. Calculation of the separation error verifies this: TDSEP separates better by about 3 orders of magnitude (JADE: E3 = 1.5 . 10- 1 , E4 = 1.4 . 10- 1 , TDSEP: E3 = 1.2 . 10- 4 , E4 = 8.7· 10- 5). Finally, in our example, estimating the audio sources with TDSEP and after this applying JADE to the orthogonal subspace, gives the optimal solution since it combines the small separation errors E3, E4 for TDSEP with the ability of JADE to separate the uniformly distributed noise. 3.3 Blockwise uncertainty estimates For a longer time series it is not only important to know which ICA channels are reliable, but also to know whether different parts of a given time series are more (or less) reliable to separate than others. To demonstrate these effects, we mixed two audio sources (8kHz, lOs - 80000 data points), where the mixtures are partly corrupted by white gaussian noise. Reliability analysis is performed on windows of length 1000, shifted in steps of 250; the resulting variance estimates are smoothed. Fig.3 shows again that the uncertainty measure is nicely correlated with the true separation error, furthermore the variance goes systematically up within the noisy part but also in other parts of the time series that do not seem to match the assumptions underlying the algorithm.3 So our reliability estimates can eventually Figure 3: Upper panel: mixtures, partly corrupted by noise. Lower panel: the blockwise variance estimate (solid line) vs the true separation error on this block (dotted line). be used to improve separation performance by removing all but the 'reliable' parts of the time series. For our example this reduces the overall separation error by 2 orders of magnitude from 2.4.10- 2 to 1.7.10-4 . This moving-window resampling can detect instabilities of the projections in two different ways: Besides the resampling variance that can be calculated for each window, one can also calculate the change of the projection directions between two windows. The later has already been used successfully by Makeig et. al. [10]. 4 Assigning Meaning: Application to Biomedical Data We now apply our reliability analysis to biomedical data that has been produced by an MEG experiment with acoustic stimulation. The stimulation was achieved by presenting alternating periods of music and silence, each of 30s length, to the subjects right ear during 30 min. of total recording time (for details see [13]). The measured DC magnetic field values, sampled at a frequency of 0.4 Hz, gave a total number of 720 sample points for each of the 49 channels. While previously 3For example, the peak in the last third of the time series can be traced back to the fact that the original time series are correlated in this region. [13] analysing the data, we found that many of the ICA components are seemingly meaningless and it took some medical knowledge to find potential meaningful projections for a later close inspection. However, our reliability assessment can also be seen as indication for meaningful projections, i.e. meaningful components should have low variance. In the experiment, BSS was performed on the 23 most powerful principal components using (a) higher order statistics (JADE) and (b) temporal decorrelation (TDSEP, time lag 0 .. 50). The results in Fig.4 show that none of higher order statistics (JADE) temporal decorrelation (TDSEP) 0.35 0.35 0.3 0.3 0.25 0.25 ::J ::J ~ 0.2 ~ 0.2 i ~ g 0.1 5 g 0.15 ::J ::J 0.1 0.1 0.05 0.05 ,~ 10 15 20 10 15 20 leA-Channel i leA-Channel i Figure 4: Resampling on the biomedical data from MEG experiment shows: (a) no JADE projection is reliable (has low uncertainty) (b) TDSEP is able to identify three sources with low uncertainty. the JADE-projections (left) have small variance whereas TDSEP (right) identifies three sources with a good reliability. In fact, these three components have physical meaning: while component 23 is an internal very low frequency signal (drift) that is always present in DC-measurements, component 22 turns out to be an artifact of the measurement; interestingly component 6 shows a (noisy) rectangular waveform that clearly displays the 1/308 on/off characteristics of the stimulus (correlation to stimulus 0.7; see Fig.5) . The clear dipole-structure of the spatial field pattern in 0.5 ~ ~O In -0.5 ~ stimulUS 1 234 5 6 7 t[min) Figure 5: Spatial field pattern, frequency content and time course of TDSEP channel 6. Fig.5 underlines the relevance of this projection. The components found by JADE do not show such a clear structure and the strongest correlation of any component to the stimulus is about 0.3, which is of the same order of magnitude as the strongest correlated PCA-component before applying JADE. 5 Discussion We proposed a simple method to estimate the reliability of ICA projections based on res amp ling techniques. After showing that our technique approximates the separation error, several directions are open(ed) for applications. First, we may like to use it for model selection purposes to distinguish between algorithms or to chose appropriate hyperparameter values (possibly even component-wise). Second, variances can be estimated on blocks of data and separation performance can be enhanced by using only low variance blocks where the model matches the data nicely. Finally reliability estimates can be used to find meaningful components. Here our assumption is that the more meaningful a component is, the more stably we should be able to estimate it. In this sense artifacts appear of course also as meaningful, whereas noisy directions are discarded easily, due to their high uncertainty. Future research will focus on applying res amp ling techniques to other unsupervised learning scenarios. We will also consider Bayesian modelings where often a variance estimate comes for free, along with the trained model. Acknowledgments K-R.M thanks Guido Nolte and the members of the Oberwolfach Seminar September 2000 in particular Lutz Dumbgen and Enno Mammen for helpful discussions and suggestions. K -R. M and A. Z. acknowledge partial funding by the EU project (IST-1999-14190 - BLISS). We thank the Biomagnetism Group of the PhysikalischTechnische Bundesanstalt (PTB) for providing the MEG-DC data. References [1] S. Amari, A. Cichocki, and H. H. Yang. A new learning algorithm for blind signal separation. In D.S. Touretzky, M.C. Mozer, and M.E. Hasselmo, editors, Advances in Neural Information Processing Systems (NIPS 95), volume 8, pages 882-893. The MIT Press, 1996. [2] A. J. Bell and T. J. Sejnowski. An information maximisation approach to blind separation and blind deconvolution. Neural Computation, 7:1129- 1159, 1995. [3] R. Beran and M.S. Srivastava. Bootstrap tests and confidence regions for functions of a covariance matrix. Annals of Statistics, 13:95- 115, 1985. [4] J.-F. Cardoso and A. Souloumiac. Blind beamforming for non Gaussian signals. IEEE Proceedings-F, 140(6):362- 370, December 1994. [5] P. Comon. Independent component analysis, a new concept? Signal Processing, 36(3):287-314, 1994. [6] G. Deco and D. Obradovic. An information-theoretic approach to neural computing. Springer, New York, 1996. [7] B. Efron and R.J. Tibshirani. An Introduction to the Bootstrap. Chapman & Hall, first edition, 1993. [8] A. Hyviirinen, J. Karhunen, and E. Oja. Independent Component Analysis. Wiley, 200l. [9] Ch. Jutten and J. Herault. Blind separation of sources, part I: An adaptive algorithm based on neuromimetic architecture. Signal Processing, 24:1- 10, 1991. [10] S. Makeig, S. Enghoff, T.-P. Jung, and T. Sejnowski. Moving-window ICA decomposition of EEG data reveals event-related changes in oscillatory brain activity. In Proc. 2nd Int. Workshop on Independent Component Analysis and Blind Source Separation (ICA '2000), pages 627- 632, Helsinki, Finland, 2000. [11] F. Meinecke, A. Ziehe, M. Kawanabe, and K-R. Muller. Assessing reliability of ica projections - a resampling approach. In ICA '01. T.-W. Lee, Ed., 200l. [12] J. Shao and D. Th. The Jackknife and Bootstrap. Springer, New York, 1995. [13] G. Wubbeler, A. Ziehe, B.-M. Mackert, K-R. Muller, L. Trahms, and G. Curio. Independent component analysis of non-invasively recorded cortical magnetic dc-fields in humans. IEEE Transactions on Biomedical Engineering, 47(5):594-599, 2000. [14] A. Ziehe and K-R. Muller. TDSEP - an efficient algorithm for blind separation using time structure. In L. Niklasson, M. Boden, and T. Ziemke, editors, Proc. Int. Conf. on Artificial Neural Networks (ICANN'9S), pages 675 - 680, Skiivde, Sweden, 1998. Springer Verlag.
|
2001
|
192
|
2,008
|
Learning from Infinite Data in Finite Time Pedro Domingos Geoff H ulten Department of Computer Science and Engineering University of Washington Seattle, WA 98185-2350, U.S.A. {pedrod, ghulten} @cs.washington.edu Abstract We propose the following general method for scaling learning algorithms to arbitrarily large data sets. Consider the model Mii learned by the algorithm using ni examples in step i (ii = (nl , ... ,nm)), and the model Moo that would be learned using infinite examples. Upper-bound the loss L(Mii' Moo ) between them as a function of ii, and then minimize the algorithm's time complexity f(ii) subject to the constraint that L(Moo , Mii ) be at most f with probability at most 8. We apply this method to the EM algorithm for mixtures of Gaussians. Preliminary experiments on a series of large data sets provide evidence of the potential of this approach. 1 An Approach to Large-Scale Learning Large data sets make it possible to reliably learn complex models. On the other hand, they require large computational resources to learn from. While in the past the factor limiting the quality of learnable models was typically the quantity of data available, in many domains today data is super-abundant, and the bottleneck is the time required to process it. Many algorithms for learning on large data sets have been proposed, but in order to achieve scalability they generally compromise the quality of the results to an unspecified degree. We believe this unsatisfactory state of affairs is avoidable, and in this paper we propose a general method for scaling learning algorithms to arbitrarily large databases without compromising the quality of the results. Our method makes it possible to learn in finite time a model that is essentially indistinguishable from the one that would be obtained using infinite data. Consider the simplest possible learning problem: estimating the mean of a random variable x. If we have a very large number of samples, most of them are probably superfluous. If we are willing to accept an error of at most f with probability at most 8, Hoeffding bounds [4] (for example) tell us that, irrespective of the distribution of x, only n = ~(R/f)2 1n (2/8) samples are needed, where R is x's range. We propose to extend this type of reasoning beyond learning single parameters, to learning complex models. The approach we propose consists of three steps: 1. Derive an upper bound on the relative loss between the finite-data and infinite-data models, as a function of the number of samples used in each step of the finite-data algorithm. 2. Derive an upper bound on the time complexity of the learning algorithm, as a function of the number of samples used in each step. 3. Minimize the time bound (via the number of samples used in each step) subject to target limits on the loss. In this paper we exemplify this approach using the EM algorithm for mixtures of Gaussians. In earlier papers we applied it (or an earlier version of it) to decision tree induction [2J and k-means clustering [3J. Despite its wide use, EM has long been criticized for its inefficiency (see discussion following Dempster et al. [1]), and has been considered unsuitable for large data sets [8J. Many approaches to speeding it up have been proposed (see Thiesson et al. [6J for a survey). Our method can be seen as an extension of progressive sampling approaches like Meek et al. [5J: rather than minimize the total number of samples needed by the algorithm, we minimize the number needed by each step, leading to potentially much greater savings; and we obtain guarantees that do not depend on unverifiable extrapolations of learning curves. 2 A Loss Bound for EM In a mixture of Gaussians model, each D-dimensional data point Xj is assumed to have been independently generated by the following process: 1) randomly choose a mixture component k; 2) randomly generate a point from it according to a Gaussian distribution with mean f-Lk and covariance matrix ~k. In this paper we will restrict ourselves to the case where the number K of mixture components and the probability of selection P(f-Lk) and covariance matrix for each component are known. Given a training set S = {Xl, ... , X N }, the learning goal is then to find the maximumlikelihood estimates of the means f-Lk. The EM algorithm [IJ accomplishes this by, starting from some set of initial means, alternating until convergence between estimating the probability p(f-Lk IXj) that each point was generated by each Gaussian (the Estep), and computing the ML estimates of the means ilk = 2::;':1 WjkXj / 2::f=l Wjk (the M step), where Wjk = p(f-Lklxj) from the previous E step. In the basic EM algorithm, all N examples in the training set are used in each iteration. The goal in this paper is to speed up EM by using only ni < N examples in the ith iteration, while guaranteeing that the means produced by the algorithm do not differ significantly from those that would be obtained with arbitrarily large N. Let Mii = (ill , . . . ,ilK) be the vector of mean estimates obtained by the finite-data EM algorithm (i.e., using ni examples in iteration i), and let Moo = (f-L1, ... ,f-LK) be the vector obtained using infinite examples at each iteration. In order to proceed, we need to quantify the difference between Mii and Moo . A natural choice is the sum of the squared errors between corresponding means, which is proportional to the negative log-likelihood of the finite-data means given the infinite-data ones: K K D L(Mii' Moo) = L Ililk - f-Lkl12 = L L lilkd f-Lkdl 2 (1) k=l k=ld=l where ilkd is the dth coordinate of il, and similarly for f-Lkd. After any given iteration of EM, lilkd - f-Lkdl has two components. One, which we call the sampling error, derives from the fact that ilkd is estimated from a finite sample, while J-Lkd is estimated from an infinite one. The other component, which we call the weighting error, derives from the fact that, due to sampling errors in previous iterations, the weights Wjk used to compute the two estimates may differ. Let J-Lkdi be the infinite-data estimate of the dth coordinate of the kth mean produced in iteration i, flkdi be the corresponding finite-data estimate, and flkdi be the estimate that would be obtained if there were no weighting errors in that iteration. Then the sampling error at iteration i is Iflkdi J-Lkdi I, the weighting error is Iflkdi flkdi I, and the total error is Iflkdi J-Lkdi 1 :::; Iflkdi flkdi 1 + Iflkdi J-Lkdi I· Given bounds on the total error of each coordinate of each mean after iteration i-I, we can derive a bound on the weighting error after iteration i as follows. Bounds on J-Lkd ,i-l for each d imply bounds on p(XjlJ-Lki ) for each example Xj, obtained by substituting the maximum and minimum allowed distances between Xjd and J-Lkd ,i-l into the expression of the Gaussian distribution. Let P}ki be the upper bound on P(XjlJ-Lki) , and Pjki be the lower bound. Then the weight of example Xj in mean J-Lki can be bounded from below by wjki = PjkiP(J-Lk)/ ~~= l P}k'iP(J-LU, and from above by W}ki = min{p}kiP(J-Lk)/ ~~=l Pjk'iP(J-LU, I}. Let w;t: = W}ki if Xj ::::: 0 and (+) h' d 1 t (- ) -'f > 0 d (- ) + th . W jki W jki ot erWlse, an e W jki W jki 1 Xj _ an W jki W jki 0 erWlse. Then Iflkdi flkdi 1 I , ~7~1 Wjki Xj I J-Lkdi ",ni uj=l Wjki {I ", ni (+) II ",ni ( - ) I} , uj=l W jki Xj , uj=l W jki Xj max J-Lkdi ",ni _ ,J-Lkdi ",ni + uj=l w jki uj=l w jki (2) < A corollary of Hoeffding's [4] Theorem 2 is that, with probability at least 1 - 8, the sampling error is bounded by Iflkdi J-Lkdi 1 :::; (3) where Rd is the range of the dth coordinate of the data (assumed known 1). This bound is independent of the distribution of the data, which will ensure that our results are valid even if the data was not truly generated by a mixture of Gaussians, as is often the case in practice. On the other hand, the bound is more conservative than distribution-dependent ones, requiring more samples to reach the same guarantees. The initialization step is error-free, assuming the finite- and infinite-data algorithms are initialized with the same means. Therefore the weighting error in the first iteration is zero, and Equation 3 bounds the total error. From this we can bound the weighting error in the second iteration according to Equation 2, and therefore bound the total error by the sum of Equations 2 and 3, and so on for each iteration until the algorithms converge. If the finite- and infinite-data EM converge in the same number of iterations m, the loss due to finite data is L(Mii" Moo ) = ~f= l ~~= llflkdm J-Lkdml 2 (see Equation 1). Assume that the convergence criterion is ~f= l IIJ-Lki J-Lk,i-111 2 :::; f. In general 1 Although a normally distributed variable has infinite range, our experiments show that assuming a sufficiently wide finite range does not significantly affect the results. (with probability specified below), infinite-data EM converges at one of the iterations for which the minimum possible change in mean positions is below ,,/, and is guaranteed to converge at the first iteration for which the maximum possible change is below "(. More precisely, it converges at one of the iterations for which ~~= l ~~= l (max{ IPkd,i- l - Pkdil-IPkd,i- l - ftkd,i - ll-IPkdi - ftkdil, O})2 ::; ,,/, and is guaranteed to converge at the first iteration for which ~~=l ~~=l (IPkd,i-l Pkdil + IPkd,i-l - ftkd,i-ll + IPkdi - ftkdil)2 ::; "/. To obtain a bound for L(Mn, Moo), finite-data EM must be run until the latter condition holds. Let I be the set of iterations at which infinite-data EM could have converged. Then we finally obtain where m is the total number of iterations carried out. This bound holds if all of the Hoeffding bounds (Equation 3) hold. Since each of these bounds fails with probability at most 8, the bound above fails with probability at most 8* = K Dm8 (by the union bound). As a result, the growth with K, D and m of the number of examples required to reach a given loss bound with a given probability is only O(v'lnKDm). The bound we have just derived utilizes run-time information, namely the distance of each example to each mean along each coordinate in each iteration. This allows it to be tighter than a priori bounds. Notice also that it would be trivial to modify the treatment for any other loss criterion that depends only on the terms IPkdm - ftkdm I (e.g., absolute loss) . 3 A Fast EM Algorithm We now apply the previous section's result to reduce the number of examples used by EM at each iteration while keeping the loss bounded. We call the resulting algorithm VFEM. The goal is to learn in minimum time a model whose loss relative to EM applied to infinite data is at most f* with probability at least 1 - 8*. (The reason to use f* instead of f will become apparent below.) Using the notation of the previous section, if ni examples are used at each iteration then the running time of EM is O(KD ~::l ni) , and can be minimized by minimizing ~::l ni. Assume for the moment that the number of iterations m is known. Then, using Equation 1, we can state the goal more precisely as follows. Goal: Minimize ~ ::l ni, subject to the constraint that ~~=l IIPkm - ftkml12 ::; f* with probability at least 1 - 8* . A sufficient condition for ~~=l IIPkm - ftkml12 ::; f* is that Vk IIPkm - ftkmll ::; Jf*/K. We thus proceed by first minimizing ~::l ni subject to IIPkm - ftkmll ::; J f* / K separately for each mean.2 In order to do this, we need to express IIPkm ftkm II as a function of the ni 'so By the triangle inequality, IIPki - ftki II ::; IIPki - ftki II + Ilftki - ftk& By Equation 3, Ilftki - ftki II::; ~R2ln(2/8) ~;~ l w;kd(~;~ l Wjki)2, where R2 = ~~=l RJ and 8 = 8* / K Dm per the discussion following Equation 4. The (~;~ l Wjki)2 / ~;~ l W;ki term is a measure of the diversity of the weights, 2This will generally lead to a suboptimal solution; improving it is a matter for future work. being equal to 1 l/Gini(W~i)' where W~i is the vector of normalized weights wjki = wjkd 2:j,i=l Wjl ki. It attains a minimum of! when all the weights but one are zero, and a maximum of ni when all the weights are equal and non-zero. However, we would like to have a measure whose maximum is independent of ni, so that it remains approximately constant whatever the value of ni chosen (for sufficiently large ni). The measure will then depend only on the underlying distribution of the data. Thus we define f3ki = (2:7~1 Wjki)2 /(ni 2:7~1 W]ki)' obtaining IliLki - ILkill :::; JR2ln(2/8)/(2f3kini). Also, IIP-ki-iLkill = J2:~= llP-kdi - iLkdil2, with lP-kdi-iLkdil bounded by Equation 2. To keep the analysis tractable, we upper-bound this term by a function proportional to IIP-kd,i-1 - ILkd,i-111. This captures the notion than the weighting error in one iteration should increase with the total error in the previous one. Combining this with the bound for IliLki - ILkill, we obtain R2 ln(2/8) 2f3kini (5) where CXki is the proportionality constant. Given this equation and IIP-kO - ILkO II = 0, it can be shown by induction that m IIP-km - ILkmll :::; ~ ~ (6) where (7) The target bound will thus be satisfied by minimizing 2::1 ni subject to 2::1 (rkd,;niJ = J E* / K. 3 Finding the n/s by the method of Lagrange multipliers yields ni = ~ (f ~rkir%j) 2 )=1 (8) This equation will produce a required value of ni for each mean. To guarantee the desired E*, it is sufficient to make ni equal to the maximum of these values. The VFEM algorithm consists of a sequence of runs of EM, with each run using more examples than the last, until the bound L(Mii' Moo) :::; E* is satisfied, with L(Mii' Moo) bounded according to Equation 4. In the first run, VFEM postulates a maximum number of iterations m, and uses it to set 8 = 8* / K Dm. If m is exceeded, for the next run it is set to 50% more than the number needed in the current run. (A new run will be carried out if either the 8* or E* target is not met.) The number of examples used in the first run of EM is the same for all iterations, and is set to 1.1(K/2)(R/E*)2ln(2/8). This is 10% more than the number of examples that would theoretically be required in the best possible case (no weighting errors in the last 3This may lead to a suboptimal solution for the ni's, in the unlikely case that Ilflkm Jtkm II increases with them. iteration, leading to a pure Hoeffding bound, and a uniform distribution of examples among mixture components). The numbers of examples for subsequent runs are set according to Equation 8. For iterations beyond the last one in the previous run, the number of examples is set as for the first run. A run of EM is terminated when L~= l L~= l (Iflkd,i- l - flkdi 1 + Iflkd,i-l - ILkd,i-l l + Iflkdi - ILkdi 1)2 :s: "( (see discussion preceding Equation 4), or two iterations after L~=l IIILk i - ILk,i-1 112 :s: "( 13, whichever comes first. The latter condition avoids overly long unproductive runs. If the user target bound is E, E* is set to min{ E, "( 13}, to facilitate meeting the first criterion above. When the convergence threshold for infinite-data EM was not reached even when using the whole training set, VFEM reports that it was unable to find a bound; otherwise the bound obtained is reported. VFEM ensures that the total number of examples used in one run is always at least twice the number n used in the previous run. This is done by, if L ni < 2n, setting the ni's instead to n~ = 2n(nil L ni). If at any point L ni > mN, where m is the number of iterations carried out and N is the size of the full training set, Vi ni = N is used. Thus, assuming that the number of iterations does not decrease with the number of examples, VFEM's total running time is always less than three times the time taken by the last run of EM. (The worst case occurs when the one-but-last run is carried out on almost the full training set.) The run-time information gathered in one run is used to set the n/s for the next run. We compute each Ctki as Ilflki - Pkill/llflk,i-l - ILk,i-lll. The approximations made in the derivation will be good, and the resulting ni's accurate, if the means' paths in the current run are similar to those in the previous run. This may not be true in the earlier runs, but their running time will be negligible compared to that of later runs, where the assumption of path similarity from one run to the next should hold. 4 Experiments We conducted a series of experiments on large synthetic data sets to compare VFEM with EM. All data sets were generated by mixtures of spherical Gaussians with means ILk in the unit hypercube. Each data set was generated according to three parameters: the dimensionality D, the number of mixture components K , and the standard deviation (Y of each coordinate in each component. The means were generated one at a time by sampling each dimension uniformly from the range (2(Y,1 - 2(Y). This ensured that most of the data points generated were within the unit hypercube. The range of each dimension in VFEM was set to one. Rather than discard points outside the unit hypercube, we left them in to test VFEM's robustness to outliers. Any ILk that was less than (vD 1 K)(Y away from a previously generated mean was rejected and regenerated, since problems with very close means are unlikely to be solvable by either EM or VFEM. Examples were generated by choosing one of the means ILk with uniform probability, and setting the value of each dimension of the example by randomly sampling from a Gaussian distribution with mean ILkd and standard deviation (Y. We compared VFEM to EM on 64 data sets of 10 million examples each, generated by using every possible combination of the following parameters: D E {4, 8,12, 16}; K E {3, 4, 5, 6}; (Y E {.01, .03, .05, .07}. In each run the two algorithms were initialized with the same means, randomly selected with the constraint that no two be less than vD 1 (2K) apart. VFEM was allowed to converge before EM's guaranteed convergence criterion was met (see discussion preceding Equation 4). All experiments were run on a 1 GHz Pentium III machine under Linux, with "( = O.OOOlDK, 8* = 0.05, and E* = min{O.Ol, "(}. Table 1: Experimental results. Values are averages over the number of runs shown. Times are in seconds, and #EA is the total number of example accesses made by the algorithm, in millions. Runs Algorithm #Runs Time #EA Loss D K rr Bound VFEM 40 217 1.21 2.51 10.5 4.2 0.029 EM 40 3457 19.75 2.51 10.5 4.2 0.029 No bound VFEM 24 7820 43.19 1.20 9.1 4.9 0.058 EM 24 4502 27.91 1.20 9.1 4.9 0.058 All VFEM 64 3068 16.95 2.02 10 4.5 0.04 EM 64 3849 22.81 2.02 10 4.5 0.04 The results are shown in Table 1. Losses were computed relative to the true means, with the best match between true means and empirical ones found by greedy search. Results for runs in which VFEM achieved and did not achieve the required E* and 8* bounds are reported separately. VFEM achieved the required bounds and was able to stop early on 62.5% of its runs. When it found a bound, it was on average 16 times faster than EM. When it did not, it was on average 73% slower. The losses of the two algorithms were virtually identical in both situations. VFEM was more likely to converge rapidly for higher D's and lower K's and rr's. When achieved, the average loss bound for VFEM was 0.006554, and for EM it was 0.000081. In other words, the means produced by both algorithms were virtually identical to those that would be obtained with infinite data.4 We also compared VFEM and EM on a large real-world data set, obtained by recording a week of Web page requests from the entire University of Washington campus. The data is described in detail in Wolman et al. [7], and the preprocessing carried out for these experiments is described in Domingos & Hulten [3]. The goal was to cluster patterns of Web access in order to support distributed caching. On a dataset with D = 10 and 20 million examples, with 8* = 0.05, I = 0.001, E* = 1/3, K = 3, and rr = 0.01, VFEM achieved a loss bound of 0.00581 and was two orders of magnitude faster than EM (62 seconds vs. 5928), while learning essentially the same means. VFEM's speedup relative to EM will generally approach infinity as the data set size approaches infinity. The key question is thus: what are the data set sizes at which VFEM becomes worthwhile? The tentative evidence from these experiments is that they will be in the millions. Databases of this size are now common, and their growth continues unabated, auguring well for the use of VFEM. 5 Conclusion Learning algorithms can be sped up by minimizing the number of examples used in each step, under the constraint that the loss between the resulting model and the one that would be obtained with infinite data remain bounded. In this paper we applied this method to the EM algorithm for mixtures of Gaussians, and observed the resulting speedups on a series of large data sets. 4The much higher loss values relative to the true means, however, indicate that infinitedata EM would often find only local optima (unless the greedy search itself only found a suboptimal match). Acknowledgments This research was partly supported by NSF CAREER and IBM Faculty awards to the first author, and by a gift from the Ford Motor Company. References [1] A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, Series B, 39:1- 38, 1977. [2] P. Domingos and G. Hulten. Mining high-speed data streams. In Proceedings of the Sixth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 71- 80, Boston, MA, 2000. ACM Press. [3] P. Domingos and G. Hulten. A general method for scaling up machine learning algorithms and its application to clustering. In Proceedings of the Eighteenth International Conference on Machine Learning, pp. 106-113, Williamstown, MA, 2001. Morgan Kaufmann. [4] W. Hoeffding. Probability inequalities for sums of bounded random variables. Journal of the American Statistical Association, 58:13- 30, 1963. [5] C. Meek, B. Thiesson, and D. Heckerman. The learning-curve method applied to clustering. Technical Report MSR-TR-01-34, Microsoft Research, Redmond, WA,2000. [6] B. Thiesson, C. Meek, and D. Heckerman. Accelerating EM for large databases. Technical Report MSR-TR-99-31, Microsoft Research, Redmond, WA, 2001. [7] A. Wolman, G. Voelker, N. Sharma, N. Cardwell, M. Brown, T. Landray, D. Pinnel, A. Karlin, and H. Levy. Organization-based analysis of Web-object sharing and caching. In Proceedings of the Second USENIX Conference on Internet Technologies and Systems, pp. 25- 36, Boulder, CO, 1999. [8] T. Zhang, R. Ramakrishnan, and M. Livny. BIRCH: An efficient data clustering method for very large databases. In Proceedings of the 1996 A CM SIGMOD International Conference on Management of Data, pp. 103- 114, Montreal, Canada, 1996. ACM Press.
|
2001
|
193
|
2,009
|
The 9 Factor: Relating Distributions on Features to Distributions on Images James M. Coughlan and A. L. Yuille Smith-Kettlewell Eye Research Institute, 2318 Fillmore Street, San Francisco, CA 94115, USA. Tel. (415) 345-2146/2144. Fax. (415) 345-8455. Email: coughlan@ski.org.yuille@ski.org Abstract We describe the g-factor, which relates probability distributions on image features to distributions on the images themselves. The g-factor depends only on our choice of features and lattice quantization and is independent of the training image data. We illustrate the importance of the g-factor by analyzing how the parameters of Markov Random Field (i.e. Gibbs or log-linear) probability models of images are learned from data by maximum likelihood estimation. In particular, we study homogeneous MRF models which learn image distributions in terms of clique potentials corresponding to feature histogram statistics (d. Minimax Entropy Learning (MEL) by Zhu, Wu and Mumford 1997 [11]) . We first use our analysis of the g-factor to determine when the clique potentials decouple for different features. Second, we show that clique potentials can be computed analytically by approximating the g-factor. Third, we demonstrate a connection between this approximation and the Generalized Iterative Scaling algorithm (GIS), due to Darroch and Ratcliff 1972 [2], for calculating potentials. This connection enables us to use GIS to improve our multinomial approximation, using Bethe-Kikuchi[8] approximations to simplify the GIS procedure. We support our analysis by computer simulations. 1 Introduction There has recently been a lot of interest in learning probability models for vision. The most common approach is to learn histograms of filter responses or, equivalently, to learn probability distributions on features (see right panel of figure (1)). See, for example, [6], [5], [4]. (In this paper the features we are considering will be extracted from the image by filters - hence we use the terms "features" and "filters" synonymously. ) An alternative approach, however, is to learn probability distributions on the images themselves. The Minimax Entropy Learning (MEL) theory [11] uses the maximum entropy principle to learn MRF distributions in terms of clique potentials determined by the feature statistics (i.e. histograms of filter responses). (We note that the maximum entropy principle is equivalent to performing maximum likelihood estimation on an MRF whose form is determined by the choice of feature statistics.) When applied to texture modeling it gives a way to unify the filter based approaches (which are often very effective) with the MRF distribution approaches (which are theoretically attractive). ) \ Figure 1: Distributions on images vs. distributions on features. Left and center panels show a natural image and its image gradient magnitude map, respectively. Right panel shows the empirical histogram (i.e. a distribution on a feature) of the image gradient across a dataset of natural images. This feature distribution can be used to create a MRF distribution over images[10]. This paper introduces the g-factor to examine connections between the distribution over images and the distribution over features. As we describe in this paper (see figure (1)), distributions on images and on features can be related by a g-factor (such factors arise in statistical physics, see [3]) . Understanding the g-factor allows us to approximate it in a form that helps explain why the clique potentials learned by MEL take the form that they do as functions of the feature statistics. Moreover, the MEL clique potentials for different features often seem to be decoupled and the g-factor can explain why, and when, this occurs. (I.e. the two clique potentials corresponding to two features A and B are identical whether we learn them jointly or independently). The g-factor is determined only by the form of the features chosen and the spatial lattice and quantization of the image gray-levels. It is completely independent of the training image data. It should be stressed that the choice of image lattice, gray-level quantization and histogram quantization can make a big difference to the g-factor and hence to the probability distributions which are the output of MEL. In Section (2), we briefly review Minimax Entropy Learning. Section (3) introduces the g-factor and determines conditions for when clique potentials are decoupled. In Section (4) we describe a simple approximation which enables us to learn the clique potentials analytically, and in Section (5) we discuss connections between this approximation and the Generalized Iterative Scaling (GIS) algorithm. 2 Minimax Entropy Learning Suppose we have training image data which we assume has been generated by an (unknown) probability distribution PT(X) where x represents an image. Minimax Entropy Learning (MEL) [11] approximates PT(X) by selecting the distribution with maximum entropy constrained by observed feature statistics i(X) = ;fobs. This gives >:. ¢(£) P(xIA) = e Z [>:] ,where A is a parameter chosen such that Lx P(xIA)¢>(X) = 'l/Jobs· Or equivalently, so that <910;{[>:] = ;fobs. We will treat the special case where the statistics i are the histogram of a shiftinvariant filter {fi(X) : i = 1, ... , N} , where N is the total number of pixels in the image. So 'l/Ja = ¢>a(x) = -tv L~l ba,' i(X) where a = 1, ... , Q indicates the (quantized) ~ ~ Q N filter response values. The potentials become A·¢>(X) = -tv La=l Li=l A(a)ba,fi(X) = -tv L~l A(fi(X)). Hence P(xl,X) becomes a MRF distribution with clique potentials given by A(fi (x)). This determines a Markov random field with the clique structure given by the filters {fd. MEL also has a feature selection stage based on Minimum Entropy to determine which features to use in the Maximum Entropy Principle. The features are evaluated by computing the entropy - Lx P(xl,X) log P(xl,X) for each choice of features (with small entropies being preferred). A filter pursuit procedure was described to determine which filters/features should be considered (our approximations work for this also). 3 The g-Factor This section defines the g-factor and starts investigating its properties in subsection (3.1). In particular, when, and why, do clique potentials decouple? More precisely, when do the potentials for filters A and B learned simultaneously differ from the potentials for the two filters when they are learned independently? We address these issues by introducing the g-factor g(;f) and the associated distribution Po (;f): x space -----+ iii space GG g(ijiJ = number of images x with histogram iii (1) Figure 2: The g-factor g(;f) counts the number of images x that have statistics ;f. Note that the g-factor depends only on the choice of filters and is independent of the training image data. Here L is the number of grayscale levels of each pixel, so that LN is the total number of possible images. The g-factor is essentially a combinational factor which counts the number of ways that one can obtain statistics ;f, see figure (2). Equivalently, Po is the default distribution on ;f if the images are generated by white noise (i.e. completely random images). We can use the g-factor to compute the induced distribution P(~I'x) on the statistics determined by MEL: A ~ ~ L ~ ~ g( ~)eX.,j; ~ L ~ X,j; P(1/1 I'\) = 6;;: 2(-)P(xl'\) = ~, Z[,\] = g(1/1)e· . (2) X'j','j' x Z[,\] ,j; Observe that both P(~I'x) and log Z[,X] are sufficient for computing the parameters X. The ,X can be found by solving either of the following two (equivalent) equations: A ~ ~ ~ ~ 8 10 zrXl ~ L:,j; P(1/1I,\) 1/1 = 1/1obs, or ;X = 1/1obs, which shows that knowledge of the g-factor and eX. ,j; are all that is required to do MEL. Observe from equation (2) that we have P(~I'x = 0) = Po(~) . In other words, setting ,X = 0 corresponds to a uniform distribution on the images x. 3.1 Decoupling Filters We now derive an important property of the minimax entropy approach. As mentioned earlier, it often seems that the potentials for filters A and B decouple. In other words, if one applies MEL to two filters A, B simultaneously bv letting ...... ....A ...... B...... ....A -B ...... "'""'A ...... B . :..tA ...... B 1/1 = (1/1 ,1/1 ), '\ = (,\ ,'\ ), and 1/1obs = (1/1 obs' 1/1 obs)' then the solutIOns'\ , '\ to the equations: LP(xl,XA, ,XB)(iA(x),iB(x)) = (~:bs'~!s)' (3) x are the same (approximately) as the solutions to the equations L:x p(xl,XA )iA(x) = ~!s and L:x P(xl,XB)iB(x) = ~!s, see figure (3) for an example. Figure 3: Evidence for decoupling of features. The left and right panels show the clique potentials learned for the features a I ax and a I ay respectively. The solid lines give the potentials when they are learned individually. The dashed lines show the potentials when they are learned simultaneously. Figure courtesy of Prof. Xiuwen Liu, Florida State University. We now show how this decoupling property arises naturally if the g-factor for the two filters factorizes. This factorization, of course, is a property only of the form of the statistics and is completely independent of whether the statistics of the two filters are dependent for the training data. Property I: Suppose we have two sufficient statistics iA(x), iB (x) which are independent on the lattice in the sense that g(~A,~B ) = gA (~A )gB(~B) , then logZ[,XA,,XB] = logZA[,XA] + logZB[,XB] and p(~A,~B ) = pA(~A)pB(~B ). This implies that the parameters XA, XB can be solved from the independent . 81ogZA[XA] _ -A 8 1ogZB[XB ] equatwns 8XA 'ljJobs' 8XB _ -B A A -A -A -A 'ljJobs or L.,j;A P ('ljJ)'ljJ = 'ljJobs' L.,j;B pB(;fB );fB = ;f~s ' Moreover, the resulting distribution PUC) can be obtained by multiplying the distributions (l/ZA)eXA .,j;A(x) and (l/ZB) eXB.,j;B(x) together. The point here is that the potential terms for the two statistics ;fA,;fB decouple if the phase factor g(;fA,;fB) can be factorized. We conjecture that this is effectively the case for many linear filters used in vision processing. For example, it is plausible that the g-factor for features 0/ ox and 0/ oy factorizes - and figure (3) shows that their clique potentials do decouple (approximately). Clearly, if factorization between filters occurs then it gives great simplification to the system. 4 Approximating the g-factor for a Single Histogram We now consider the case where the statistic is a single histogram. Our aim is to understand why features whose histograms are of stereotypical shape give rise to potentials of the form given by figure (3). Our results, of course, can be directly extended to multiple histograms if the filters decouple, see subsection (3.1). We first describe the approximation and then discuss its relevance for filter pursuit. We rescale the X variables by N so that we have: eNX.¢(x) A _ _ eNX.,j; P(X'I-\) = Z[X] , P('ljJ I-\) = g('ljJ) Z[X] , (4) We now consider the approximation that the filter responses {Ii} are independent of each other when the images are uniformly distributed. This is the multinomial approximation. (We attempted a related approximation [1] which was less successful.) It implies that we can express the phase factor as being proportional to a multinomial distribution: (nt:) LN N! N1/Jl N1/JQ n (nt:) _ N! N1/Jl N1/JQ 9 <P = (N'ljJd!. .. (N'ljJQ)!o ... 0Q ' TO <p (N'ljJd!. .. (N'ljJQ)!Ol "'OQ (5) where L.~= 1 'ljJa = 1 (by definition) and the {oa} are the means of the components Na } with respect to the distribution Po (;f). As we will describe later, the {oa} will be determined by the filters {fi}. See Coughlan and Yuille, in preparation, for details of how to compute the {oa}. This approximation enables us to calculate MEL analytically. Theorem With the multinomial approximation the log partition function is: Q log Z[X] = N log L + N log{~= e " a+1og aa } , (6) a=l and the "potentials" P a} can be solved in terms of the observed data {'ljJobs,a} to be: \ - I 'ljJobs,a Aa - og--, Oa a = 1, ... ,Q. (7) Figure 4: Top row: the multinomial approximation. Bottom row: full implementation of MEL (see text). (Left panels) the potentials, (center panels) synthesized images, and (right panels) the difference between the observed histogram (dashed line) and the histogram of the synthesized images (bold line). Filters were d/dx and d/dy. We note that there is an ambiguity Aa r-+ Aa + K where K is an arbitrary number (recall that L~=l 'IjJ(a) = 1). We fix this ambiguity by setting X = 0 if a. = "Jobs. Proof. Direct calculation. Our simulation results show that this simple approximation gives the typical potential forms generated by Markov Chain Monte Carlo (MCMC) algorithms for Minimax Entropy Learning. Compare the multinomial approximation results with those obtained from a full implementation of MEL by the algorithm used in [11], see figure (4). Filter pursuit is required to determine which filters carry most information. MEL [11] prefers filters (statistics) which give rise to low entropy distributions (this is the "Min" part of Minimax). The entropy is given by H(P) = - Lx P(xIX) log P(xIX) = log Z[X] L~=l Aa'IjJa · For the multinomial approximation this can be computed to be N log L - N L~= l 'ljJa log ~. This gives an intuitive interpretation of feature pursuit: we should prefer filters whose statistical response to the image training data is as large as possible from their responses to uniformly distributed images. This is measured by the Kullback-Leibler divergence L~= l 'ljJa log ~. Recall that if the multinomial approximation is used for multiple filters then we should simply add together the entropies of different filters. 5 Connections to Generalized Iterative Scaling In this section we demonstrate a connection between the multinomial approximation and Generalized Iterative Scaling (GIS)[2]. GIS is an iterative procedure for calculating clique potentials that is guaranteed to converge to the maximum likelihood values of the potentials given the desired empirical filter marginals (e.g. filter histograms). We show that estimating the potentials by the multinomial approximation is equivalent to the estimate obtained after performing the first iteration of GIS. We also outline an efficient procedure that allows us to continue additional GIS iterations to improve upon the multinomial approximation. The GIS procedure calculates a sequence of distributions on the entire image (and is guaranteed to converge to the correct maximum likelihood distribution), with an update rule given by p(t+1)(x) ex P(O)(x)Il~=l{ :F; } <pa(x), where 'lfJit ) =< <Pa(X) >P(t)(x) is the expected histogram for the distribution at time t. This implies that the corresponding clique potential update equation is given by: >.it +1) = >.it ) + log 'lfJ~bs - log 'lfJit ). If we initialize GIS so that the initial distribution is the uniform distribution, i.e. p(O) (x) = L -N, then the distribution after one iteration is p(1) (x) ex e2::a <Pa(X) log (1j;~bs /aa) . In other words, the distribution after one iteration is the MEL distribution with clique potential given by the multinomial approximation. (The result can be adapted to the case of multiple filters, as explained in Coughlan and Yuille, in preparation.) We can iterate GIS to improve the estimate of the clique potentials beyond the accuracy of the multinomial approximation. The main difficulty lies in estimating 'lfJit ) for t > 0 (at t = 0 this expectation is just the mean histogram with respect to the uniform distribution, <l:a, which may be calculated efficiently as described in Coughlan and Yuille, in preparation). One way to approximate these expectations is to apply a Bethe-Kikuchi approximation technique [8], used for estimating marginals on Markov Random Fields, to our MEL distribution. Our technique, which was inspired by the Unified Propagation and Scaling Algorithm [7], consists of writing the Bethe free energy [8] for our 2-d image lattice, simplifying it using the shift invariance of the lattice (which enables the algorithm to run swiftly), and using the Convex-Concave Procedure (CCCP) [9] procedure to obtain an iterative update equation to estimate the histogram expectations. The GIS algorithm is then run using these histogram expectations (the results were accurate and did not improve appreciably by using the higher-order Kikuchi free energy approximation). See Coughlan and Yuille, in preparation, for details of this procedure. 6 Discussion This paper describes the g-factor, which depends on the lattice and quantization and is independent of the training image data. Alternatively it can be thought of as being proportional to the distribution of feature responses when the input images are uniformly distributed. We showed that the g-factor can be used to relate probability distributions on features to distributions on images. In particular, we described approximations which, when valid, enable MEL to be computed analytically. In addition, we can determine when the clique potentials for features decouple, and evaluate how informative each feature is. Finally, we establish a connection between the multinomial approximation and GIS, and outline an efficient procedure based on Bethe-Kikuchi approximations that allows us to continue additional GIS iterations to improve upon the multinomial approximation. Acknowledgements We would like to thank Michael Jordan and Yair Weiss for introducing us to Generalized Iterative Scaling and related algorithms. We also thank Anand Rangarajan, Xiuwen Liu, and Song Chun Zhu for helpful conversations. Sabino Ferreira gave useful feedback on the manuscript. This work was supported by the National Institute of Health (NEI) with grant number R01-EY 12691-01. References [1] J.M. Coughlan and A.L. Yuille. "A Phase Space Approach to Minimax Entropy Learning and The Minutemax approximation". In Proceedings NIPS'98. 1998. [2] J. N. Darroch and D. Ratcliff. "Generalized Iterative Scaling for Log-Linear Models". The Annals of Mathematical Statistics. 1972. Vol. 43, No.5, 14701480. [3] C. Domb and M.S. Green (Eds). Phase Transitions and Critical Phenomena. Vol. 2. Academic Press. London. 1972. [4] S. M. Konishi, A.L. Yuille, J.M. Coughlan and Song Chun Zhu. "Fundamental Bounds on Edge Detection: An Information Theoretic Evaluation of Different Edge Cues." In Proceedings Computer Vision and Pattern Recognition CVPR'99. Fort Collins, Colorado. June 1999. [5] A.B. Lee, D.B. Mumford, and J. Huang. "Occlusion Models of Natural Images: A Statistical Study of a Scale-Invariant Dead Leaf Model". International Journal of Computer Vision. Vol. 41, No.'s 1/2. January/February 2001. [6] J. Portilla and E. P. Simoncelli. "Parametric Texture Model based on Joint Statistics of Complex Wavelet Coefficients". International Journal of Computer Vision. October 2000. [7] Y. W. Teh and M. Welling. "The Unified Propagation and Scaling Algorithm." In Proceedings NIPS'01. 2001. [8] J.S. Yedidia, W.T. Freeman, Y. Weiss, "Generalized Belief Propagation." In Proceedings NIPS'OO. 2000. [9] A.L. Yuille. "CCCP Algorithms to Minimize the Bethe and Kikuchi Free Energies," Neural Computation. In press. 2002. [10] S.C. Zhu and D. Mumford. "Prior Learning and Gibbs Reaction-Diffusion." PAMI vo1.19, no.11, pp1236-1250, Nov. 1997. [11] S.C. Zhu, Y. Wu, and D. Mumford. "Minimax Entropy Principle and Its Application to Texture Modeling". Neural Computation. Vol. 9. no. 8. Nov. 1997.
|
2001
|
194
|
2,010
|
Distribution of Mutual Information Marcus Hutter IDSIA, Galleria 2, CH-6928 Manno-Lugano, Switzerland marcus@idsia.ch http://www.idsia.ch/- marcus Abstract The mutual information of two random variables z and J with joint probabilities {7rij} is commonly used in learning Bayesian nets as well as in many other fields. The chances 7rij are usually estimated by the empirical sampling frequency nij In leading to a point estimate J(nij In) for the mutual information. To answer questions like "is J (nij In) consistent with zero?" or "what is the probability that the true mutual information is much larger than the point estimate?" one has to go beyond the point estimate. In the Bayesian framework one can answer these questions by utilizing a (second order) prior distribution p( 7r) comprising prior information about 7r. From the prior p(7r) one can compute the posterior p(7rln), from which the distribution p(Iln) of the mutual information can be calculated. We derive reliable and quickly computable approximations for p(Iln). We concentrate on the mean, variance, skewness, and kurtosis, and non-informative priors. For the mean we also give an exact expression. Numerical issues and the range of validity are discussed. 1 Introduction The mutual information J (also called cross entropy) is a widely used information theoretic measure for the stochastic dependency of random variables [CT91, SooOO]. It is used, for instance, in learning Bayesian nets [Bun96, Hec98] , where stochastically dependent nodes shall be connected. The mutual information defined in (1) can be computed if the joint probabilities {7rij} of the two random variables z and J are known. The standard procedure in the common case of unknown chances 7rij is to use the sample frequency estimates n~; instead, as if they were precisely known probabilities; but this is not always appropriate. Furthermore, the point estimate J (n~; ) gives no clue about the reliability of the value if the sample size n is finite. For instance, for independent z and J, J(7r) =0 but J(n~;) = O(n- 1/ 2 ) due to noise in the data. The criterion for judging dependency is how many standard deviations J(":,;) is away from zero. In [KJ96, Kle99] the probability that the true J(7r) is greater than a given threshold has been used to construct Bayesian nets. In the Bayesian framework one can answer these questions by utilizing a (second order) prior distribution p(7r),which takes account of any impreciseness about 7r. From the prior p(7r) one can compute the posterior p(7rln), from which the distribution p(Iln) of the mutual information can be obtained. The objective of this work is to derive reliable and quickly computable analytical expressions for p(1ln). Section 2 introduces the mutual information distribution, Section 3 discusses some results in advance before delving into the derivation. Since the central limit theorem ensures that p(1ln) converges to a Gaussian distribution a good starting point is to compute the mean and variance of p(1ln). In section 4 we relate the mean and variance to the covariance structure of p(7rln). Most non-informative priors lead to a Dirichlet posterior. An exact expression for the mean (Section 6) and approximate expressions for the variance (Sections 5) are given for the Dirichlet distribution. More accurate estimates of the variance and higher central moments are derived in Section 7, which lead to good approximations of p(1ln) even for small sample sizes. We show that the expressions obtained in [KJ96, Kle99] by heuristic numerical methods are incorrect. Numerical issues and the range of validity are briefly discussed in section 8. 2 Mutual Information Distribution We consider discrete random variables Z E {l, ... ,r} and J E {l, ... ,s} and an i.i.d. random process with samples (i,j) E {l , ... ,r} x {l, ... ,s} drawn with joint probability 7rij. An important measure of the stochastic dependence of z and J is the mutual information T S 7rij 1( 7r) = L L 7rij log ~ = L 7rij log7rij - L 7ri+ log7ri+ - L 7r +j log7r +j' (1) i=1 j = 1 H +J ij i j log denotes the natural logarithm and 7ri+ = Lj7rij and 7r +j = L i7rij are marginal probabilities. Often one does not know the probabilities 7rij exactly, but one has a sample set with nij outcomes of pair (i,j). The frequency irij := n~j may be used as a first estimate of the unknown probabilities. n:= L ijnij is the total sample size. This leads to a point (frequency) estimate 1(ir) = Lij n~j logn:~:j for the mutual information (per sample). Unfortunately the point estimation 1(ir) gives no information about its accuracy. In the Bayesian approach to this problem one assumes a prior (second order) probability density p( 7r) for the unknown probabilities 7rij on the probability simplex. From this one can compute the posterior distribution p( 7rln) cxp( 7r) rr ij7r~;j (the nij are multinomially distributed). This allows to compute the posterior probability density of the mutual information.1 p(Iln) = f 8(1(7r) - I)p(7rln)dTS7r (2) 2The 80 distribution restricts the integral to 7r for which 1(7r) =1. For large sam1 I(7r) denotes the mutual information for the specific chances 7r, whereas I in the context above is just some non-negative real number. I will also denote the mutual information random variable in the expectation E [I] and variance Var[I]. Expectaions are always w.r.t. to the posterior distribution p(7rln). 2Since O~I(7r) ~Imax with sharp upper bound Imax :=min{logr,logs}, the integral may be restricted to J:mam, which shows that the domain of p(Iln) is [O,Imax] . pIe size n ---+ 00, p(7rln) is strongly peaked around 7r = it and p(Iln) gets strongly peaked around the frequency estimate I = I(it). The mean E[I] = fooo Ip(Iln) dI = f I(7r)p(7rln)dTs7r and the variance Var[I] =E[(I - E[I])2] = E[I2]- E[Ij2 are of central interest. 3 Results for I under the Dirichlet P (oste )rior Most3 non-informative priors for p(7r) lead to a Dirichlet posterior distribution ( I) IT nij -1 ·th· t t t· -,,, h ' th b p 7r n ex: ij 7rij WI III erpre a IOn nij - nij + nij , were nij are e num er of samples (i,j), and n~j comprises prior information (1 for the uniform prior, ~ for Jeffreys' prior, 0 for Haldane's prior, -?:s for Perks' prior [GCSR95]). In principle this allows to compute the posterior density p(Iln) of the mutual information. In sections 4 and 5 we expand the mean and variance in terms of n- 1 : E[I] ~ nij I nijn (r - 1)(8 - 1) O( -2) L...J og --- + + n , .. n ni+n+j 2n 'J (3) Var[I] The first term for the mean is just the point estimate I(it). The second term is a small correction if n » r· 8. Kleiter [KJ96, Kle99] determined the correction by Monte Carlo studies as min {T2~1 , 8;;;,1 }. This is wrong unless 8 or rare 2. The expression 2E[I]/n they determined for the variance has a completely different structure than ours. Note that the mean is lower bounded by co~st. +O(n- 2 ), which is strictly positive for large, but finite sample sizes, even if z and J are statistically independent and independence is perfectly represented in the data (I (it) = 0). On the other hand, in this case, the standard deviation u= y'Var(I) '" ~ ",E[I] correctly indicates that the mean is still consistent with zero. Our approximations (3) for the mean and variance are good if T~8 is small. The central limit theorem ensures that p(Iln) converges to a Gaussian distribution with mean E[I] and variance Var[I]. Since I is non-negative it is more appropriate to approximate p(II7r) as a Gamma (= scaled X2 ) or log-normal distribution with mean E[I] and variance Var[I], which is of course also asymptotically correct. A systematic expansion in n -1 of the mean, variance, and higher moments is possible but gets arbitrarily cumbersome. The O(n- 2) terms for the variance and leading order terms for the skewness and kurtosis are given in Section 7. For the mean it is possible to give an exact expression 1 E[I] = - L nij[1jJ(nij + 1) -1jJ(ni+ + 1) -1jJ(n+j + 1) + 1jJ(n + 1)] (4) n .. 'J with 1jJ(n+1)=-,),+L~= lt=logn+O(~) for integer n. See Section 6 for details and more general expressions for 1jJ for non-integer arguments. There may be other prior information available which cannot be comprised in a Dirichlet distribution. In this general case, the mean and variance of I can still be 3But not all priors which one can argue to be non-informative lead to Dirichlet posteriors. Brand [Bragg] (and others), for instance, advocate the entropic prior p( 7r) ex e-H(rr). related to the covariance structure of p(7fln), which will be done in the following Section. 4 Approximation of Expectation and Variance of I In the following let frij := E[7fij]. Since p( 7fln) is strongly peaked around 7f = fr for large n we may expand J(7f) around fr in the integrals for the mean and the variance. With I:::..ij :=7fij -frij and using L:ij7fij = 1 = L:ijfrij we get for the expansion of (1) ( fr .. ) 1:::..2 . 1:::..2 1:::..2 . J(7f) = J(fr) + 2)og ~ I:::..ij + L ----}J--L ~-L ~+O(1:::..3). (5) .. 7fi+7f+j .. 27fij . 27fi+ . 27f+j 2J 2J 2 J Taking the expectation, the linear term E[ I:::..ij ] = a drops out. The quadratic terms E[ I:::..ij I:::..kd = Cov( 7fij ,7fkl) are the covariance of 7f under distribution p( 7fln) and are proportional to n- 1 . It can be shown that E[1:::..3] ,,-,n-2 (see Section 7). [ ] ( A) 1", (bikbjl bik bjl) ( ) (-2) EJ = J7f +-~ -A- -A- -ACOV7fij,7fkl +On . 2 ijkl 7fij 7fi+ 7f +j (6) The Kronecker delta bij is 1 for i = j and a otherwise. The variance of J in leading order in n - 1 is (7) where :t means = up to terms of order n -2. So the leading order variance and the leading and next to leading order mean of the mutual information J(7f) can be expressed in terms of the covariance of 7f under the posterior distribution p(7fln). 5 The Second Order Dirichlet Distribution Noninformative priors for p(7f) are commonly used if no additional prior information is available. Many non-informative choices (uniform, Jeffreys' , Haldane's, Perks', prior) lead to a Dirichlet posterior distribution: 1 II n;j - 1 ( ) N(n) .. 7fij b 7f++ - 1 with normalization 2J N(n) (8) where r is the Gamma function, and nij = n~j + n~j, where n~j are the number of samples (i,j), and n~j comprises prior information (1 for the uniform prior, ~ for Jeffreys' prior, a for Haldane's prior, -!s for Perks' prior). Mean and covariance of p(7fln) are A E[] nij 7fij:= 7fij =-, n (9) Inserting this into (6) and (7) we get after some algebra for the mean and variance of the mutual information I(7r) up to terms of order n- 2 : E[I] J (r - 1)(8 - 1) O( -2) + 2(n + 1) + n , (10) Var[I] ~1 (K - J2) + 0(n-2), n+ (11) (12) (13) J and K (and L, M, P, Q defined later) depend on 7rij = ":,j only, i.e. are 0(1) in n. Strictly speaking we should expand n~l = ~+0(n-2), i.e. drop the +1, but the exact expression (9) for the covariance suggests to keep the +1. We compared both versions with the exact values (from Monte-Carlo simulations) for various parameters 7r. In most cases the expansion in n~l was more accurate, so we suggest to use this variant. 6 Exact Value for E[I] It is possible to get an exact expression for the mean mutual information E[I] under the Dirichlet distribution. By noting that xlogx= d~x,6I,6= l' (x = {7rij,7ri+ ,7r+j}), one can replace the logarithms in the last expression of (1) by powers. From (8) we see that E[ (7rij ),6] = ~i~:~ t~~~;l. Taking the derivative and setting ,8 = 1 we get d 1 E[7rij log 7rij] = d,8E[(7rij) ,6],6=l = ;;: 2:::: nij[1j!(nij + 1) -1j!(n + 1)]. "J The 1j! function has the following properties (see [AS74] for details) dlogf(z) f'(z) 1 1 1 1j!(z) = dz = f(z)' 1j!(z + 1) = log z + 2z 12z2 + O( Z4)' n - l 1 n 1 1j!(n) = -"( + L k' 1j!(n +~) = -"( + 2log2 + 2 L 2k _ l' (14) k=l k=l The value of the Euler constant "( is irrelevant here, since it cancels out. Since the marginal distributions of 7ri+ and 7r+j are also Dirichlet (with parameters ni+ and n+j) we get similarly 1 - L n+j[1j!(n+j + 1) -1j!(n + 1)]. n . J Inserting this into (1) and rearranging terms we get the exact expression4 1 E[I] = - L nij[1j!(nij + 1) -1j!(ni+ + 1) -1j!(n+j + 1) + 1j!(n + 1)] (15) n .. 4This expression has independently been derived in [WW93]. For large sample sizes, 'Ij;(z+ 1) ~ logz and (15) approaches the frequency estimate I(7r) as it should be. Inserting the expansion 'Ij;(z+ 1) = logz + 2\ + ... into (15) we also get the correction term (r - 11~s - 1) of (3). The presented method (with some refinements) may also be used to determine an exact expression for the variance of I(7f). All but one term can be expressed in terms of Gamma functions. The final result after differentiating w.r.t. (31 and (32 can be represented in terms of 'Ij; and its derivative 'Ij;' . The mixed term E[( 7fi+ )131 (7f +j )132] is more complicated and involves confluent hypergeometric functions, which limits its practical use [WW93] . 7 Generalizations A systematic expansion of all moments of p(Iln) to arbitrary order in n -1 is possible, but gets soon quite cumbersome. For the mean we already gave an exact expression (15), so we concentrate here on the variance, skewness and the kurtosis of p(Iln). The 3rd and 4th central moments of 7f under the Dirichlet distribution are ( )2( ) [27ra7rb7rc - 7ra7rbc5bc - 7rb7rcc5ca - 7rc7rac5ab + 7rac5abc5bc] n+l n+2 (16) ~2 [37ra7rb7rc7rd - jrc!!d!!a c5ab - A7rbjrdA7rac5ac - A7rbA7rcA7rac5ad (17) -7fa7fd7fbc5bc - 7fa7fc7fbc5bd - 7fa7fb7fcc5cd +7ra7rcc5abc5cd + 7ra7rbc5acc5bd + 7ra7rbc5adc5bc] + O(n-3) with a=ij, b= kl, ... E {1, ... ,r} x {1, ... ,8} being double indices, c5ab =c5ik c5jl , ... 7rij = n~j • Expanding D..k = (7f_7r)k in E[D..aD..b ... ] leads to expressions containing E[7fa7fb ... ], which can be computed by a case analysis of all combinations of equal/unequal indices a,b,c, ... using (8). Many terms cancel leading to the above expressions. They allow to compute the order n- 2 term of the variance of I(7f). Again, inspection of (16) suggests to expand in [(n+l)(n+2)]-1, rather than in n-2 . The variance in leading and next to leading order is Var[I] K - J2 + M + (r - 1)(8 1)(~ - J) - Q + O(n- 3) n + 1 (n + l)(n + 2) (18) M L (~- _1 _ _ _ 1_ +~) nij log nijn , ij nij ni+ n+j n ni+n+j (19) Q 2 l-L~· ij ni+n+j (20) J and K are defined in (12) and (13). Note that the first term ~+f also contains second order terms when expanded in n -1. The leading order terms for the 3rd and 4th central moments of p(Iln) are L .'""" nij I nij n ~- og--j n ni+n+j 32 [K - J 2F + O(n- 3 ), n from which the skewness and kurtosis can be obtained by dividing by Var[Ij3/2 and Var[IF respectively. One can see that the skewness is of order n- 1/ 2 and the kurtosis is 3 + 0 (n - 1). Significant deviation of the skewness from a or the kurtosis from 3 would indicate a non-Gaussian I. They can be used to get an improved approximation for p(Iln) by making, for instance, an ansatz and fitting the parameters b, c, jJ" and (j-2 to the mean, variance, skewness, and kurtosis expressions above. Po is the Normal or Gamma distribution (or any other distribution with Gaussian limit). From this, quantiles p(I>I*ln):= fI:'p(Iln) dI, needed in [KJ96, Kle99], can be computed. A systematic expansion of arbitrarily high moments to arbitrarily high order in n- 1 leads, in principle, to arbitrarily accurate estimates. 8 Numerics There are short and fast implementations of'if;. The code of the Gamma function in [PFTV92], for instance, can be modified to compute the 'if; function. For integer and half-integer values one may create a lookup table from (14). The needed quantities J, K, L, M, and Q (depending on n) involve a double sum, P only a single sum, and the r+s quantities Ji+ and J+j also only a single sum. Hence, the computation time for the (central) moments is of the same order O(r·s) as for the point estimate (1). "Exact" values have been obtained for representative choices of 7rij, r, s, and n by Monte Carlo simulation. The 7rij := Xij / x++ are Dirichlet distributed, if each Xij follows a Gamma distribution. See [PFTV92] how to sample from a Gamma distribution. The variance has been expanded in T~S, so the relative error Var [I]app"o.-Var[I] .. act of the approximation (11) and (18) are of the order of T'S and Var[Il e• act n (T~S)2 respectively, if z and J are dependent. If they are independent the leading term (11) drops itself down to order n -2 resulting in a reduced relative accuracy O( T~S) of (18). Comparison with the Monte Carlo values confirmed an accurracy in the range (T~S)1...2. The mean (4) is exact. Together with the skewness and kurtosis we have a good description for the distribution of the mutual information p(Iln) for not too small sample bin sizes nij' We want to conclude with some notes on useful accuracy. The hypothetical prior sample sizes n~j = {a, -!S' ~,1} can all be argued to be non-informative [GCSR95]. Since the central moments are expansions in n- 1 , the next to leading order term can be freely adjusted by adjusting n~j E [0 ... 1]. So one may argue that anything beyond leading order is free to will, and the leading order terms may be regarded as accurate as we can specify our prior knowledge. On the other hand, exact expressions have the advantage of being safe against cancellations. For instance, leading order of E[I] and E[I2] does not suffice to compute the leading order of Var[I]. Acknowledgements I want to thank Ivo Kwee for valuable discussions and Marco Zaffalon for encouraging me to investigate this topic. This work was supported by SNF grant 200061847.00 to Jiirgen Schmidhuber. References [AS74] [Bra99] [Bun96] [CT91] M. Abramowitz and 1. A. Stegun, editors. Handbook of mathematical functions. Dover publications, inc., 1974. M. Brand. Structure learning in conditional probability models via an entropic prior and parameter extinction. Neural Computation, 11(5):1155- 1182, 1999. W. Buntine. A guide to the literature on learning probabilistic networks from data. IEEE Transactions on Knowledge and Data Engineering, 8:195- 210, 1996. T. M. Cover and J. A. Thomas. Elements of Information Theory. Wiley Series in Telecommunications. John Wiley & Sons, New York, NY, USA, 1991. [GCSR95] A. Gelman, J. B. Carlin, H. S. Stern, and D. B. Rubin. Bayesian Data Analysis. Chapman, 1995. [Hec98] D. Heckerman. A tutorial on learning with Bayesian networks. Learnig in Graphical Models, pages 301-354, 1998. [KJ96] G. D. Kleiter and R. Jirousek. Learning Bayesian networks under the control of mutual information. Proceedings of the 6th International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems (IPMU-1996), pages 985- 990, 1996. [Kle99] G. D. Kleiter. The posterior probability of Bayes nets with strong dependences. Soft Computing, 3:162- 173, 1999. [PFTV92] W. H. Press, B. P. Flannery, S. A. Teukolsky, and W. T. Vetterling. Numerical Recipes in C: The Art of Scientific Computing. Cambridge University Press, Cambridge, second edition, 1992. [SooOO] E. S. Soofi. Principal information theoretic approaches. Journal of the American Statistical Association, 95:1349- 1353, 2000. [WW93] D. R. Wolf and D. H. Wolpert. Estimating functions of distributions from A finite set of samples, part 2: Bayes estimators for mutual information, chisquared, covariance and other statistics. Technical Report LANL-LA-UR-93833, Los Alamos National Laboratory, 1993. Also Santa Fe Insitute report SFI-TR-93-07 -047.
|
2001
|
195
|
2,011
|
A Maximum-Likelihood Approach to Modeling Multisensory Enhancement Hans Colonius* Institut fUr Kognitionsforschung Carl von Ossietzky Universitat Oldenburg, D-26111 hans. colonius@uni-oldenburg.de Adele Diederich School of Social Sciences International University Bremen Bremen, D-28725 a. diederich@iu-bremen.de Abstract Multisensory response enhancement (MRE) is the augmentation of the response of a neuron to sensory input of one modality by simultaneous input from another modality. The maximum likelihood (ML) model presented here modifies the Bayesian model for MRE (Anastasio et al.) by incorporating a decision strategy to maximize the number of correct decisions. Thus the ML model can also deal with the important tasks of stimulus discrimination and identification in the presence of incongruent visual and auditory cues. It accounts for the inverse effectiveness observed in neurophysiological recording data, and it predicts a functional relation between uni- and bimodal levels of discriminability that is testable both in neurophysiological and behavioral experiments. 1 Introduction In a typical environment stimuli occur at various positions in space and time. In order to produce a coherent assessment of the external world an individual must constantly discriminate between signals relevant for action planning (targets) and signals that need no immediate response (distractors). Separate sensory channels process stimuli by modality, but an individual must determine which stimuli are related to one another, i.e., it is must construct a perceptual event by integrating information from several modalities. For example, stimuli that occur at the same time and space are likely to be interrelated by a common cause. However, if the visual and auditory cues are incongruent, e.g., when dubbing one syllable onto a movie showing a person mouthing a different syllable, listeners typically report hearing a third syllable that represents a combination of what was seen and heard (McGurk effect, cf. [1]). This indicates that cross-modal synthesis is particularly important for stimulus identification and discrimination, not only for detection. Evidence for multisensory integration at the neural level has been well documented in a series of studies in the mammalian midbrain by Stein, Meredith and Wallace (e.g., [2]; for a review, see [3]). The deep layers of the superior colliculus (DSC) • www.uni-oldenburg.de/psychologie /hans.colonius /index.html integrate multisensory input and trigger orienting responses toward salient targets. Individual DSC neurons can receive inputs from multiple sensory modalities (visual, auditory, and somatosensory), there is considerable overlap between the receptive fields of these individual multisensory neurons, and the number of neural impulses evoked depends on the spatial and temporal relationships of the multisensory stimuli. Multisensory response enhancement refers to the augmentation of the response of a DSC neuron to a multisensory stimulus compared to the response elicited by the most effective single modality stimulus. A quantitative measure of the percent enhancement is MRE = CM - SMmax x 100, SMmax (1) where CM is the mean number of impulses evoked by the combined-modality stimulus in a given time interval, and S Mmax refers to the response of the most effective single-modality stimulus (cf. [4]). Response enhancement in the DSC neurons can be quite impressive, with values of M RE sometimes reaching values above 1000. Typically, this enhancement is most dramatic when the unimodal stimuli are weak and/or ambiguous, a principle referred to in [4] as "inverse effectiveness" . Since DSC neurons play an important role in orienting responses (like eye and head movements) to exogenous target stimuli, it is not surprising that multisensory enhancement is also observed at the behavioral level in terms of, for example, a lowering of detection thresholds or a speed-up of (saccadic) reaction time (e.g., [5], [6], [7]; see [8] for a review). Inverse effectiveness makes intuitive sense in the behavioral situation: the detection probability for a weak or ambiguous stimulus gains more from response enhancement by multisensory integration than a highintensity stimulus that is easily detected by a single modality alone. A model of the functional significance of multisensory enhancement has recently been proposed by Anastasio, Patton, and Belkacem-Boussaid [9]. They suggested that the responses of individual DSC neurons are proportional to the Bayesian probability that a target is present given their sensory inputs. Here, this Bayesian model is extended to yield a more complete account of the decision situation an organism is faced with. As noted above, in a natural environment an individual is confronted with the task of discriminating between stimuli important for survival (" targets") and stimuli that are irrelevant (" distractors"). Thus, an organism must not only keep up a high rate of detecting targets but, at the same time, must strive to minimize "false alarms" to irrelevant stimuli. An optimally adapted system will be one that maximizes the number of correct decisions. It will be shown here that this can be achieved already at the level of individual DSC neurons by appealing to a maximum-likelihood principle, without requiring any more information than is assumed in the Bayesian model. The next section sketches the Bayesian model by Anastasio, Patton, and BelkacemBoussaid (Bayesian model, for short), after which a maximum-likelihood model of multisensory response enhancement will be introduced. 2 The Bayesian Model of Multisensory Enhancement DSC neurons receive input from the visual and auditory systems elicited by stimuli occurring within their receptive fields! According to the Bayesian model, these vii An extension to the trimodal situation, including somatosensory input, could be easily attained in the models discussed here. sual and auditory inputs are represented by random variables V and A, respectively. A binary random variable T indicates whether a signal is present (T = 1) or not (T = 0). The central assumption of the model is that a DSC neuron computes the Bayesian (posterior) probability that a target is present in its receptive field given its sensory input: P(T = I V = A = ) = P(V = v, A = a I T = I)P(T = 1) 1 v, a P(V = v, A = a) , (2) where v and a denote specific values of the sensory input variables. Analogous expressions hold for the two unimodal situations. The response of the DSC neuron (number of spikes in a unit time interval) is postulated to be proportional to these probabilities. In order to arrive at quantitative predictions two more specific assumptions are made: (1) the distributions of V and A, given T = 1 or T = 0, are conditionally independent, i.e., P(V = v, A = a I T) = P(V = v IT) P(A = a I T) for any v, a; (2) the distribution of V , given T = 1 or T = 0, is Poisson with Al or Ao, resp., and the distribution of A, given T = 1 or T = 0, is Poisson with {-tl or {-to, resp. The conditional independence assumption means that the visibility of a target indicates nothing about its audibility, and vice-versa. The choice of the Poisson distribution is seen as a reasonable first approximation that requires only one single parameter per distribution. Finally, the computation of the posterior probability that a target is present requires specification of the a-priori probability of a target, P(T = 1). The parameters Ao and {-to denote the mean intensity of the visual and auditory input, resp., when no target is present (spontaneous input), while Al and {-tl are the corresponding mean intensities when a target is present (driven input). By an appropriate choice of parameter values, Anastasio et al. [9] show that the Bayesian model reproduces values of multisensory response enhancement in the order of magnitude observed in neurophysiological experiments [10]. In particular, the property of inverse effectiveness, by which the enhancement is largest for combined stimuli that evoke only small unimodal responses, is reflected by the model. 3 The Maximum Likelihood Model of Multisensory Enhancement 3.1 The decision rule The maximum likelihood model (ML model, for short) incorporates the basic decision problem an organism is faced with in a typical environment: to discriminate between relevant stimuli (targets), i.e., signals that require immediate reaction, and irrelevant stimuli (distractors), i.e., signals that can be ignored in a given situation. In the signal-detection theory framework (cf. [11]), P(Yes I T = 1) denotes the probability that the organism (correctly) decides that a target is present (hit), while P(Yes I T = 0) denotes the probability of deciding that a target is present when in fact only a distractor is present (false alarm). In order to maximize the probability of a correct response, P(C) = P(Yes I T = 1) P(T = 1) + [1- P(Yes I T = O)]P(T = 0), (3) the following maximum likelihood decision rule must be adopted (cf. [12]) for, e.g., the unimodal visual case: If P(T = 11 V = v) > P(T = 0 I V = v), then decide "Yes", otherwise decide "No" . The above inequality is equivalent to P(T=IIV=v) P(T=I)P(v=vIT=I) P(T = 0 I V = v) P(T = 0) P(V = v IT = 0) > 1, where the right-most ratio is a function of V , L(V), the likelihood ratio. Thus, the above rule is equivalent to: If L(v) > 1 - P then decide "Yes" otherwise decide "No" , , , P with p = P(T = 1). Since L(V) is a random variable, the probability to decide "Yes" , given a target is present, is P (Yes I T = 1) = P (L(V) > 1; PIT = 1) . Assuming Poisson distributions, this equals with P (exP(Ao - Ad U~) v > ~ I T = 1) = P(V > ciT = 1), In (l;P) + Al - AO c=---'--------'-----;-----;--In U~) In analogy to the Bayesian model, the ML model postulates that the response of a DSC neuron (number of spikes in a unit time interval) to a given target is proportional to the probability to decide that a target is present computed under the optimal (maximum likelihood) strategy defined above. 3.2 Predictions for Hit Probabilities In order to compare the predictions of the ML model for unimodal vs. bimodal inputs, consider the likelihood ratio for bimodal Poisson input under conditional independence: L(V, A) P(V = v, A = a I T = 1) P(V = v, A = a I T = 0) exp(Ao _ Ad (~~) v exp(po _ pd (~~) A The probability to decide "Yes" given bimodal input amounts to, after taking logarithms, P (In (~~) V + In (~~) A > In (1; p) + Al - AO + PI -Po IT = 1) Table 1: Hit probabilities and MRE for different bimodal inputs Mean Driven Input Prob (Hit) Al J.Ll V Driven A Driven V A Driven MRE Low 6 7 .000 .027 .046 704 7 7 .027 .027 .117 335 8 8 .112 .112 .341 204 8 9 .112 .294 .528 79 8 10 .112 .430 .562 31 Medium 12 12 .652 .652 .872 33 12 13 .652 .748 .895 20 High 16 16 .873 .873 .984 13 16 20 .873 .961 .990 3 Note: A-priori target probability is set at p = O.l. Visual and auditory inputs have spontaneous means of 5 impulses per unit time. V Driven (A Driven, V A Driven) columns refer to the hit probabilities given a unimodal visual (resp. auditory, bimodal) target. Multisensory response enhancement (last column) is computed using Eq. (1) For Ad Ao = J.Ld J.Lo this probability is computed directly from the Poisson distribution with mean (AI + J.Ld. Otherwise, hit probabilities follow the distribution of a linear combination of two Poisson distributed variables. Table 1 presents2 hit probabilities and multisensory response enhancement values for different levels of mean driven input. Obviously, the ML model imitates the inverse effectiveness relation: combining weak intensity unimodal stimuli leads to a much larger response enhancement than medium or high intensity stimuli. 3.3 Predictions for discriminability measures The ML model allows to assess the sensitivity of an individual DSC neuron to discriminate between target and distract or signals. Intuitively, this sensitivity should be a (decreasing) function of the amount of overlap between the driven and the spontaneous likelihood (e.g., P(V = v IT = 1) and P(V = v I T = 0)). One possible appropriate measure of sensitivity for the Poisson observer is (cf. [12]) Al - Ao J.Ll J.Lo Dy = (AI AO)I/4 and DA = (J.LIJ.LO)l /4 (4) for the visual and auditory unimodal inputs, resp. A natural choice for the bimodal measure of sensitivity then is D (AI + J.Ll) (J.Lo + Ao) y A = [(AI + J.Ld(Ao + J.Lo)Jl/4 . (5) Note that, unlike the hit probabilities, the relative increase in discriminability by combining two unimodal inputs does not decrease with the intensity of the driven input (see Table 2). Rather, the relation between bimodal and unimodal discriminability measures for the input values in Table 2 is approximately of Euclidean 2For input combinations with >'1 =I- J.t1 hit probabilities are estimated from samples of 1,000 pseudo-random numbers. Table 2: Discriminability measure values and % increase for different bimodal inputs Mean Driven Input Discriminability Value Al J.Ll Dv DA DVA % Increase 7 7 .82 .82 1.16 41 8 8 1.19 1.19 1.69 41 8 10 1.19 1.88 2.18 16 12 12 2.52 2.52 3.57 41 16 16 3.68 3.68 5.20 41 16 20 3.68 4.74 5.97 26 Note: Visual and auditory inputs have spontaneous means of 5 impulses per unit time. % Increase of Dv A over Dv and DA (last column) is computed in analogy to Eq. (1) distance form: (6) For Al = J.Ll this amounts to Dv A = V2Dv yielding the 41 % increase in discriminability. The fact that the discriminability measures do not follow the inverse effectiveness rule should not be not surprising: whether two stimuli are easy or hard to discriminate depends on their signal-to-noise ratio, but not on the level of intensity. 4 Discussion and Conclusion The maximum likelihood model of multisensory enhancement developed here assumes that the response of a DSC neuron to a target stimulus is proportional to the hit probability under a maximum likelihood decision strategy. Obviously, no claim is made here that the neuron actually performs these computations, only that its behavior can be described approximately in this way. Similar to the Bayesian model suggested by Anastasio et al. [9], the neuron's behavior is solely based on the a-priori probability of a target and the likelihood function for the different sensory inputs. The ML model predicts the inverse effectiveness observed in neurophysiological experiments. Moreover, the model allows to derive a measure of the neuron's ability to discriminate between targets and non-targets. It makes specific predictions how un i- and bimodal discriminability measures are related and, thereby, opens up further avenues for testing the model assumptions. The ML model, like the Bayesian model, operates at the level of a single DSC neuron. However, an extension of the model to describe multisensory population responses is desirable: First, this would allow to relate the model predictions to numerous behavioral studies about multisensory effects (e.g., [13], [14]), and, second, as a recent study by Kadunce et al. [15) suggests, the effects of multisensory spatial coincidence observed in behavioral experiments may only be reconcilable with the degree of spatial resolution achievable by a population of DSC neurons with overlapping receptive fields. Moreover, this extension might also be useful to relate behavioral and single-unit recording results to recent findings on multisensory brain areas using functional imaging techniques (e.g., King and Calvert [16]). Acknowledgments This research was partially supported by a grant from Deutsche Forschungsgemeinschaft-SFB 517 Neurokognition to the first author. References [1] McGurk, H. & MacDonald, J. (1976). Hearing lips and seeing voices. Nature, 264, 746-748. [2] Wallace, M. T ., Meredith, M. A., & Stein, B. E. (1993) . Converging influences from visual, auditory, and somatosensory cortices onto output neurons of the superior colliculus. Journal of Neurophysiology, 69, 1797-1809. [3] Stein, B. E., & Meredith, M. A. (1996). The merging of the senses. Cambridge, MA: MIT Press. [4] Meredith, M. A. & Stein, B. E. (1986a). Spatial factors determine the activity of multisensory neurons in cat superior colliculus. Brain Research, 365(2), 350-354. [5] Frens, van Opstal, & van der Willigen (1995) . Spatial and temporal factors determine auditory-visual interactions in human saccadic eye movements. Perception fj Psychophysics, 57, 802-816. [6] Colonius, H. & Arndt, P. A. (2001). A two stage-model for visual-auditory interaction in saccadic latencies. Perception fj Psychophysics, 63, 126-147. [7] Stein, B. E., Meredith, M. A., Huneycutt, W. S., & McDade, L. (1989). Behavioral indices of multisensory integration: Orientation to visual cues is affected by auditory stimuli. Journal of Cognitive Neurosciences, 1, 12-24. [8] Welch, R. B., & Warren, D. H. (1986). Intersensory interactions. In K R. Boff, L. Kaufman, & J. P. Thomas (eds.), Handbook of perception and human performance, Volume I: Sensory process and perception (pp. 25-1-25-36) New York: Wiley [9] Anastasio" T. J., Patton, P. E., & Belkacem-Boussaid, K (2000). Using Bayes' rule to model multisensory enhancement in the superior colliculus. Neural Computation, 12, 1165-1187. [10] Meredith, M. A. & Stein, B. E. (1986b). Visual, auditory, and somatosensory convergence on cells in superior colliculus results in multisensory integration. Journal of Neurophysiology, 56(3), 640-662. [11] Green, D. M., & Swets, J. A. (1974). Signal detection theory and psychophysics. New York: Krieger Pub!. Co. [12] Egan, J. P. (1975) . Signal detection theory and ROC analysis. New York: Academic Press. [13] Craig, A., & Colquhoun, W. P. (1976). Combining evidence presented simultaneously to the eye and the ear: A comparison of some predictive models. Perception fj Psychophysics, 19, 473-484. [14] Stein, B. E., London, N., Wilkinson, L. K , & Price, D. D. (1996). Enhancement of perceived visual intensity by auditory stimuli: A psychophysical analysis. Journal of Cognitive Neuroscience, 8, 497-506. [15] Kadunce, D. C., Vaughan, J. W ., Wallace, M. T ., & Stein, B. E. (2001) . The influence of visual and auditory receptive field organization on multisensory integration in the superior colliculus. Experimental Brain Research, 139, 303-310. [16] King, A. J., & Calvert, G. A. (2001). Multisensory integration: Perceptual grouping by eye and ear. Current Biology, 11, 322-325.
|
2001
|
196
|
2,012
|
Model-Free Least Squares Policy Iteration Michail G. Lagoudakis Department of Computer Science Duke University Durham, NC 27708 mgl@cs.duke.edu Ronald Parr Department of Computer Science Duke University Durham, NC 27708 parr@cs.duke.edu Abstract We propose a new approach to reinforcement learning which combines least squares function approximation with policy iteration. Our method is model-free and completely off policy. We are motivated by the least squares temporal difference learning algorithm (LSTD), which is known for its efficient use of sample experiences compared to pure temporal difference algorithms. LSTD is ideal for prediction problems, however it heretofore has not had a straightforward application to control problems. Moreover, approximations learned by LSTD are strongly influenced by the visitation distribution over states. Our new algorithm, Least Squares Policy Iteration (LSPI) addresses these issues. The result is an off-policy method which can use (or reuse) data collected from any source. We have tested LSPI on several problems, including a bicycle simulator in which it learns to guide the bicycle to a goal efficiently by merely observing a relatively small number of completely random trials. 1 Introduction Linear least squares function approximators offer many advantages in the context of reinforcement learning. While their ability to generalize is less powerful than black box methods such as neural networks, they have their virtues: They are easy to implement and use, and their behavior is fairly transparent, both from an analysis standpoint and from a debugging and feature engineering standpoint. When linear methods fail, it is usually relatively easy to get some insight into why the failure has occurred. Our enthusiasm for this approach is inspired by the least squares temporal difference learning algorithm (LSTD) [4]. LSTD makes efficient use of data and converges faster than other conventional temporal difference learning methods. Although it is initially appealing to attempt to use LSTD in the evaluation step of a policy iteration algorithm, this combination can be problematic. Koller and Parr [5] present an example where the combination of LSTD style function approximation and policy iteration oscillates between two bad policies in an MDP with just 4 states. This behavior is explained by the fact that linear approximation methods such as LSTD compute an approximation that is weighted by the state visitation frequencies of the policy under evaluation. Further, even if this problem is overcome, a more serious difficulty is that the state value function that LSTD learns is of no use for policy improvement when a model of the process is not available. This paper introduces the Least Squares Policy Iteration (LSPI) algorithm, which extends the benefits of LSTD to control problems. First, we introduce LSQ, an algorithm that learns least squares approximations of the state-action ( ) value function, thus permitting action selection and policy improvement without a model. Next we introduce LSPI which uses the results of LSQ to form an approximate policy iteration algorithm. This algorithm combines the policy search efficiency of policy iteration with the data efficiency of LSTD. It is completely off policy and can, in principle, use data collected from any reasonable sampling distribution. We have evaluated this method on several problems, including a simulated bicycle control problem in which LSPI learns to guide the bicycle to the goal by observing a relatively small number of completely random trials. 2 Markov Decision Processes We assume that the underlying control problem is a Markov Decision Process (MDP). An MDP is defined as a 4-tuple
where: is a finite set of states; is a finite set of actions; is a Markovian transition model where
represents the probability of going from state to state with action ; and is a reward function IR, such that !
represents the reward obtained when taking action in state and ending up in state " . We will be assuming that the MDP has an infinite horizon and that future rewards are discounted exponentially with a discount factor #%$'& ()+*"
. (If we assume that all policies are proper, our results generalize to the undiscounted case.) A stationary policy , for an MDP is a mapping ,-./10 , where ,2!"
is the action the agent takes at state . The state-action value function 43 5
is defined over all possible combinations of states and actions and indicates the expected, discounted total reward when taking action in state and following policy , thereafter. The exact -values for all state-action pairs can be found by solving the linear system of the Bellman equations : 3 !6 7
28'9:!6 7
<;=#>@?BA !
3 ! ,2
C where 9:7
D8FE ? A GH
IJ!6 KL
. In matrix format, the system becomes M3 8 9N;O#QP 3R 3 , where S3 and 9 are vectors of size T DTUT 0MT and P 3 is a stochastic matrix of size T DTLT 0 T7OT DTLT 0MT
. P 3 describes the transitions from pairs 7
to pairs !@,2!H
@
. For every MDP, there exists an optimal policy, ,V , which maximizes the expected, discounted return of every state. Policy iteration is a method of discovering this policy by iterating through a sequence of monotonically improving policies. Each iteration consists of two phases. Value determination computes the state-action values for a policy ,.WYX[Z by solving the above system. Policy improvement defines the next policy as ,.WYX[\^]@ZC"
J8`_a bc_d)e 3fhgLi 7
. These steps are repeated until convergence to an optimal policy, often in a surprisingly small number of steps. 3 Least Squares Approximation of Q Functions Policy iteration relies upon the solution of a system of linear equations to find the Q values for the current policy. This is impractical for large state and action spaces. In such cases we may wish to approximate 43 with a parametric function approximator and do some form of approximate policy iteration. We now address the problem of finding a set of parameters that maximizes the accuracy of our approximator. A common class of approximators is the so called linear architectures, where the value function is approximated as a linear weighted combination of basis functions (features): 3 !
8 > ] !6 7
8 !7
S where is a set of weights (parameters). In general,
-T DTLT 0 T and so, the linear system above now becomes an overconstrained system over the parameters : 9F;O#QP 3 # P 3
9 where is a T DTLT 0MT! )
matrix. We are interested in a set of weights 3 that yields a fixed point in value function space, that is a value function 3 8 3 that is invariant under one step of value determination followed by orthogonal projection to the space spanned by the basis functions. Assuming that the columns of are linearly independent this is
] 9N;O#QP 3 3
8 3 8 3 8 # P 3
] 9 We note that this is the standard fixed point approximationmethod for linear value functions with the exception that the problem is formulated in terms of Q values instead of state values. For any P 3 , the solution is guaranteed to exist for all but finitely many # [5]. 4 LSQ: Learning the State-Action Value Function In the previous section we assumed that a model M P 3
of the underlying MDP is available. In many practical applications, such a model is not available and the value function or, more precisely, its parameters have to be learned from sampled data. These sampled data are tuples of the form: L
, meaning that in state , action was taken, a reward was received, and the resulting state was . These data can be collected from actual (sequential) episodes or from random queries to a generative model of the MDP. In the extreme case, they can be experiences of other agents on the same MDP. We know that the desired set of weights can be found as the solution of the system, 3 8! , where 8 # P 3
and 8 9 . The matrix P 3 and the vector 9 are unknown and so, and cannot be determined a priori. However, and can be approximated using samples. Recall that , P 3 , and 9 are of the form: "$# %& & & ' (*),+.-0/213-4 5 60606 (*),+./7184 5 60606 (*),+:9 ;<9=/ 139 >?9=4 5 @A A A B CED "$# %& & & ' FHG A:I ),+.-0/ 1J-0/+0KL42(?),+0KM/ N),+K4 4 5 6O606 FPG A I ),+./21J/+ K 42(?),+ K / NQ),+ K 4 4 5 6O606 FPG A I ),+ 9 ;<9 /21 9 >*9 /+ K 42(?),+ K / NQ),+ K 4 4 5 @A A A B RS# % & & & ' F G ATI ),+ /21 /+ K 42UV),+ /21 /+ K 4 606O6 FHG A I ),+:/71J/+ K 42UV),+./ 1J/ + K 4 606O6 F G A I ),+T9 ;J9=/ 139 >?9W/+0KL42UV),+:9 ;<9X/ 139 >?9Y/+0K4 @ A A A B Given a set of samples, Z 8\["^]0_GJ]0_ K ]0_ :]0_@
+T<`M8 *6OaRbcb dfe , where the :]0_G J]0_I
are sampled from according to distribution g and the 6 ] _ are sampled according to K ]0_ T :]0_GJ]0_I
, we can construct approximate versions of , P 3 , and 9 as follows : 8 hi i i j ]lk ]lk
bc ] _C ] _I
bc !:]m <]m
nbo o o p q P 3 8 hi i i i j !K ]lk @,sr K ]lk
t cb K ] _ @, r K ] _
t cb !H ]m @, r K ]m
t nbo o o o p 9 8 hi i i j ]lk bb ] _ bb :]m nbo o o p These approximations can be thought of as first sampling rows from according to g and then, conditioned on these samples, as sampling terms from the summations in the corresponding rows of P 3 and 9 . The sampling distribution from the summations is governed by the underlying dynamics ( U
) of the process as the samples in Z are taken directly from the MDP. Given , q P 3 , and 9 , and can be approximated as 8 # q P 3
and 8 9 With d uniformly distributed samples over pairs of states and actions !7
, the approximations and are consistent approximations of the true and :
8 d T TUT T and
28 d T TUT T The Markov property ensures that the solution 3 will converge to the true solution 3 for sufficiently large d whenever 3 exists: 3
8 ] C
8 d T DTUT T
] d T DTUT T
28 ] 8$ 3 In the more general case, where g is not uniform, we will compute a weighted projection, which minimizes the g weighted distance in the projection step. Thus, state is implicitly assigned weight g!K
and the projection minimizes the weighted sum of squared errors with respect to g . In LSTD, for example, g is the stationary distribution of P 3 , giving high weight to frequently visited states, and low weight to infrequently visited states. As with LSTD, it is easy to see that approximations ( ] ] 6 ) derived from different sets of samples ( Z ] Z ) can be combined additively to yield a better approximation that corresponds to the combined set of samples: 8 ] ; and 8 ] ; This observation leads to an incremental update rule for and . Assume that initially 8 ( and 8 ( . For a fixed policy, a new sample !6
contributes to the approximation according to the following update equation : ; 5
+ 5
# ! ,2
@
and .; 7
We call this new algorithm LSQ due to its similarity to LSTD. However, unlike LSTD, it computes Q functions and does not expect the data to come from any particular Markov chain. It is a feature of this algorithm that it can use the same set of samples to compute Q values for any policy representation that offers an action choice for each in the set. The policy merely determines which !,2K!@,2!H
@
is added to for each sample. Thus, LSQ can use every single sample available to it no matter what policy is under evaluation. We note that if a particular set of projection weights are desired, it is straightforward to reweight the samples as they are added to . Notice that apart from storing the samples, LSQ requires only J7
space independently of the size of the state and the action space. For each sample in Z , LSQ incurs a cost of J
to update the matrices and and a one time cost of J7 "
to solve the system and find the weights. Singular value decomposition (SVD) can be used for robust inversion of as it is not always a full rank matrix. LSQ includes LSTD as a special case where there is only one action available. It is also possible to extend LSQ to LSQ( ) in a way that closely resembles LSTD( ) [3], but in that case the sample set must consist of complete episodes generated using the policy under evaluation, which again raises the question of bias due to sampling distribution, and prevents the reusability of samples. LSQ is also applicable in the case of infinite and continuous state and/or action spaces with no modification. States and actions are reflected only through the basis functions of the linear approximation and the resulting value function can cover the entire state-action space with the appropriate set of continuous basis functions. 5 LSPI: Least Squares Policy Iteration The LSQ algorithm provides a means of learning an approximate state-action value function, S3 !6 7
, for any fixed policy , . We now integrate LSQ into an approximate policy iteration algorithm. Clearly, LSQ is a candidate for the value determination step. The key insight is that we can achieve the policy improvement step without ever explicitly representing our policy and without any sort of model. Recall that in policy improvement, ,WYX[\^]@Z will pick the action that maximizes 43 !7
. Since LSQ computes Q functions directly, we do not need a model to determine our improved policy; all the information we need is contained implicitly in the weights parameterizing our Q functions1: , WUX[\^]IZ !
8 _a bc_d e 7
8 _a b2c_d e !7
We close the loop simply by requiring that LSQ performs this maximization for each when constructing the matrix for a policy. For very large or continuous action spaces, explicit maximization over may be impractical. In such cases, some sort of global nonlinear optimization may be required to determine the optimal action. Since LSPI uses LSQ to compute approximate Q functions, it can use any data source for samples. A single set of samples may be used for the entire optimization, or additional samples may be acquired, either through trajectories or some other scheme, for each iteration of policy iteration. We summarize the LSPI algorithm in Figure 1. As with any approximate policy iteration algorithm, the convergence of LSPI is not guaranteed. Approximate policy iteration variants are typically analyzed in terms of a value function approximation error and an action selection error [2]. LSPI does not require an approximate policy representation, e.g., a policy function or “actor” architecture, removing one source of error. Moreover, the direct computation of linear Q functions from any data source, including stored data, allows the use of all available data to evaluate every policy, making the problem of minimizing value function approximation error more manageable. 6 Results We initially tested LSPI on variants of the problematic MDP from Koller and Parr [5], essentially simple chains of varying length. LSPI easily found the optimal policy within a few iterations using actual trajectories. We also tested LSPI on the inverted pendulum problem, where the task is to balance a pendulum in the upright position by moving the cart to which it is attached. Using a simple set of basis functions and samples collected from random episodes (starting in the upright position and following a purely random policy), LSPI was able to find excellent policies using a few hundred such episodes [7]. Finally, we tried a bicycle balancing problem [12] in which the goal is to learn to balance and ride a bicycle to a target position located 1 km away from the starting location. Initially, the bicycle’s orientation is at an angle of 90 to the goal. The state description is a sixdimensional vector D
, where is the angle of the handlebar, is the vertical 1This is the same principle that allows action selection without a model in Q-learning. To our knowledge, this is the first application of this principle in an approximate policy iteration algorithm. LSPI ( /( / //*N./ ) // : Number of basis functions // ( : Basis functions // : Discount factor // : Stopping criterion // N : Initial policy, given as , N # N),+./ 4 (default: # ) // : Initial set of samples, possibly empty # N K # N // In essence, K # repeat Update (optional) // Add/remove samples, or leave unchanged N # N K // # K N K = LSQ ( / /( / /*N ) // K = LSQ ( / /( /
/ ) until ( NPN K ) // that is, ( K ) return N // return Figure 1: The LSPI algorithm. angle of the bicycle, and is the angle of the bicycle to the goal. The actions are the torque applied to the handlebar (discretized to [ a7 (RC; a<e ) and the displacement of the rider (discretized to [ (3 ( aR()G;(3 ( aJe ). In our experiments, actions are restricted to be either or (or nothing) giving a total of 5 actions2. The noise in the system is a uniformly distributed term in & ( (8a7G;( (8a added to the displacement component of the action. The dynamics of the bicycle are based on the model described by Randløv and Alstrøm [12] and the time step of the simulation is set to (3 (R* seconds. The state-action value function !7
for a fixed action is approximated by a linear combination of 20 basis functions: <* D D
C where 8 , for ( and 8 , for ( . Note that the state variable is completely ignored. This block of basis functions is repeated for each of the 5 actions, giving a total of 100 basis functions and weights. Training data were collected by initializing the bicycle to a random state around the equilibrium position and running small episodes of 20 steps each using a purely random policy. LSPI was applied on training sets of different sizes and the average performance is shown in Figure 2(a). We used the same data set for each run of policy iteration and usually obtained convergence in 6 or 7 iterations. Successful policies usually reached the goal in approximately 1 km total, near optimal performance. We also show an annotated set of trajectories to demonstrate the performance improvement over multiple steps of policy iteration in Figure 2(b). The following design decisions influenced the performance of LSPI on this problem: As is typical with this problem, we used a shaping reward [10] for the distance to the goal. In this case, we used (3 (R* of the net change (in meters) in the distance to the goal. We found that when using full random trajectories, most of our sample points were not very useful; they occurred after the bicycle had already entered into a “death spiral” from which recovery was impossible. This complicated our learning efforts by biasing the samples towards hopeless parts of the space, so we decided to cut off trajectories after 20 steps. This created an additional problem because there was no terminating reward signal to indicate failure. We approximated this with an additional shaping reward, which was proportional to the 2Results are similar for the full 9-action case, but required more training data. 0 500 1000 1500 2000 2500 3000 0 10 20 30 40 50 60 70 80 90 100 Number of training episodes Percentage of trials reaching the goal −200 0 200 400 600 800 1000 1200 −800 −600 −400 −200 0 200 1st iteration 2nd iteration (crash) 4th and 8th iteration 5th and 7th iteration 3rd iteration 6th iteration (crash) Starting Position Goal (a) (b) Figure 2: The bicycle problem: (a) Percentage of final policies that reach the goal, averaged over 200 runs of LSPI for each training set size; (b) A sample run of LSPI based on 2500 training trials. This run converged in 8 iterations. Note that iterations 5 and 7 had different Q-values but very similar policies. This was true of iterations 4 and 8 as well. The weights of the ninth differed from the eighth by less than *H( ] in , indicate convergence. The curves at the end of the trajectories indicating where the bicycle has looped back for a second pass through the goal. net change in the square of the vertical angle. This roughly approximated the likeliness of falling at the end of a truncated trajectory. Finally, we used a discount of ( ( , which seemed to yield more robust performance. We admit to some slight unease about the amount of shaping and adjusting of parameters that was required to obtain good results on this problem. To verify that we had not eliminated the learning problem entirely through shaping, we reran some experiments using a discount of ( . In this case LSQ simply projects the immediate reward function into the column space of the basis functions. If the problem were tweaked too much, acting to maximize the projected immediate reward would be sufficient to obtain good performance. On the contrary, these runs always produced immediate crashes in trials. 7 Discussion and Conclusions We have presented a new, model-free approximate policy iteration algorithm called LSPI, which is inspired by the LSTD algorithm. This algorithm is able to use either a stored repository of samples or samples generated dynamically from trajectories. It performs action selection and approximate policy iteration entirely in value function space, without any need for model. In contrast to other approaches to approximate policy iteration, it does not require any sort of approximate policy function. In comparison to the memory based approach of Ormoneit and Sen [11], our method makes stronger use of function approximation. Rather than using our samples to implicitly construct an approximate model using kernels, we operate entirely in value function space and use our samples directly in the value function projection step. As noted by Boyan [3] the matrix used by LSTD and LSPI can be viewed as an approximate, compressed model. This is most compelling if the columns of are orthonormal. While this provides some intuitions, a proper transition function cannot be reconstructed directly from , making a possible interpretation of LSPI as a model based method an area for future research. In comparison to direct policy search methods [9, 8, 1, 13, 6], we offer the strength of policy iteration. Policy search methods typically make a large number of relatively small steps of gradient-based policy updates to a parameterized policy function. Our use of policy iteration generally results in a small number of very large steps directly in policy space. Our experimental results demonstrate the potential of our method. We achieved good performance on the bicycle task using a very small number of randomly generated samples that were reused across multiple steps of policy iteration. Achieving this level of performance with just a linear value function architecture did require some tweaking, but the transparency of the linear architecture made the relevant issues much more salient than would be the case with any “black box” approach. We believe that the direct approach to function approximation and data reuse taken by LSPI will make the algorithm an intuitive and easy to use first choice for many reinforcement learning tasks. In future work, we plan to investigate the application of our method to multi-agent systems and the use of density estimation to control the projection weights in our function approximator. Acknowledgments We would like to thank J. Randløv and P. Alstrøm for making their bicycle simulator available. We also thank C. Guestrin, D. Koller, U. Lerner and M. Littman for helpful discussions. The first author would like to thank the Lilian-Boudouri Foundation in Greece for partial financial support. References [1] J. Baxter and P.Bartlett. Reinforcement learning in POMDP’s via direct gradient ascent. In Proc. 17th International Conf. on Machine Learning, pages 41–48. Morgan Kaufmann, San Francisco, CA, 2000. [2] D. Bertsekas and J. Tsitsiklis. Neuro-Dynamic Programming. Athena Scientific, Belmont, Massachusetts, 1996. [3] Justin A. Boyan. Least-squares temporal difference learning. In I. Bratko and S. Dzeroski, editors, Machine Learning: Proceedings of the Sixteenth International Conference, pages 49– 56. Morgan Kaufmann, San Francisco, CA, 1999. [4] S. Bradtke and A. Barto. Linear least-squares algorithms for temporal difference learning. Machine Learning, 22(1/2/3):33–57, 1996. [5] D. Koller and R. Parr. Policy iteration for factored mdps. In Proceedings of the Sixteenth Conference on Uncertainty in Artificial Intelligence (UAI-00). Morgan Kaufmann, 2000. [6] V. Konda and J. Tsitsiklis. Actor-critic algorithms. In NIPS 2000 editors, editor, Advances in Neural Information Processing Systems 12: Proceedings of the 1999 Conference. MIT Press, 2000. [7] M. G. Lagoudakis and R. Parr. Model-Free Least-Squares policy iteration. Technical Report CS-2001-05, Department of Computer Science, Duke University, December 2001. [8] A. Ng and M. Jordan. PEGASUS: A policy search method for large MDPs and POMDPs. In Proceedings of the Sixteenth Conference on Uncertainty in Artificial Intelligence (UAI-00). Morgan Kaufmann, 2000. [9] A. Ng, R. Parr, and D. Koller. Policy search via density estimation. In Advances in Neural Information Processing Systems 12: Proceedings of the 1999 Conference. MIT Press, 2000. [10] Andrew Y. Ng, Daishi Harada, and Stuart Russell. Policy invariance under reward transformations: theory and application to reward shaping. In Proc. 16th International Conf. on Machine Learning, pages 278–287. Morgan Kaufmann, San Francisco, CA, 1999. [11] D. Ormoneit and S. Sen. Kernel-based reinforcement learning. To appear, Machine Learning, 2001. [12] J. Randløv and P. Alstrøm. Learning to drive a bicycle using reinforcement learning and shaping. In The Fifteenth International Conference on Machine Learning, 1998. Morgan Kaufmann. [13] R. Sutton, D. McAllester, S. Singh, and Y. Mansour. Policy gradient methods for reinforcement learning with function approximation. In Advances in Neural Information Processing Systems 12: Proceedings of the 1999 Conference, 2000. MIT Press.
|
2001
|
197
|
2,013
|
Fragment completion in humans and machines David Jacobs NEC Research Institute 4 Independence Way, Princeton, NJ 08540 dwj@research.nj.nec.com Bas Rokers Psychology Department at UCLA PO Box 951563, Los Angeles, CA 90095 rokers@psych.ucla.edu Archisman Rudra CS Department at NYU 251 Mercer St., New York, NY 10012 archi@cs.nyu.edu Zili Liu Psychology Department at UCLA PO Box 951563, Los Angeles CA 90095 zili@psych.ucla.edu Abstract Partial information can trigger a complete memory. At the same time, human memory is not perfect. A cue can contain enough information to specify an item in memory, but fail to trigger that item. In the context of word memory, we present experiments that demonstrate some basic patterns in human memory errors. We use cues that consist of word fragments. We show that short and long cues are completed more accurately than medium length ones and study some of the factors that lead to this behavior. We then present a novel computational model that shows some of the flexibility and patterns of errors that occur in human memory. This model iterates between bottom-up and top-down computations. These are tied together using a Markov model of words that allows memory to be accessed with a simple feature set, and enables a bottom-up process to compute a probability distribution of possible completions of word fragments, in a manner similar to models of visual perceptual completion. 1 Introduction This paper addresses the problem of retrieving items in memory from partial information. Human memory is remarkable for its flexibility in handling a wide range of possible retrieval cues. It is also very accurate, but not perfect; some cues are more easily used than others. We hypothesize that memory errors occur in part because a trade-off exists between memory accuracy and the complexity of neural hardware needed to perform complicated memory tasks. If this is true, we can gain insight into mechanisms of human memory by studying the patterns of errors humans make, and we can model human memory with systems that produce similar patterns as a result of constraints on computational resources. We experiment with word memory questions of the sort that arise in a game called superghost. Subjects are presented with questions of a form: ‘*p*l*c*’. They must find a valid English word that matches this query, by replacing each ‘*’ with zero or more letters. So for this example, ‘place’, ’application’, and ‘palace’ would all be valid answers. In effect, the subject is given a set of letters and must think of a word that contains all of those letters, in that order, with other letters added as needed. Most of the psychological literature on word completion involves the effects of priming certain responses with recent experience (Shacter and Tulving[18]). However, priming is only able to account for about five percent of the variance in a typical fragment completion task (Olofsson and Nyberg[13], Hintzman and Hartry[6]). We describe experiments that show that the difficulty of a query depends on what we call its redundancy. This measures the extent to which all the letters in the query are needed to find a valid answer. We show that when we control for the redundancy of queries, we find that the difficulty of answering questions increases with their length; queries with many letters tend to be easy only because they tend to be highly redundant. We then describe a model that mimics these and other properties of human memory. Our model is based on the idea that a large memory system can gain efficiency by keeping the comparison between input and items in memory as simple as possible. All comparisons use a small, fixed set of features. To flexibly handle a range of queries, we add a bottomup process that computes the probability that each feature is present in the answer, given the input and a generic, Markov model of words. So the complexity of the bottom-up computation does not grow with the number of items in memory. Finally, the system is allowed to iterate between this bottom up and a top down process, so that a new generic model of words is constructed based on a current probability distribution over all words in memory, and this new model is combined with the input to update the probability that each feature is present in the answer. Previous psychological research has compared performance of word-stem and wordfragment completion. In the former a number of letters (i.e. a fragment) is given beginning with the first letter(s) of the word. In the latter, the string of letters given may begin at any point in the word, and adjacent letters in the fragment do not need, but may, be adjacent in the completed word. For example, for stem completion the fragment “str” may be completed into “string”, but for fragment completion also into “satire”. Performance for wordfragment completion is lower than word-stem completion (Olofsson and Nyberg[12]). In addition words, for which the ending fragment is given, show performance closer to wordstem completion than to word-fragment completion (Olofsson and Nyberg[13]). Seidenberg[17] proposed a model based on tri-grams. Srinivas et al.[21] indicate that assuming orthographic encoding is in most cases sufficient to describe word completion performance in humans. Orthographic Markov models of words have often been used computationally, as, for example, in Shannon’s[19] famous work. Following this work, our model is also orthographic. We find that a bigram rather than a trigram representation is sufficient, and leads to a simpler model. Contradicting evidence exists for the influence of fragment length on word completion. Oloffsson and Nyberg [12] failed to find a difference between two and three letter fragments on words of length of five to eight letters. However this might have been due to the fact that in their task, each fragment has a unique completion. Many recurrent neural networks have been proposed as models of associative memory (Anderson[1] contains a review). Perhaps most relevant to our work are models that use an input query to activate items from a complete dictionary in memory, and then use these items to alter the activations of the input. For example, in the Interactive Activation model of Rumelhart and McClelland[16], the presence of letters activates words, which boost the activity of the letters they contain. In Adaptive Resonance models (Carpenter and Grossberg[3]) activated memory items are compared to the input query and de-activated if they do not match. Also similar in spirit to our approach is the bidirectional model of Kosko[10] (for more recent work see, eg., Sommer and Palm[20]). Other models iteratively combine top-down and bottom-up information (eg., Hinton et al.[5], Rao and Ballard[14]), although these are not used as part of a memory system with complete items stored in memory. Our model differs from all of these in using a Markov model as an intermediate layer between the input and the dictionary. This allows the model to answer superghost queries, and leads to different computational mechanisms that we will detail. We find that superghost queries seem more natural to people than associative memory word problems (compare the superghost query “think of a word with an a” to the associative memory query “think of a word whose seventh letter is an a”). However, it is not clear how to extend most models of associative memory to handle superghost problems. Our use of features is more related to feedforward neural nets, and especially the “information bottleneck” approach of Tishby, Pereira and Bialek[22] (see also Baum, et al.[2]). Our work differs from feedforward methods in that our method is iterative, and uses features symmetrically to relate the memory to input in both directions. Our approach is also related to work on visual object recognition that combines perceptual organization and top-down knowledge (see Ullman[23]). Our model is inspired by Mumford’s[11] and Williams and Jacobs’[24] use of Markov models of contours for bottom-up perceptual completion. Especially relevant to our work is that of Grimes and Mozer[4]. Simultaneous with our work ([8]) they use a bigram model to solve anagram problems, in which letters are unscrambled to match words in a dictionary. They also use a Markov model to find letter orderings that conform with the statistics of English spelling. Their model is quite different in how this is done, due to the different nature of the anagram problem. They view anagram solving as a mix of low-level processing and higher level cognitive processes, while it is our goal to focus just on lower level memory. 2 Experiments with Human Subjects In our experiments, fragments and matching words were drawn from a large standard corpus of English text. The frequency of a word is the number of times it appears in this corpus. The frequency of a fragment is the sum of the frequency of all words that the fragment matches. We used fragments of length two to eight, discarding any fragments with frequency lower than one thousand. Fragments selected for an experiment were presented in random order. In our first experiment we systematically varied the length of the fragments, but otherwise selected them from a uniform, random distribution. Consequently, shorter fragments tended to match more words, with greater total frequency. In the second experiment, fragments were selected so that a uniform distribution of frequencies was ensured over all fragment lengths. For example, we used length two fragments that matched unusually few words. As a result the average frequency in experiment two is also much lower than in experiment one. A fragment was presented on a computer screen with spaces interspersed, indicating the possibility of letter insertion. The subject was required to enter a word that would fit the fragment. A subject was given 10 seconds to produce a completion, with the possibility to give up. For each session 50 fragments were presented, with a similar number of fragments of each length. Reaction times were recorded by measuring the time elapsed between the fragment first appearing on screen and the subject typing the first character of a matching word. Words that did not match the fragment or did not exist in the corpus were marked as not completed. Each experiment was completed by thirty-one subjects. The subjects were undergraduate students at Rutgers University, participating in the experiment for partial credit. Total time 1 2 3 4 5 6 7 8 9 0 0.2 0.4 0.6 0.8 1 Fragment Length Fraction Completed 1 2 3 4 5 6 7 8 9 0 0.2 0.4 0.6 0.8 1 Fragment Length Fraction Completed 1 2 3 4 5 6 7 8 9 0 0.2 0.4 0.6 0.8 1 Fragment Length Fraction Completed R0 R1 R2 R3 R4 Figure 1: Fragment completion as a function of fragment length for randomly chosen cues (top-left) and cues of equal frequency (top-right). On the bottom, the equal frequency cues are divided into five groups, from least redundancy (R0) to most (R5) . spent on the task varied from 15 minutes to close to one hour. Results For each graph we plot the number of fragments completed divided by the number of fragments presented (Figure 1). Error bars are calculated as
, where is the percent correct in the sample, and
is the number of trials. This assumes that all decisions are independent and correct with probability ; more precise results can be obtained by accounting for between-subject variance, but roughly the same results hold. For random, uniformly chosen fragments, there is a U-shaped dependence of performance on length. Controlling for frequency reduces performance because on average lower frequency fragments are selected. The U-shaped curve is flattened, but persists; hence Ushaped performance is not just due to frequency Finally, we divide the fragments from the two experiments into five groups, according to their redundancy. This is a rough measure of how important each letter is in finding a correct answer to the overall question. It is the probability that if we randomly delete a letter from the fragment and find a matching word, that this word will match the full fragment. Specifically, let denote the frequency of a query fragment of length (total frequency of words that match it). Let denote the frequency of the fragment that results when we delete the ’th letter from the query (note, ). Then redundancy is: !"# . In all cases where there is a significant difference, greater redundancy leads to better performance. In almost all cases, when we control for redundancy performance decreases with length. We will discuss the implications of these experiments after describing corresponding experiments with our model. 3 Using Markov Models for Word Retrieval We now describe a model of word memory in which matching between the query and memory is mediated by a simple set of features. Specifically, we use bigrams (adjacent pairs of letters) as our feature set. We denote the beginning and end of a word using the symbols ‘0’ and ‘1’, respectively, so that bigram probabilities also indicate how often individual letters begin or end a word. Bottom up processing of a cue is done using this as a Markov model of words. Then bigram probabilities are used to trigger words in memory that might match the query. Our algorithm consists of three steps. First, we compute a prior distribution on how likely each word in memory is to match our query. In our simulations, we just use a uniform distribution. However, this distribution could reflect the frequency with which each word occurs in English. It could also be used to capture priming phenomena; for example, if a word has been recently seen, its prior probability could increase, making it more likely that the model would retrieve this word. Then, using these we compute a probability that each bigram will appear if we randomly select a bigram from a word selected according to our prior distribution. Second, we use these bigram probabilities as a Markov model, and compute the expected number of times each bigram will occur in the answer, conditioned on the query. That is, as a generic model of words we assume that each letter in the word depends on the adjacent letters, but is conditionally independent of all others. This conditional independence allows us to decompose our problem into a set of small, independent problems. For example, consider the query ‘*p*l*c*’. Implicitly, each query begins with ‘0’ and ends with ‘1’, so the expected number of times any bigram will appear in the completed word is the sum of the number of times it appears in the completions of the fragments: ‘0*p’, ‘p*l’, ‘l*c’, and ‘c*1’. To compute this, we assume a prior distribution on the number of letters that will replace a ‘*’ in the completed word. We use an exponential model, setting the probability of
letters to be (in practice we truncate
at 5 and normalize the probabilities). A similar model is used in the perceptual completion of contours ([11, 24]). Using these priors, it becomes straightforward to compute a probability distribution on the bigrams that will appear in the completed cue. For a fixed
, we structure this problem as a belief net with
bigrams, and each bigram depending on only its neighbors. The conditional probability of each bigram given its neighbor comes from the Markov model, and we can solve the problem with belief propagation. Beginning the third step of the algorithm, we know the expected number of times that each bigram appears in the completed cue. Each bigram then votes for all words containing that bigram. The weight of this vote is the expected number of times each bigram appears in the completed cue, divided by the prior probability of each bigram, computed in step 1. We combine these votes multiplicatively. We update the prior for each word as the product of these votes with the previous probability. We can view this an approximate computation of the probability of each word being the correct answer, based on the likelihood that a bigram appears in the completed cue, and our prior on each word being correct. After the third step, we once again have a probability that each word is correct, and can iterate, using this probability to initialize step one. After a small number of iterations, we terminate the algorithm and select the most probable word as our answer. Empirically, we find that the answer the algorithm produces often changes in the first one or two iterations, and then generally remains the same. The answer may or may not actually match the input cue, and by this we judge whether it is correct or incorrect. We can view this algorithm as an approximate computation of the probability that each 1 2 3 4 5 6 7 8 9 0.4 0.5 0.6 0.7 0.8 0.9 1 Fragment Length Fraction Completed 1 2 3 4 5 6 7 8 9 0.4 0.5 0.6 0.7 0.8 0.9 1 Fragment Length Fraction Completed 1 2 3 4 5 6 7 8 9 0.4 0.5 0.6 0.7 0.8 0.9 1 Fragment Length Fraction Completed R0 R1 R2 R3 R4 Figure 2: Performance as a function of cue length, for cues of frequency between 4 and 22 (top-left) and between 1 and 3 (top-right). On the bottom, we divide the first set of cues into five groups ranging from the least redundant (R0) to the most (R4). word matches the cue, where the main approximation comes from using a small set of features to bring the cue into contact with items in memory. Denote the number of features by (with a bigram representation, ), the number of features in each word by (ie., the word length plus one), the number of words by , and the maximum number of blanks replacing a ‘*’ by
. Then steps one and three require O(mw) computation, and step two requires O(Fn) computation. In a neural network, the primary requirement would be bidirectional connections between each feature (bigram) and each item in memory. Therefore, computational simplicity is gained by using a small feature set, at the cost of some approximation in the computation. Experiments We have run experiments to compare the performance of this model to that of human subjects. For simplicity, we used a memory of 6,040 words, each with eight characters. First, we simulated the conditions described in Olofsson and Nyberg[12] comparing word stem and word fragment completion. To match their experiments, we used a modified algorithm that handled cues in which the number of missing letters can be specified. We used cues that specified the first three letters of a word, the last three letters, or three letters scattered throughout the word. The algorithm achieved accuracy of 95% in the first case, 87% in the second, and 80% in the third. This qualitatively matches the results for human subjects. Note that our algorithm treats the beginning and end of words symmetrically. Therefore, the fact that it performs better when the first letters of the word are given than when the last are given is due to regularities in English spelling, and is not built into the algorithm. Next we simulated conditions comparable to our own experiments on human subjects, using superghost cues. First we selected cues of varying length that match between four and twenty-two words in the dictionary. Figure 2-top-left shows the percentage of queries the algorithm correctly answered, for cues of lengths two to seven. This figure shows a U-shaped performance curve qualitatively similar to that displayed by human subjects. We also ran these experiments using cues that matched one to three words (Figure 2-topright). These very low frequency cues did not display this U-shaped behavior. The algorithm performs differently on fragments with very low frequency because in our corpus the shorter of these cues had especially low redundancy and the longer fragments had especially high redundancy, in comparison to fragments with frequencies between 4 and 22. Next (Figure 2-bottom) we divided the cues into five groups of equal size, according to their redundancy. We can see that performance increases with redundancy and decreases with cue length. Discussion Our experiments indicate two main effects in human word memory that our model also shares. First, performance improves with the redundancy of cues. Second, when we control for this, performance drops with cue length. Since redundancy tends to increase with cue length, this creates two conflicting tendencies that result in a U-shaped memory curve. We conjecture that these factors may be present in many memory tasks, leading to U-shaped memory curves in a number of domains. In our model, the fact that performance drops with cue length is a result of our use of a simple feature set to mediate matching the cue to words in memory. This means that not all the information present in the cue is conveyed to items in memory. When the length of a cue increases, but its redundancy remains low, all the information in the cue remains important in getting a correct answer, but the amount of information in the cue increases, making it harder to capture it all with a limited feature set. This can account for the performance of our model; similar mechanisms may account for human performance as well. On the other hand, the extent to which redundancy grows with cue length is really a product of the specific words in memory and the cues chosen. Therefore, the exact shape of the performance curve will also depend on these factors. This may partly explain some of the quantitative differences between our model and human performance. Finally, we also point out that our measure of redundancy is rather crude. In particular, it tends to saturate at very high or very low levels. So, for example, if we add a letter to a cue that is already highly redundant, the new letter may not be needed to find a correct answer, but that is not reflected by much of an increase in the cue’s redundancy. 4 Conclusions We have proposed superghost queries as a domain for experimenting with word memory, because it seems a natural task to people, and requires models that can flexibly handle somewhat complicated questions. We have shown that in human subjects, performance on superghost improves with the redundancy of a query, and otherwise tends to decrease with word length. Together, these effects results in a U-shaped performance curve. We have proposed a computational model that uses a simple, generic model of words to map a superghost query onto a simple feature set of bigrams. This means that somewhat complicated questions can be answered while keeping comparisons between the fragments and words in memory very simple. Our model displays the two main trends we have found in human memory. It also does better at word stem completion than word fragment completion, which agrees with previous work on human memory. Future work will investigate the modification of our model to account for priming effects in memory. References [1] J. Anderson. An Introduction to Neural Networks, MIT Press, Cambridge MA. 1995. [2] E. Baum, J. Moody and F. Wilczek. “Internal Representations for Associative Memory,” Biological Cybernetics, 59:217-228, 1988. [3] G. Carpenter, and S. Grossberg. “ART 2: Self-Organization of Stable Category Recognition Codes for Analog Input Patterns,”Applied Optics, 26:4919-4930, 1987. [4] D. Grimes and M. Mozer. “The interplay of symbolic and subsymbolic processes in anagram problem solving,”NIPS, 2001. [5] G. Hinton, P. Dayan, B. Frey, and R. Neal. “The ‘Wake-Sleep’ Algorithm for Unsupervised Neural Networks,”Science, 268:1158-1161, 1995. [6] D.L. Hintzman and A.L. Hartry. Item effects in recognition and fragment completion: Contingency relations vary for different sets of words. JEP: Learning, Memory and Cognition, 17: 341-345, 1990. [7] J. Hopfield. “Neural networks and Physical Systems with Emergent Collective Computational Abilities.”Proc. of the Nat. Acad. of Science, 79:2554-2558, 1982. [8] D. Jacobs and A. Rudra. “An Iterative Projection Model of Memory,” NEC Research Institute Technical Report, 2000. [9] G.V. Jones. Fragment and schema models for recall. Memory and Cognition, 12(3):250-63, 1984. [10] B. Kosko. “Adaptive Bidirectional Associative Memory”, Applied Optics, 26(23):4947-60, 1987. [11] D. Mumford. “Elastica and Computer Vision.”C. Bajaj (Ed), Algebraic Geometry and its Applications New York: Springer-Verlag. 1994. [12] U. Olofsson and L. Nyberg. Swedish norms for completion of word stems and unique word fragments. Scandinavian Journal of Psychology, 33(2):108-16, 1992. [13] U. Olofsson and L. Nyberg. Determinants of word fragment completion. Scandinavian Journal of Psychology, 36(1):59-64, 1995. [14] R. Rao and D. Ballard. “Dynamic Model of Visual Recognition Predicts Neural Response Properties in the Visual Cortex,”Neural Computation, 9(4):721-763, 1997. [15] R.H. Ross and G.H. Bower. Comparisons of models of associative recall. Memory and Cognition, 9(1):1-16, 1981. [16] D. Rumelhart and J. McClelland. “An interactive activation model of context effects in letter perception: part 2. The contextual enhancement effect and some tests and extensions of the model”, Psychological Review, 89:60-94, 1982. [17] M.S. Seidenberg. Sublexical structures in visual word recognition: Access units or orthographic redundancy? In M. Coltheart (Ed.), Attention and performance XII, 245-263. Hillsdale, NJ: Erlbaum. 1987. [18] D.L. Shacter and E. Tulving. Memory systems. Cambridge, MA: MIT Press. 1994. [19] C. Shannon. “Prediction and Entropy of Printed English,” Bell Systems Technical Journal, 30:50-64, 1951. [20] Sommer, F., and Palm, G., 1997, NIPS:676-681. [21] K. Srinivas, H.L. Roediger 3d and S. Rajaram. The role of syllabic and orthographic properties of letter cues in solving word fragments. Memory and Cognition, 20(3):219-30, 1992. [22] N. Tishby, F. Pereira and W. Bialek. “The Information Bottleneck Method,”37th Allerton Conference on Communication, Control, and Computing. 1999. [23] S. Ullman. High-level Vision, MIT Press, Cambridge, MA. 1996. [24] L. Williams & D. Jacobs. “Stochastic Completion Fields: A Neural Model of Illusory Contour Shape and Salience”. Neural Computation, 9:837–858, 1997. Acknowledgements The authors would like to thank Nancy Johal for her assistance in conducting the psychological experiments presented in this paper.
|
2001
|
2
|
2,014
|
K-Local Hyperplane and Convex Distance Nearest Neighbor Algorithms Pascal Vincent and Yoshua Bengio Dept. IRO, Universit´e de Montr´eal C.P. 6128, Montreal, Qc, H3C 3J7, Canada vincentp,bengioy @iro.umontreal.ca http://www.iro.umontreal.ca/ vincentp Abstract Guided by an initial idea of building a complex (non linear) decision surface with maximal local margin in input space, we give a possible geometrical intuition as to why K-Nearest Neighbor (KNN) algorithms often perform more poorly than SVMs on classification tasks. We then propose modified K-Nearest Neighbor algorithms to overcome the perceived problem. The approach is similar in spirit to Tangent Distance, but with invariances inferred from the local neighborhood rather than prior knowledge. Experimental results on real world classification tasks suggest that the modified KNN algorithms often give a dramatic improvement over standard KNN and perform as well or better than SVMs. 1 Motivation The notion of margin for classification tasks has been largely popularized by the success of the Support Vector Machine (SVM) [2, 15] approach. The margin of SVMs has a nice geometric interpretation1: it can be defined informally as (twice) the smallest Euclidean distance between the decision surface and the closest training point. The decision surface produced by the original SVM algorithm is the hyperplane that maximizes this distance while still correctly separating the two classes. While the notion of keeping the largest possible safety margin between the decision surface and the data points seems very reasonable and intuitively appealing, questions arise when extending the approach to building more complex, non-linear decision surfaces. Non-linear SVMs usually use the “kernel trick” to achieve their non-linearity. This conceptually corresponds to first mapping the input into a higher-dimensional feature space with some non-linear transformation and building a maximum-margin hyperplane (a linear decision surface) there. The “trick” is that this mapping is never computed directly, but implicitly induced by a kernel. In this setting, the margin being maximized is still the smallest Euclidean distance between the decision surface and the training points, but this time measured in some strange, sometimes infinite dimensional, kernel-induced feature space rather than the original input space. It is less clear whether maximizing the margin in this new space, is meaningful in general (see [16]). 1for the purpose of this discussion, we consider the original hard-margin SVM algorithm for two linearly separable classes. A different approach is to try and build a non-linear decision surface with maximal distance to the closest data point as measured directly in input space (as proposed in [14]). We could for instance restrict ourselves to a certain class of decision functions and try to find the function with maximal margin among this class. But let us take this even further. Extending the idea of building a correctly separating non-linear decision surface as far away as possible from the data points, we define the notion of local margin as the Euclidean distance, in input space, between a given point on the decision surface and the closest training point. Now would it be possible to find an algorithm that could produce a decision surface which correctly separates the classes and such that the local margin is everywhere maximal along its surface? Surprisingly, the plain old Nearest Neighbor algorithm (1NN) [5] does precisely this! So why does 1NN in practice often perform worse than SVMs? One typical explanation, is that it has too much capacity, compared to SVM, that the class of function it can produce is too rich. But, considering it has infinite capacity (VC-dimension), 1NN is still performing quite well... This study is an attempt to better understand what is happening, based on geometrical intuition, and to derive an improved Nearest Neighbor algorithm from this understanding. 2 Fixing a broken Nearest Neighbor algorithm 2.1 Setting and definitions The setting is that of a classical classification problem in (the input space). We are given a training set of points
and their corresponding class label
! #"$ %" '&(
)!* where )!* is the number of different classes. The + , pairs are assumed to be samples drawn from an unknown distribution ./ 0 . Barring duplicate inputs, the class labels associated to each 1 define a partition of : let * 1 32 4 576 . The problem is to find a decision function 8 97: <; " that will generalize well on new points drawn from .= >0 . 8 9 should ideally minimize the expected classification error, i.e. minimize ?A@CB DFE G'HJILKM NFO$P where ?A@ denotes the expectation with respect to .= 0 and DFE G(HJQKM N R denotes the indicator function, whose value is & if 8 9 TS and U otherwise. In the previous and following discussion, we often refer to the concept of decision surface, also known as decision boundary. The function 8 9 corresponding to a given algorithm defines for any class 6 #" two regions of the input space: the region V * W 2 8 9 5X6 and its complement Y V * . The decision surface for class 6 is the “boundary” between those two regions, i.e. the contour of V * , and can be seen as a Z Y & dimensional manifold (a “surface” in ) possibly made of several disconnected components. For simplicity, when we mention the decision surface in our discussion we consider only the case of two class discrimination, in which there is a single decision surface. When we mention a test point, we mean a point 1 that does not belong to the training set and for which the algorithm is to decide on a class 8 9 . By distance, we mean the usual Euclidean distance in input-space . The distance between two points [ and \ will be written ] [ \ or alternatively ^[ Y \(^ . The distance between a single point and a set of points _ is the distance to the closest point of the set: ] + _ `ba<cJdfeg'h ] + i . The K-neighborhood jLk of a test point is the set of the l points of whose distance to is smallest. The K-c-neighborhood jmk * of a test point is the set of l points of * whose distance to is smallest. By Nearest Neighbor algorithm (1NN) we mean the following algorithm: the class of a test point is decided to be the same as the class of its closest neighbor in _ . By K-Nearest Neighbor algorithm (KNN) we mean the following algorithm: the class of a test point is decided to be the same as the class appearing most frequently among the K-neighborhood of . 2.2 The intuition Figure 1: A local view of the decision surface produced by the Nearest Neighbor (left) and SVM (center) algorithms, and how the Nearest Neighbor solution gets closer to the SVM solution in the limit, if the support for the density of each class is a manifold which can be considered locally linear (right). Figure 1 illustrates a possible intuition about why SVMs outperforms 1NNs when we have a finite number of samples. For classification tasks where the classes are considered to be mostly separable,2 we often like to think of each class as residing close to a lowerdimensional manifold (in the high dimensional input space) which can reasonably be considered locally linear3. In the case of a finite number of samples, “missing” samples would appear as “holes” introducing artifacts in the decision surface produced by classical Nearest Neighbor algorithms. Thus the decision surface, while having the largest possible local margin with regard to the training points, is likely to have a poor small local margin with respect to yet unseen samples falling close to the locally linear manifold, and will thus result in poor generalization performance. This problem fundamentally remains with the K Nearest Neighbor (KNN) variant of the algorithm, but, as can be seen on the figure, it does not seem to affect the decision surface produced by SVMs (as the surface is constrained to a particular smooth form, a straight line or hyperplane in the case of linear SVMs). It is interesting to notice, if the assumption of locally linear class manifolds holds, how the 1NN solution approaches the SVM solution in the limit as we increase the number of samples. To fix this problem, the idea is to somehow fantasize the missing points, based on a local linear approximation of the manifold of each class. This leads to modified Nearest Neighbor algorithms described in the next sections.4 2By “mostly separable” we mean that the Bayes error is almost zero, and the optimal decision surface has not too many disconnected components. 3i.e. each class has a probability density with a “support” that is a lower-dimensional manifold, and with the probability quickly fading, away from this support. 4Note that although we never generate the “fantasy”points explicitly, the proposed algorithms are really equivalent to classical 1NN with fantasized points. 2.3 The basic algorithm Given a test point , we are really interested in finding the closest neighbor, not among the training set , but among an abstract, virtually enriched training set that would contain all the fantasized “missing” points of the manifold of each class, locally approximated by an affine subspace. We shall thus consider, for each class 6 , the local affine subspace that passes through the l points of the K-c neighborhood of . This affine subspace is typically l Y & dimensional or less, and we will somewhat abusively call it the “local hyperplane”.5 Formally, the local hyperplane can be defined as k * 5 i i k N ) N
(1) where )
>) j k * . Another way to define this hyperplane, that gets rid of the constraint & , is to take a reference point within the hyperplane as an origin, for instance the centroid6 ) k k N ) . This same hyperplane can then be expressed as k * 5 i i ) k N Y ; (2) where Y ; ) Y ) . Our modified nearest neighbor algorithm then associates a test point to the class 6 whose hyperplane k * is closest to . Formally 8 9 A$a<c d * g ] + k * , where ] F k * , is logically called K-local Hyperplane Distance, hence the name K-local Hyperplane Distance Nearest Neighbor algorithm (HKNN in short). Computing, for each class 6 ] F k * , a!c d eg !#" $ H QK ^ Y i ^ a<cJd % g'& (*),+ + + + + Y ) Y k N Y ; + + + + + (3) amounts to solving a linear system in , that can be easily expressed in matrix form as: .- /0 / 1- / Y ) (4) where and ) are Z dimensional column vectors,
k - , and is a Z2 l matrix whose columns are the Y ; vectors defined earlier.7 5Strictly speaking a hyperplane in an 3 dimensional input space is an 35476 affine subspace, while our “local hyperplanes”can have fewer dimensions. 6We could be using one of the 8 neighbors as the reference point, but this formulation with the centroid will prove useful later. 7Actually there is an infinite number of solutions to this system since the 4 9 :7; are linearly dependent: remember that the initial formulation had an equality constraint and thus only 8<4=6 effective degrees of freedom. But we are interested in >@?BADCFEHGJI K ?BA LFL not in M so any solution will do. Alternatively, we can remove one of the 4 9 :7; from the system so that it has a unique solution. 2.4 Links with other paradigms The proposed HKNN algorithm is very similar in spirit to the Tangent Distance Algorithm [13]. k * can be seen as a tangent hyperplane representing a set of local directions of transformation (any linear combination of the Y ; vectors) that do not affect the class identity. These are invariances. The main difference is that in HKNN these invariances are inferred directly from the local neighborhood in the training set, whereas in Tangent Distance, they are based on prior knowledge. It should be interesting (and relatively easy) to combine both approaches for improved performance when prior knowledge is available. Previous work on nearest-neighbor variations based on other locally-defined metrics can be found in [12, 9, 6, 7], and is very much related to the more general paradigm of Local Learning Algorithms [3, 1, 10]. We should also mention close similarities between our approach and the recently proposed Local Linear Embedding [11] method for dimensionality reduction. The idea of fantasizing points around the training points in order to define the decision surface is also very close to methods based on estimating the class-conditional input density [14, 4]. Besides, it is interesting to look at HKNN from a different, less geometrical angle: it can be understood as choosing the class that achieves the best reconstruction (the smallest reconstruction error) of the test pattern through a linear combination of l particular prototypes of that class (the l neighbors). From this point of view, the algorithm is very similar to the Nearest Feature Line (NFL) [8] method. They differ in the fact that NFL considers all pairs of points for its search rather than the local l neighbors, thus looking at many ( ) lines (i.e. 2 dimensional affine subspaces), rather than at a single l Y & dimensional one. 3 Fixing the basic HKNN algorithm 3.1 Problem arising for large K One problem with the basic HKNN algorithm, as previously described, arises as we increase the value of l , i.e. the number of points considered in the neighborhood of the test point. In a typical high dimensional setting, exact colinearities between input patterns are rare, which means that as soon as l Z , any pattern of (including nonsensical ones) can be produced by a linear combination of the l neighbors. The “actual” dimensionality of the manifold may be much less than l . This is due to “near-colinearities” producing directions associated to small eigenvalues of the covariance matrix /0 that are but noise, that can lead the algorithm to mistake those noise directions for “invariances”, and may hurt its performance even for smaller values of l . Another related issue is that the linear approximation of the class manifold by a hyperplane is valid only locally, so we might want to restrict the “fantasizing” of class members to a smaller region of the hyperplane. We considered two ways of dealing with these problems.8 3.2 The convex hull solution One way to avoid the above mentioned problems is to restrict ourselves to considering the convex hull of the neighbors, rather than the whole hyperplane they support (of which the convex hull is a subset). This corresponds to adding a constraint of U to equation (1). Unlike the problem of computing the distance to the hyperplane, the distance to the convex hull cannot be found by solving a simple linear system, but typically requires solving a quadratic programming problem (very similar to the one of SVMs). While this 8A third interesting avenue, which we did not have time to explore, would be to keep only the most relevant principal components of : , ignoring those corresponding to small eigenvalues. is more complex to implement, it should be mentioned that the problems to be solved are of a relatively small dimension of order l , and that the time of the whole algorithm will very likely still be dominated by the search of the l nearest neighbors within each class. This algorithm will be referred to as K-local Convex Distance Nearest Neighbor Algorithm (CKNN in short). 3.3 The “weight decay” penalty solution This consists in incorporating a penalty term to equation (3) to penalize large values of (i.e. it penalizes moving away from the centroid, especially in non essential directions): ] + k * a<cJd % g'& (0)+ + + + + Y ) Y k N Y ; + + + + + k N (5) The solution for is given by solving the linear system / D / / Y ) where D is the Z72TZ identity matrix. This is equation (4) with an additional diagonal term. The resulting algorithm is a generalization of HKNN (basic HKNN corresponds to U ). 4 Experimental results We performed a number of experiments, to highlight different properties of the algorithms: A first 2D toy example (see Figure 2) graphically illustrates the qualitative differences in the decision surfaces produced by KNN, linear SVM and CKNN. Table 1 gives quantitative results on two real-world digit OCR tasks, allowing to compare the performance of the different old and new algorithms. Figure 3 illustrates the problem arising with large l , mentioned in Section 3, and shows that the two proposed solutions: CKNN and HKNN with an added weight decay , allow to overcome it. In our final experiment, we wanted to see if the good performance of the new algorithms absolutely depended on having all the training points at hand, as this has a direct impact on speed. So we checked what performance we could get out of HKNN and CKNN when using only a small but representative subset of the training points, namely the set of support vectors found by a Gaussian Kernel SVM. The results obtained for MNIST are given in Table 2, and look very encouraging. HKNN appears to be able to perform as well or better than SVMs without requiring more data points than SVMs. Figure 2: 2D illustration of the decision surfaces produced by KNN (left, K=1), linear SVM (middle), and CKNN (right, K=2). The “holes”are again visible in KNN. CKNN doesn’t suffer from this, but keeps the objective of maximizing the margin locally. 5 Conclusion From a few geometrical intuitions, we have derived two modified versions of the KNN algorithm that look very promising. HKNN is especially attractive: it is very simple to implement on top of a KNN system, as it only requires the additional step of solving a small and simple linear system, and appears to greatly boost the performance of standard KNN even above the level of SVMs. The proposed algorithms share the advantages of KNN (no training required, ideal for fast adaptation, natural handling of the multi-class case) and its drawbacks (requires large memory, slow testing). However our latest result also indicate the possibility of substantially reducing the reference set in memory without loosing on accuracy. This suggests that the algorithm indeed captures essential information in the data, and that our initial intuition on the nature of the flaw of KNN may well be at least partially correct. References [1] C. G. Atkeson, A. W. Moore, and S. Schaal. Locally weighted learning. Artificial Intelligence Review, 1996. [2] B. Boser, I. Guyon, and V. Vapnik. An algorithm for optimal margin classifiers. In Fifth Annual Workshop on Computational Learning Theory, pages 144–152, Pittsburgh, 1992. [3] L. Bottou and V. Vapnik. Local learning algorithms. Neural Computation, 4(6):888–900, 1992. [4] Olivier Chapelle, Jason Weston, L´eon Bottou, and Vladimir Vapnik. Vicinal risk minimization. In T.K. Leen, T.G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems, volume 13, pages 416–422, 2001. [5] T.M. Cover and P.E. Hart. Nearest neighbor pattern classification. IEEE Transactions on Information Theory, 13(1):21–27, 1967. [6] J. Friedman. Flexible metric nearest neighbor classification. Technical Report 113, Stanford University Statistics Department, 1994. [7] Trevor Hastie and Robert Tibshirani. Discriminant adaptive nearest neighbor classification and regression. In David S. Touretzky, Michael C. Mozer, and Michael E. Hasselmo, editors, Advances in Neural Information Processing Systems, volume 8, pages 409–415. The MIT Press, 1996. [8] S.Z. Li and J.W. Lu. Face recognition using the nearest feature line method. IEEE Transactions on Neural Networks, 10(2):439–443, 1999. [9] J. Myles and D. Hand. The multi-class measure problem in nearest neighbour discrimination rules. Pattern Recognition, 23:1291–1297, 1990. [10] D. Ormoneit and T. Hastie. Optimal kernel shapes for local linear regression. In S. A. Solla, T. K. Leen, and K-R. Mller, editors, Advances in Neural Information Processing Systems, volume 12. MIT Press, 2000. [11] Sam Roweis and Lawrence Saul. Nonlinear dimensionality reduction by locally linear embedding. Science, 290(5500):2323–2326, Dec. 2000. [12] R. D. Short and K. Fukunaga. The optimal distance measure for nearest neighbor classification. IEEE Transactions on Information Theory, 27:622–627, 1981. [13] P. Y. Simard, Y. A. LeCun, J. S. Denker, and B. Victorri. Transformation invariance in pattern recognition — tangent distance and tangent propagation. Lecture Notes in Computer Science, 1524, 1998. [14] S. Tong and D. Koller. Restricted bayes optimal classifiers. In Proceedings of the 17th National Conference on Artificial Intelligence (AAAI), pages 658–664, Austin, Texas, 2000. [15] V.N. Vapnik. The Nature of Statistical Learning Theory. Springer, New York, 1995. [16] Bin Zhang. Is the maximal margin hyperplane special in a feature space? Technical Report HPL-2001-89, Hewlett-Packards Labs, 2001. Table 1: Test-error obtained on the USPS and MNIST digit classification tasks by KNN, SVM (using a Gaussian Kernel), HKNN and CKNN. Hyper parameters were tuned on a separate validation set. Both HKNN and CKNN appear to perform much better than original KNN, and even compare favorably to SVMs. Data Set Algorithm Test Error Parameters used USPS KNN 4.98% l & (6291 train, SVM 4.33% & U(U 1000 valid., HKNN 3.93% l &f U 2007 test points) CKNN 3.98% l U MNIST KNN 2.95% l (50000 train, SVM 1.30%
& U'U 10000 valid., HKNN 1.26% l & U 10000 test points) CKNN 1.46% l
U 0.012 0.014 0.016 0.018 0.02 0.022 0.024 0.026 0.028 0.03 0.032 0 20 40 60 80 100 120 error rate K CKNN basic HKNN HKNN, lambda=1 HKNN, lambda=10 Figure 3: Error rate on MNIST as a function of l for CKNN, and HKNN with different values of . As can be seen the basic HKNN algorithm performs poorly for large values of l . As expected, CKNN is relatively unaffected by this problem, and HKNN can be made robust through the added “weight decay”penalty controlled by . Table 2: Test-error obtained on MNIST with HKNN and CKNN when using a reduced training set made of the 16712 support vectors retained by the best Gaussian Kernel SVM. This corresponds to 28% of the initial 60000 training patterns. Performance is even better than when using the whole dataset. But here, hyper parameters l and were chosen with the test set, as we did not have a separate validation set in this setting. It is nevertheless remarkable that comparable performances can be achieved with far fewer points. Data Set Algorithm Test Error Parameters used MNIST (16712 train s.v., HKNN 1.23% l U & U 10000 test points) CKNN 1.36% l
|
2001
|
20
|
2,015
|
Thomas L .. Griffiths & Joshua B. Tenenbaum Department of Psychology Stanford University, Stanford, CA 94305 {gruffydd,jbt}©psych. stanford. edu Abstract Estimating the parameters of sparse multinomial distributions is an important component of many statistical learning tasks. Recent approaches have used uncertainty over the vocabulary of symbols in a multinomial distribution as a means of accounting for sparsity. We present a Bayesian approach that allows weak prior knowledge, in the form of a small set of approximate candidate vocabularies, to be used to dramatically improve the resulting estimates. We demonstrate these improvements in applications to text compression and estimating distributions over words in newsgroup data. 1 Introduction Sparse multinomial distributions arise in many statistical domains, including natural language processing and graphical models. Consequently, a number of approaches to parameter estimation for sparse multinomial distributions have been suggested [3]. These approaches tend to be domain-independent: they make little use of prior knowledge about a specific domain. In many domains where multinomial distributions are estimated there is often at least weak prior knowledge about' the potential structure of distributions, such as a set of hypotheses about restricted vocabularies from which the symbols might be generated. Such knowledge can be solicited from experts or obtained from unlabeled data. We present a method for Bayesian_parameter estimation in sparse discrete domains that exploits this weak form of prior knowledge to improve estimates over knowledge-free approaches. 1.1 Bayesian parameter estimation for multinomial distributions Following the presentation in [4], we consider a language ~ containing L distinct symbols. A multinomial distribution is specified by a parameter vector f) == (Ol, ... ,f)L), where f)i is the probability of an observation being symbol i. Consequently, we have the constraints that Ef==l f)i == 1 and (h ~ 0, i == 1, ... ,L. The task of multinomial estimation is to take a data set D and produce a'vector f) that results in a good approximation to the distribution that produced D. In this case, D consists of N independent observations Xl, ... xN drawn from the distribution to be estimated, which can be summarized by the statistics Ni specifying the number of times the ith symbol occurs in the data. D also determines the set ~o of symbols that occur in the data. Stated in this way, multinomial estimation involves predicting the next observation based on the data. Specifically, we wish to calculate P(XN+1ID). The Bayesian estimate for this probability is given by PL(xN+lID) =I p(XN+1IB)P(BID)dB where P(XN +1 10) follows from the multinomial distribution corresponding to O. The posterior probability P(OID) can be obtained via Bayes rule L P(OID) oc P(DIO)P(O) == P(8}II ONi i==l where P(O) is the prior probability of a given O. Laplace used this method with a uniform prior over 0 to give the famous "law of succession" [6J. A more general approach is to assume a Dirichlet prior over (), which is conjugate to the multinomial distribution and gives P(XN+l = ilD) = Ni +LCY.i (1) N + l:j==l O!.j where the ai are the hyperparameters of the Dirichlet distribution. Different estimates are obtained for different choices of the ai, with most approaches making the simplifying assumption that ai == O!. for all i. Laplace's law results from a == 1. The case with a == 0.5 is the Jeffreys-Perks law or Expected Likelihood Estimation [2] [5J [9J, while using arbitrary O!. is Lidstone's law [7]. 1.2 EstiIllating sparse Illultinomial distributions Several authors have extended the Bayesian approach to sparse multinomial distributions, in which only a restricted vocabulary of symbols are used, by maintaining uncertainty over these vocabularies. In [10], Ristad uses assumptions about the probability of strings based upon different vocabularies to give the estimate { (Ni + l)/(N +L) if kO == L PR (X N +1 == ilD) == (Ni + l)(N + 1 - kO)/(N2 + N + 2kO) if kO < L 1\ Ni > 0 kO(kO + l)/(L - kO)(N2 + N + 2kO) otherwise where kO == I~oIis the size of the smallest vocabulary consistent with the data. A different approach is taken by Friedman and Singer in [4], who point out that Ristad's method is a special case of their framework. Friedman and Singer consider the vocabulary V ~ :E to be a random variable, allowing them to write p.(XN +1 == ilD) == L p(XN +1 == ilV, D)P(VID) (2) v where P{XN +1 == ilV, D) results from a Dirichlet prior over the symbols in V, p(XN +1 == ilV, D) == {~it'ja if i E v: (3) o otherWIse and by Bayes' rule and the properties of Dirichlet priors P(VID) oc P(DIV)P(V) { ~fJ~~(a) niE~O r(~·t~a) P(V) EO ~ V ( ) otherwise 4 Friedman and Singer assume a hierarchical prior over V, such that all vocabularies of cardinality k are given equal probability, namely P(S == k)/(t), where P(S == k) is the probability that the size of the vocabulary (IVI) is k. It follows that if i E ~o, p(XN +I == ilD) == Lk :+1~P(S == kiD). If i ¢ ~o, it is necessary to estimate the proportion of V that contain i for a given k. The simplified result is PF(XN +1 == ilD) == { %tt~aC(D,L) L-kO(1- C(D,L)) where . h P(S k) - k! r(ka:) WIt mk ==. == (k-kO)!· r(N+ka:) . 2 Ivlaking use of weak prior knowiedge if i E ~o otherwise (5) Friedman and Singer assume a prior that gives equal probability to all vocabularies of a given cardinality. However, many real-world tasks provide limited knowledge about the structure of distributions that we can build into our methods for parameter estimation. In the context of sparse multinomial estimation, one instance of such knowledge the importance of specific vocabularies. For example, in predicting the next character in a file, our predictions could be facilitated by considering the fact that most files either use a vocabulary consisting of ASCII printing characters (such as text files), or all possible characters (suc~ as object files). This kind of structural knowledge about a domain is typically easier to solicit from experts than accurate distributional information, and forms a valuable informational resource. If we have this kind of prior knowledge, we can restrict our attention to a subset of the 2L possible vocabularies. fu particular, we can specify a set of vocabularies V which we consider as hypotheses for the vocabulary used in producing D, where the elements of V are specified by our knowledge of the domain. This stands as a compromise between Friedman and Singer's approach, in which V consists of all vocabularies, and traditional Bayesian parameter estimation as represented by Equation 1, in which V consists of only the vocabulary containing all words. To do this, we explicitly evaluate the sum given in Equation 2, where the sum over V includes all V E V. This sum remains tractable when V is a small subset of the possible vocabularies, and the efficiency is aided by the fact that P(DIV) shares common terms across all V which can cancel in normalization. The intuition behind this approach is that it attempts to classify thetarget distribution as using one of a known set of vocabularies, where the vocabularies are obtained either from experts or from unlabeled data. Applying standard Bayesian multinomial estimation within this vocabulary gives enough flexibility for the method to capture a range of distributions, while making use of our weak prior knowledge. 2.1 An illustration: Text compression Text compression is an effective test of methods for multinomial estimation. Adaptive coding can be performed by specifying a method for calculating a distribution over the probability of the next byte in a file based upon the preceding bytes [1]. The extent to which the file is compressed depends upon the quality of these predictions. To illustrate the utility of including prior knowledge, we follow Ristad in using the Calgary text compression corpus [1]. This corpus consists of 19 files of Table 1: Text compression lengths (in bytes) on the Calgary corpus file size kO NH(Ni/N ) Pv PF PR PL PJ bib 111261 81 72330 18 89 92 269 174 book1 768771 82 435043 219 105 116 352 219 book2 610856 96 365952 94 115 124 329 212 geo 102400 256 72274 161 162 165 165 161 nellS 377109 98 244633 89 113 116 304 201 obj1 21504 256 15989 126 127 129 129 126 obj2 246814 256 193144 182 184 190 189 182 paper1 53161 95 33113 71 94 100 236 156 paper2 82199 91 47280 75 94 105 259 167 paper3 46526 84 27132 70 85 92 238 154 paper4 13286 80 7806 58 72 79 190 126 paper5 11954 91 7376 57 79 83 181 122 paper6 38105 93 23861 68 90 95 223 149 pic 513216 159 77636 205 16~ 216 323 205 progc 39611 92 25743 68 - 89 91 222 150 progl 71646 87 42720 74 91 97 253 164 progp 49379 89 30052 71 89 94 236 155 trans 93695 99 64800 169 101 105 252 169 several different types, each using some subset of 256 possible characters (L == 256). The files include Bib'IEXsource (bib), formatted English text (book*, paper*), geological data (geo), newsgroup articles (news), object files (obj*), a bit-mapped picture (pic), programs in three different languages (prog*) and a terminal transcript (trans). The task was to estimate the distribution from which characters in the file were drawn based upon the first N characters and thus predict the N + 1st character. Performance was measured in terms of the length of the resulting file, where the contribution of the N + 1st character to the length is log2 P(XN+lID). The results are expressed as the number of bytes required to encode the file relative to the empirical entropy NH(Ni/N) as assessed by Ristad [10]. Results are shown in Table 1. Pv is the restricted vocabulary model outlined above, with V consisting of just two hypotheses: one corresponding to binary files, containing all 256 characters, and one consisting of a 107 character vocabulary representing formatted English. The latter vocabulary was estimated from 5MB of English text, C code, Bib'IEXsource, and newsgroup data from outside the Calgary corpus. PF is Friedman and Singer's method. For both of these approaches, a was set to 0.5, to allow direct comparison to the Jeffreys-Perks law, PJo PR and PL are Ristad's and Laplace's laws respectively. Py outperformed the other methods on all files based upon English text, bar bookl, and all files using all 256 symbolsl . The high performance followed from rapid classification of these files as using the appropriate vocabulary in V. When the vocabulary included all symbols Py performed as PJ, which gave the best predictions for these files. 1A number of excellent techniques for· text compression exist that outperform all of those presented here. We have not included these techniques for comparison because our interest is in using text compression as a means of assessing estimation procedures, rather than as an end in itself. We thus consider only methods for multinomial estimation as our comparison. group. 2.2 Maintaining uncertainty in vocabularies The results for book1 illustrate a weakness of the approach outlined above. The file length for Py is higher than those for PF and PR , despite the fact that the file uses a text-based vocabulary. This file contains two characters that were not encountered in the data used to construct V. These characters caused Py to default to the unrestricted vocabulary of all 256 characters. From that point Py corresponded to PJ, which gave poor results on this file. This behavior results from the assumption that the candidate vocabularies in V are completely accurate. Since in many cases the knowledge that informs the vocabularies in V may be imperfect, it is desirable to allow for uncertainty in vocabularies. This uncertainty will be reflected in the fact that symbols outside V are expected to occur with a vocabulary-specific probability ty, p(XN+1 == ilV, D) == { (1 - (L -IVI)ty) N~~t~la if i E V ty otherwise where Ny == I:iEY N i · It follows that r(IVla) r(Ni + a) P(DIV) = (1 - (L -IVJ)€V)NV€t"-Nv r(N + 1V1a:) r(a:) y iE:EonY which replaces Equations 3-4 in specifying Py . When V is determined by the judgments of domain experts, ty is the probability that an unmentioned word actually belongs to a particular vocabulary. While it may not be the most efficient use of such data, the V E V can also be estimated from some form of unlabeled data. In this case, Friedman and Singer's approach provides a means of setting ty. Friedman and Singer explicitly calculate the probability that an unseen word is in V based upon a dataset: from the second condition of Equation 5, we find that we should set ty == L_1IYI (1- C(D, L)). We use this method below. 3 Bayesian parameter estimation in natural language Statistical natural language processing often uses sparse multinomial distributions over large vocabularies of words. In different contexts, different vocabularies will be used. By specifying a basis set ofvocabularies, we can perform parameter estimation by classifying distributions according to their vocabulary. This idea was examined using data from 20 different Usenet newsgroups. This dataset is commonly used in testing text classification algorithms (eg. [8]). Ten newsgroups were used to estimate a set of vocabularies V with corresponding ty. These vocabularies were used in estimating multinomial distributions on these newsgroups and ten others. The dataset was 20news-18827, which consists of the 20newsgroups data with headers and duplicates removed, and was preprocessed to remove all punctuation, capitalization, and distinct numbers. The articl~s in each of the 20 newsgroups were then divided into three sets. The first 500 articles from ten newsgroups were used to estimate the candidate vocabularies V and uncertainty parameters ty. Articles 501700 for all 20 newsgroups were used as training data for multinomial estimation. Articles 701-900 for all 20 newgroups were used as testing data. Following [8], a dictionary was built up by running over the 13,000 articles resulting from this division, and all words that occurred only once were mapped to an "unknown" word. The resulting dictionary contained L == 54309 words. As before, the restricted vocabulary method (Py), Friedman and Singer's method (PF), and Ristad's (PR ), Laplace's (PL ) and the Jeffreys-Perks (PJ ) laws were ap~~~ soc.religion.christian talk.politics.guns comp.sys.ibm.pc.hardware ~~~ rec.sport.hockey scLelectronics comp.windows.x ~~~ rec.autos rec.sport.baseball scLcrypt ~~~ scLmed comp.os.ms-windows.misc misc.forsale ~::=:= ~-;""", r.?';'~;:;" ~~~ comp.sys.mac.hardware talk.religion.misc comp.graphics ~~~ rec.motorcycles scLspace alt.atheism .... " talk.politics. mise talk.politics.mideast 17 '. 18 . 16 18 11 100 10000 50000 100 10000 Number of words F~gure 1: Cross-entropy of predictions on newsgroup data as a function of the logarithm of the number of words. The abscissa is at the empirical entropy of the test distribution. The top ten panels (talk.polities.mideast and those to its right) are for the newsgroups with unknown vocabularies. The bottom ten are for those that contributed vocabularies to V, trained and tested on novel data. PL and PJ are both indicated with dotted lines, but PJ always performs better than PL. The box on talk.polities.mideast indicates the point at which Pv defaults to the full vocabulary, as the number of unseen words makes this vocabulary more likely. At this point, the line for Pv joins the line for PJ , since both methods give the same estimates of the distribution. plied to the task. Both Pv and PF used a == 0.5 to facilitate comparison with PJ . 'V featured one vocabulary that contained all words in the dictionary, and ten vocabularies each corresponding to the words used in the first 500 articles of one of the newsgroups designated for this purpose. €y was estimated as outlined above. Testing for each newsgroup consisted of taking words from the 200 articles assigned for training purposes, estimating a. distribution using each method, and then computing the cross-entropy between that distribution and an empirical estimate of the true distribution. The cross-entropy is H{Q; P) == Ei Qi log2 Pi, where Q is the true distribution and P is the distribution produced by the estimation method. Q was given by the maximum likelihood estimate formed from the word frequencies in all 200 articles assigned for testing purposes. The testing procedure was conducted with just 100 words, and then in increments of 450 up to a total of 10000 words. Long-run performance was examined on talk.polities.mideast and talk.polities.mise, each trained with 50000 words. The results are shown in Figure 1. As expected, Py consistently outperformed the other methods on the newsgroups that contributed to V. However, performance on novel newsgroups was also greatly improved. As can be seen in Figure 2, the novel newsgroups were classified to appropriate vocabularies - for example all words rec.autos I-----------------rec.motorcycles r--+-ta1k.politics.guns ,-------T---+----------- talk.politics.mideast rl\--_f-------T----'r----------- alt.atheism frt-'\.;:::::::::::;f:=.='=f--t---------- soc.religion.christian l-+-if-Hf-------t---------- scLspace ,-~~?.;~~~~~~g~ey scLmed rec.sport.baseball scLcrypt misc.forsale comp.graphics comp.sys.mac.hardware talk.religion. misc talk.politics.misc comp.os.ms-windows.m~c'-----------------com~sy&ibm.p~hardware o 10000 Number of words Figure 2: Classification of newsgroup vocabularies. The lines illustrate the vocabulary which had maximum posterior probability for each of the ten test newsgroups after exposure to differing numbers of words. The vocabularies in V are listed along the left hand side of the axis, and the lines are identified with newsgroups by the labels on the right hand side. Lines are offset to facilitate identification. talk.religion.misc had the highest posterior probability for alt.atheism and soc. religion. christian, while rec. autos had highest posterior probability for rec .motorcycles. The proportion of word types occurring in the test data but not the vocabulary to which the novel newsgroups were classified ranged between 30.5% and 66.2%, with a mean of 42.2%. This illustrates that even approximate knowledge can facilitate predictions: the basis set of vocabularies allowed the high frequency words in the data to be modelled effectively, without excess mass being attributed to the low frequency novel word tokens. Long-run performance on talk.politics .mideast illustrates the same defaulting behavior that was shown for text classification: when the data become more probable under the vocabulary containing all words than under a restricted vocabulary the method defaults to the Jeffreys-Perks law. This guarantees that the method will tend to perform no worse than PJ when unseen words are encountered in sufficient proportions. This is desirable, since PJ gives good estimates once N becomes large. 4 Discussion Bayesian approaches to parameter estimation for sparse multinomial distributions have employed the notion of a restricted vocabulary from which symbols are produced. In many domains where such distributions are estimated; there is often at least limited knowledge about the structure of these vocabularies. By considering just the vocabularies suggested by such knowledge, together with some uncertainty concerning those vocabularies, we can achieve very good estimates of distributions in these domains. We have presented a Bayesian approach that employs limited prior knowledge, and shown that it outperforms a range of approaches to multinomial estimation for both text compression and a task involving natural language. While our applications in this paper estimated approximate vocabularies from data, the real promise of this approach lies with domain knowledge solicited from experts. Experts are typically better at providing qualitative structural information than quantitative distributional information, and our approach provides a way of using this information in parameter estimation. For example, in the context of parameter estimation for graphical models to be used in medical diagnosis, distinguishing classes of symptoms might be informative in determining the parameters governing their relationship to diseases. This form of knowledge is naturally translated into a set of vocabularies to be considered for each such distribution. More complex applications to natural language lllay also be possible, such as using syntactic information in estimating probabilities for n-gram models. The approach we have presented in this paper provides a simple way to allow this kind of limited domain knowledge to be useful in Bayesian parameter estimation. References [1] T. C. Bell, J. G. Cleary, and 1. H. Witten. Text compression. Prentice Hall, 1990. [2] G. E. P. Box and G. C. Tiao. Bayesian Inference in Statistical Analysis. AddisonWesley, 1973. [3] S. F. Chen and J. Goodman. An empirical study of smoothing techniques for language modeling. Technical Report TR-10-98, Center for Research in Computing Technology, Harvard University, 1998. [4] N. Friedman and Y. Singer. Efficient Bayesian parameter estimation in large discrete domains. In Neural Information Processing Systems, 1998. [5] H. Jeffreys. An invariant form for the prior probability in estimation problems. Proceedings of the Royal Society A, 186:453-461, 1946. [6] P.-S. Laplace. Philosophical Essay on Probabilities. Springer-Verlag, 1995. Originally published 1825. [7] G. Lidstone. Note on the general case of the Bayes-Laplace formula for inductive or a posteriori probabilities. Transactions of the Faculty of Actuaries, 8:182-192, 1920. [8] K. Nigam, A. K. Mccallum, S. Thrun, and T. Mitchell. Text classification fro'in labeled and unlabeled documents using EM. Machine Learning, 39:103-134, 2000. [9] W. Perks. Some observations on inverse probability, including a new indifference rule. Journal of the Institute of Actuaries, 73:285-312, 1947. [10] E. S. Ristad. A natural law ·of succession. Technical Report CS-TR-895-95, Department of Computer Science, Princeton University, 1995.
|
2001
|
21
|
2,016
|
Bayesian time series classification Peter Sykacek Department of Engineering Science University of Oxford Oxford, OX1 3PJ, UK psyk@robots.ox.ac.uk Stephen Roberts Department of Engineering Science University of Oxford Oxford, OX1 3PJ, UK sjrob@robots.ox.ac.uk Abstract This paper proposes an approach to classification of adjacent segments of a time series as being either of classes. We use a hierarchical model that consists of a feature extraction stage and a generative classifier which is built on top of these features. Such two stage approaches are often used in signal and image processing. The novel part of our work is that we link these stages probabilistically by using a latent feature space. To use one joint model is a Bayesian requirement, which has the advantage to fuse information according to its certainty. The classifier is implemented as hidden Markov model with Gaussian and Multinomial observation distributions defined on a suitably chosen representation of autoregressive models. The Markov dependency is motivated by the assumption that successive classifications will be correlated. Inference is done with Markov chain Monte Carlo (MCMC) techniques. We apply the proposed approach to synthetic data and to classification of EEG that was recorded while the subjects performed different cognitive tasks. All experiments show that using a latent feature space results in a significant improvement in generalization accuracy. Hence we expect that this idea generalizes well to other hierarchical models. 1 Introduction Many applications in signal or image processing are hierarchical in the sense that a probabilistic model is built on top of variables that are the coefficients of some feature extraction technique. In this paper we consider a particular problem of that kind, where a Gaussian and Multinomial observation hidden Markov model (GMOHMM) is used to discriminate coefficients of an Auto Regressive (AR) process as being either of classes. Bayesian inference is known to give reasonable results when applied to AR models ([RF95]). The situation with classification is similar, see for example the seminal work by [Nea96] and [Mac92]. Hence we may expect to get good results if we apply Bayesian techniques to both stages of the decision process separately. However this is suboptimal since it meant to establish a no probabilistic link between feature extraction and classification. Two arguments suggest the building of one probabilistic model which combines feature extraction and classification: Since there is a probabilistic link, the generative classifier acts as a prior for feature extraction. The advantage of using this prior is that it naturally encodes our knowledge about features as obtained from training data and other sensors. Obviously this is the only setup that is consistent with Bayesian theory ([BS94]). Since all inferences are obtained from marginal distributions, information is combined according to its certainty. Hence we expect to improve results since information from different sensors is fused in an optimal manner. 2 Methods 2.1 A Gaussian and Multinomial observation hidden Markov model As we attempt to classify adjacent segments of a time series, it is very likely that we find correlations between successive class labels. Hence our model has a hidden Markov model ([RJ86]) like architecture, with diagonal Gaussian observation models for continuous variables and Multinomial observation models for discrete variables. We call the architecture a Gaussian and Multinomial observation hidden Markov model or GMOHMM for short. Contrary to the classical approach, where each class is represented by its own trained HMM, our model has class labels which are child nodes of the hidden state variables. Figure 1 shows the directed acyclic graph (DAG) of our model. We use here the convention found in [RG97], where circular nodes are latent and square nodes are observed variables. 2.1.1 Quantities of interest We regard all variables in the DAG that represent the probabilistic model of the time series as quantities of interest. These are the hidden states, , the variables of the latent feature space, , and , the class labels, , the sufficient statistics of the AR process, , and the segments of the time series, . The DAG shows the observation model only for the
-th state. We have latent feature variables, , which represent the coefficients of the preprocessing model for of the
-th segment at sensor . The state conditional distributions, , are modeled by diagonal Gaussians. Variable is the latent model indicator which represents the model order of the preprocessing model and hence the dimension of . The corresponding observation model is a Multinomial-one distribution. The third observation, , represents the class label of the
-th segment. The observation model for is again a Multinomial-one distribution. Note that depending on whether we know the class label or not, can be a latent variable or observed. The child node of and is the observed variable , which represents a sufficient statistics of the corresponding time series segment. The proposed approach requires to calculate the likelihoods !
repeatedly. Hence using the sufficient statistics is a computational necessity. Finally we use " to represent the precision of the residual noise model. The noise level is a nuisance parameter which is integrated over. 2.1.2 Model coefficients Since we integrate over all unknown quantities, there is no conceptual difference between model coefficients and the variables described above. However there is a qualitative difference. Model parameters exist only once for the entire GMOHMM, whereas there is an individual quantity of interest for every segment
. Furthermore the model coefficients are only updated during model inference whereas all quantities of interest are updated during model inference and for prediction. We have three different prior counts, #%$ , #& and #' , which define the Dirichlet priors of the corresponding probabilities. Variable ( denotes the transition probabilities, that is )* ,+ /. (1032 . The model assumes a stationary hidden state sequence. This allows us to obtain the unconditional prior probability of states 4 from the recurrence relation 56$7 8 /. (159 8:; . The prior probability of the first hidden state, 5 $ , is therefore the normalized eigenvector of the transition probability matrix ( that corresponds to the eigenvalue . Variable represents the probabilities of class , )* . 032 , which are conditional on as well. The prior probabilities for observing the model indicator are represented by 5 . The probability )* 8 . 5 0 2 is again conditional on the state . As was mentioned above, represents the model order of the time series model. Hence another interpretation of 5 is that of state dependent prior probabilities for observing particular model orders. The observation models for are dynamic mixtures of Gaussians, with one model for each sensor . Variables and 1 represent the coefficients of all Gaussian kernels. Hence 1 8 is a variate Gaussian distribution. Another interpretation is that the discrete indicator variables 8 and determine together with and 1 a Gaussian prior over . The nodes , , , , and / define a hierarchical prior setting which is discussed below. s 1 ϕ s ϕ λ λ λ s λ δP 1 h 1 g 1 α 1 Σ β1 1 ξ 1 κ δ δ T di−1 t i T W di+1 id W 1 µ i,1 i,1 Xi,1 i,1 I i,1 P i,s Xi,s I i,s P δ P κs ξs αs µs Σs s β hs gs i,s i,s α α Figure 1: This figure illustrates the details of the proposed model as a directed acyclic graph. The graph shows the model parameters and all quantities of interest: denotes the hidden states of the HMM; are the class labels of the corresponding time series segments; are the latent coefficients of the time series model and the corresponding model indicator variables; is the precision of the residual noise. For tractable inference, we extract from the time series the sufficient statistics . All other variables denote model coefficients: ( are the transition probabilities; are the probabilities for class 3 ; and 1 are mean vectors and covariance matrices of the Gaussian observation model for sensor ; and ) are the probabilities for observing . 2.2 Likelihood and priors for the GMOHMM Suppose that we are provided with segments of training data, .
. The likelihood function of the GMOHMM parameters is then obtained by summation over all possible sequences, , of latent states, . The sums and integrals under the product make the likelihood function of Equation (1) highly nonlinear. This may be resolved by using Gibbs sampling [GG84], which uses tricks similar to those of the expectation maximization algorithm. )* /. )*
-3 " -3 -3 " (1) "! )* :
2 #$ 7 %&' Gibbs sampling requires that we obtain full conditional distributions1 we can sample from. The conjugate priors are adopted from [RG97]. Below square brackets and index ( are used to denote a particular component of a vector or matrix. Each component mean, 032 , is given a Gaussian prior: 0 2*),+ - : , with denoting the mean and : the inverse covariance matrix. As we use diagonal covariance matrices, we may give each diagonal element an independent Gamma prior: * 0 2.( (0/ :;)1 (0/ , where denotes the shape parameter and (0/ denotes the inverse scale parameter. The hyperparameter, 2! , gets a component wise Gamma hyper prior: (0/ )31 /4(0/ . The state conditional class probabilities, 032 , get a Dirichlet prior: 032 )65 # & ' ' # & . The transition probabilities, ( 0 2 , get a Dirichlet prior: ( 0 2 )75 #$ ' ' #$ . The probabilities for observing different model orders, 5 3 0 2 , depend on the state . Their prior is Dirichlet 5 032 )85 # ' '9' # '/ . The precision gets a Jeffreys’ prior, i.e. the scale parameter :<; is set to 0. Values for are between and = , 8 is set between > ' = and and /?(0/ is typically between A@AB (?/ ! and > @AB (?/ ! , with B (0/ denoting the input range of maximum likelihood estimates for (0/
. The mean, , is the midpoint of the maximum likelihood estimates (0/
. The inverse covariance matrix A(?/ . 4@?B C(0/ ! , where B A(0/ is again the range of the estimates at sensor . We set the prior counts #& and #$ and #' to . 2.3 Sampling from the posterior During model inference we need to update all unobserved variables of the DAG, whereas for predictions we update only the variables summarized in section 2.1.1. Most of the updates are done using the corresponding full conditional distributions, which have the same functional forms as the corresponding priors. These full conditionals follow closely from what was published previously in [Syk00], with some modifications necessary (see e.g. [Rob96]), because we need to consider the Markov dependency between successive hidden states. As the derivations of the full conditionals do not differ much from previous work, we will omit them here and instead concentrate on an illustration how to update the latent feature space,
. 2.3.1 A representation of the latent feature space The AR model in Equation (2) is a linear regression model. We use :<D 2 to denote the AR coefficients, to denote the model order and E4 / to denote a sample from the noise process, which we assume to be Gaussian with precision . F / .HG 2 D : D 2 F GJI /LKE4 / (2) As is indicated by the subscript , the value of the I -th AR coefficient depends on the model order. Hence AR coefficients are not a convenient representation of the latent feature 1These are the distributions obtained when we condition on all other variables of the DAG. space. A much more convenient representation is provided by using reflection coefficients, , (statistically speaking they are partial correlation coefficients), which relate to AR coefficients via the order recursive Levinson algorithm. Below we use vector notation and the symbol 2 to denote the upside down version of the AR coefficient vector. 2 +. 2 K 2 +- 2 2 +- (3) We expect to observe only such data that was generated from dynamically stable AR processes. For such processes, the latent density is defined on G / 2
2 . This is in contrast with the proposed DAG, where we use a finite Gaussian mixture as probabilistic model for the latent variable, which is is defined on 2 . In order to avoid this mismatch, we reparameterise the space of reflection coefficients by applying ! , to obtain a more convenient representation of the latent features. . ! ;" (4) 2.3.2 Within dimensional updates The within dimensional updates can be done with a conventional Metropolis Hastings step. Integrating out , we obtain a Student t distributed likelihood function of the AR coefficients. In order to obtain likelihood ratio 1, we propose from the multivariate Student-t distribution shown below, reparameterise in terms of reflection coefficients and apply the ! transformation. $# . ! ;" /%&# (5) where # ) ' (*) with ) . + :;-, . + :; B *G , $ + :;-, =. G The proposal uses + to denote the -dimensional sample auto-covariance matrix, B is the sample variance, , . B ' '9' B 2 +/ $ is a vector of sample autocorrelations at lags to K and N denotes the number of samples of the time series . The proposal in Equation (5) gives a likelihood ratio of . The corresponding acceptance probability is : ..0/ 2134 # 65 5 587 92 7 9 2 5 5 5 5 5 5 7 2 7 2 5 5 5 :<; = ' (6) The determinant of the Jacobian arises because we transform the AR coefficients using Equations (3) and (4). 2.3.3 Updating model orders Updating model orders requires us to sample across different dimensional parameter spaces. One way of doing this is by using the reversible jump MCMC which was recently proposed in [Gre95]. We implement the reversible jump move from parameter space > 3 2 to parameter space > 2 +- as partial proposal. That is we propose a reflection coefficient from a distribution that is conditional on the AR coefficient / . Integrating out the precision of the noise model ! we obtain again a Student-t distributed likelihood. This suggests the following proposal: # . ! ; / (7) where ) ' ( with . G ! . G ! = G . G . B K = $ , K2 $ + ! . B +"! K = , $ K2 $ + * ' Equation (7) makes use of the sufficient statistics of the K -dimensional AR process, . B ' ' B +"! . We use to denote the number of observations and B to denote the estimated auto covariance at time lag to obtain , $ . B ' ' B +/ and + as dimensional sample covariance matrix. Assuming that the probability of proposing this move is independent of , the proposal from > 3 2 to > 2 +- has acceptance probability : . . / G ! ! ! :
= 1 :;! 1 ! # K G ! 8' (8) If we attempt an update from > 2 +- to > 2 , we have to invert the second argument of the .0/ operation in Equation (8). 3 Experiments Convergence of all experiments is analysed by applying the method suggested in [RL96] to the sequence of observed data likelihoods (equation (1), when filling in all variables). 3.1 Synthetic data Our first evaluation uses synthetic data. We generate a first order Markov sequence as target labels (2 state values) with 200 samples used for training and 600 used for testing. Each sample is used as label of a segment with 200 samples from an auto regressive process. If the label is , we generate data using reflection coefficients > ' G > ' > ' . If the label is = , we use the model > ' G > ' > ' . The driving noise has variance . Due to sampling effects we obtain a data set with Bayes error > . In order to make the problem more realistic, we use a second state sequence to replace =0> of the segments with white noise. These “artifacts” are not correlated with the class labels. In order to assess the effect of using a latent feature space, we perform three different tests: In the first run we use conventional feature extraction with a third order model and estimates found with maximum likelihood; In a second run we use again a third order model but integrate over feature values; Finally the third test uses the proposed architecture with a prior over model order which is “flat” between > and . When compared with conditioning on feature estimates, the latent features show increased likelihood. The likelihood gets even larger when we regard both the feature values and the model orders of the preprocessing stage as random variables. As can be seen in figure 2, this effect is also evident when we look at the generalization probabilities which become larger as well. We explain this by sharper “priors” over feature values and model orders, which are due to the information provided by temporal context2 of every segment. This reduces the variance of the observation models which in turn increases likelihoods and target probabilities. Table 1 shows that these higher probabilities correspond to a significant improvement in generalization accuracy. 50 100 150 200 250 300 350 400 450 500 550 600 0 0.5 1 Probabilities from conditioning 50 100 150 200 250 300 350 400 450 500 550 600 0 0.5 1 Probabilities from integrating over features 50 100 150 200 250 300 350 400 450 500 550 600 0 0.5 1 Probabilities from integrating over model orders and features Figure 2: This figure shows the generalization probabilities obtained with different settings. We see that the class probabilities get larger when we regard features as random variables. This effect is even stronger when both the features and the model orders are random variables. 3.2 Classification of cognitive tasks The data used in these experiments is EEG recorded from 5 young, healthy and untrained subjects while they perform different cognitive tasks. We classify 2 task pairings: auditorynavigation and left motor-right motor imagination. The recordings were taken from 3 electrode sites: T4, P4 (right tempero-parietal for spatial and auditory tasks), C3’ , C3” (left motor area for right motor imagination) and C4’ , C4” (right motor area for left motor imagination). The ground electrode was placed just lateral to the left mastoid process. The data were recorded using an ISO-DAM system (gain of and fourth order band pass filter with pass band between Hz and Hz). These signals were sampled with 384 Hz and 12 bit resolution. Each cognitive experiment was performed times for seconds. Classification uses again the same settings as with the synthetic problem. The summary in table 1 shows results obtained from fold cross validation, where one experiment is used for testing whereas all remaining data is used for training. We observe again significantly improved results when we regard features and model orders as latent variables. The values in brackets are the significance levels for comparing integration of features with conditioning and full integration with integration over feature values only. 4 Discussion We propose in this paper a novel approach to hierarchical time series processing which makes use of a latent feature representation. This understanding of features and model orders as random variables is a direct consequence of applying Bayesian theory. Empirical 2In a multi sensor setting there is spatial context as well. Table 1: Generalization accuracies of different experiments experiment conditioning marginalize features full integration synthetic L' L' = ( ' > : - ) ' ( > ' >0> = ) left vs. right motor ' <' ( ' = > : ) ' ( > ' > ) auditory vs. navigation ' = ' > ' > = ' ( = ' > : ) evaluations show that theoretical arguments are confirmed by significant improvements in generalization accuracy. The only disadvantage of having a latent feature space is that all computations get more involved, since there are additional variables that have to be integrated over. However this additional complexity does not render the method intractable since the algorithm remains polynomial in the number of segments to be classified. Finally we want to point out that the improvements observed in our results can only be attributed to the idea of using a latent feature space. This idea is certainly not limited to time series classification and should generalize well to other hierarchical architectures. Acknowledgments We want to express gratitude to Dr. Rezek, who made several valuable suggestions in the early stages of this work. We also want to thank Prof. Stokes, who provided us with the EEG recordings that were used in the experiments section. Finally we are also grateful for the valuable comments provided by the reviewers of this paper. Peter Sykacek is currently funded by grant Nr. F46/399 kindly provided by the BUPA foundation. References [BS94] J. M. Bernardo and A. F. M. Smith. Bayesian Theory. Wiley, Chichester, 1994. [GG84] S. Geman and D. Geman. Stochastic relaxation, Gibbs distributions and the Bayesian restoration of images. IEEE Transactions on Pattern Analysis and Machine Intelligence, 6:721–741, 1984. [Gre95] P. J. Green. Reversible jump Markov chain Monte Carlo computation and Bayesian model determination. Biometrika, 82:711–732, 1995. [Mac92] D. J. C. MacKay. The evidence framework applied to classification networks. Neural Computation, 4:720–736, 1992. [Nea96] R. M. Neal. Bayesian Learning for Neural Networks. Springer, New York, 1996. [RF95] J. J. K. ´O Ruanaidh and W. J. Fitzgerald. Numerical Bayesian Methods Applied to Signal Processing. Springer-Verlag, New York, 1995. [RG97] S. Richardson and P. J. Green. On Bayesian analysis of mixtures with an unknown number of components. Journal Royal Stat. Soc. B, 59:731–792, 1997. [RJ86] L. R. Rabiner and B. H. Juang. An introduction to Hidden Markov Models. IEEE ASSP Magazine, 3(1):4–16, 1986. [RL96] A. E. Raftery and S. M. Lewis. Implementing MCMC. In W.R. Gilks, S. Richardson, and D.J. Spiegelhalter, editors, Markov Chain Monte Carlo in practice, chapter 7, pages 115– 130. Chapman & Hall, London, Weinheim, New York, 1996. [Rob96] C. P. Robert. Mixtures of distributions: inference and estimation. In W. R. Gilks, S. Richardson, and D.J. Spiegelhalter, editors, Markov Chain Mont Carlo in Practice, pages 441–464. Chapman & Hall, London, 1996. [Syk00] P. Sykacek. On input selection with reversible jump Markov chain Monte Carlo sampling. In S.A. Solla, T.K. Leen, and K.-R. M¨uller, editors, Advances in Neural Information Processing Systems 12, pages 638–644, Boston, MA, 2000. MIT Press.
|
2001
|
22
|
2,017
|
Computing Time Lower Bounds for Recurrent Sigmoidal Neural Networks Michael Schmitt Lehrstuhl Mathematik und Informatik, Fakultat fUr Mathematik Ruhr-Universitat Bochum, D- 44780 Bochum, Germany mschmitt@lmi.ruhr-uni-bochum.de Abstract Recurrent neural networks of analog units are computers for realvalued functions. We study the time complexity of real computation in general recurrent neural networks. These have sigmoidal, linear, and product units of unlimited order as nodes and no restrictions on the weights. For networks operating in discrete time, we exhibit a family of functions with arbitrarily high complexity, and we derive almost tight bounds on the time required to compute these functions. Thus, evidence is given of the computational limitations that time-bounded analog recurrent neural networks are subject to. 1 Introduction Analog recurrent neural networks are known to have computational capabilities that exceed those of classical Turing machines (see, e.g., Siegelmann and Sontag, 1995; Kilian and Siegelmann, 1996; Siegelmann, 1999). Very little, however, is known about their limitations. Among the rare results in this direction, for instance, is the one of Sima and Orponen (2001) showing that continuous-time Hopfield networks may require exponential time before converging to a stable state. This bound, however, is expressed in terms of the size of the network and, hence, does not apply to fixed-size networks with a given number of nodes. Other bounds on the computational power of analog recurrent networks have been established by Maass and Orponen (1998) and Maass and Sontag (1999). They show that discretetime recurrent neural networks recognize only a subset of the regular languages in the presence of noise. This model of computation in recurrent networks, however, receives its inputs as sequences. Therefore, computing time is not an issue since the network halts when the input sequence terminates. Analog recurrent neural networks, however, can also be run as "real" computers that get as input a vector of real numbers and, after computing for a while, yield a real output value. No results are available thus far regarding the time complexity of analog recurrent neural networks with given size. We investigate here the time complexity of discrete-time recurrent neural networks that compute functions over the reals. As network nodes we allow sigmoidal units, linear units, and product units- that is, monomials where the exponents are adjustable weights (Durbin and Rumelhart, 1989). We study the complexity of real computation in the sense of Blum et aI. (1998). That means, we consider real numbers as entities that are represented exactly and processed without restricting their precision. Moreover, we do not assume that the information content of the network weights is bounded (as done, e.g., in the works of Balcazar et aI., 1997; Gavalda and Siegelmann, 1999). With such a general type of network, the question arises which functions can be computed with a given number of nodes and a limited amount of time. In the following, we exhibit a family of real-valued functions ft, l 2: 1, in one variable that is computed by some fixed size network in time O(l). Our main result is, then, showing that every recurrent neural network computing the functions ft requires at least time nW/4). Thus, we obtain almost tight time bounds for real computation in recurrent neural networks. 2 Analog Computation in Recurrent Neural Networks We study a very comprehensive type of discrete-time recurrent neural network that we call general recurrent neural network (see Figure 1). For every k, n E N there is a recurrent neural architecture consisting of k computation nodes YI , . . . ,Yk and n input nodes Xl , ... , xn . The size of a network is defined to be the number ofits computation nodes. The computation nodes form a fully connected recurrent network. Every computation node also receives connections from every input node. The input nodes play the role of the input variables of the system. All connections are parameterized by real-valued adjustable weights. There are three types of computation nodes: product units, sigmoidal units, and linear units. Assume that computation node i has connections from computation nodes weighted by Wil, ... ,Wi k and from input nodes weighted by ViI, .. . ,Vi n. Let YI (t) , . . . ,Yk (t) and Xl (t), ... ,Xn (t) be the values of the computation nodes and input nodes at time t, respectively. If node i is a product unit, it computes at time t + 1 the value (1) that is, after weighting them exponentially, the incoming values are multiplied. Sigmoidal and linear units have an additional parameter associated with them, the threshold or bias ()i . A sigmoidal unit computes the value where (J is the standard sigmoid (J(z ) = 1/ (1 + e- Z ). If node i is a linear unit, it simply outputs the weighted sum We allow the networks to be heterogeneous, that is, they may contain all three types of computation nodes simultaneously. Thus, this model encompasses a wide class of network types considered in research and applications. For instance, architectures have been proposed that include a second layer of linear computation nodes which have no recurrent connections to computation nodes but serve as output nodes (see, e.g., Koiran and Sontag, 1998; Haykin, 1999; Siegelmann, 1999). It is clear that in the definition given here, the linear units can function as these output nodes if the weights of the outgoing connections are set to O. Also very common is the use of sigmoidal units with higher-order as computation nodes in recurrent networks (see, e.g., Omlin and Giles, 1996; Gavalda and Siegelmann, 1999; Carrasco et aI., 2000). Obviously, the model here includes these higher-order networks as a special case since the computation of a higher-order sigmoidal unit can be simulated by first computing the higher-order terms using product units and then passing their computation nodes input nodes Xl I . I sigmoidal, product, and linear units Yl . Yk t Xn I Figure 1: A general recurrent neural network of size k. Any computation node may serve as output node. outputs to a sigmoidal unit. Product units, however, are even more powerful than higher-order terms since they allow to perform division operations using negative weights. Moreover, if a negative input value is weighted by a non-integer weight, the output of a product unit may be a complex number. We shall ensure here that all computations are real-valued. Since we are mainly interested in lower bounds, however, these bounds obviously remain valid if the computations of the networks are extended to the complex domain. We now define what it means that a recurrent neural network N computes a function f : ~n --+ llt Assume that N has n input nodes and let x E ~n. Given tE N, we say that N computes f(x) in t steps if after initializing at time 0 the input nodes with x and the computation nodes with some fixed values, and performing t computation steps as defined in Equations (1), (2), and (3), one of the computation nodes yields the value f(x). We assume that the input nodes remain unchanged during the computation. We further say that N computes f in time t if for every x E ~n , network N computes f in at most t steps. Note that t may depend on f but must be independent of the input vector. We emphasize that this is a very general definition of analog computation in recurrent neural networks. In particular, we do not specify any definite output node but allow the output to occur at any node. Moreover, it is not even required that the network reaches a stable state, as with attractor or Hopfield networks. It is sufficient that the output value appears at some point of the trajectory the network performs. A similar view of computation in recurrent networks is captured in a model proposed by Maass et al. (2001). Clearly, the lower bounds remain valid for more restrictive definitions of analog computation that require output nodes or stable states. Moreover, they hold for architectures that have no input nodes but receive their inputs as initial values of the computation nodes. Thus, the bounds serve as lower bounds also for the transition times between real-valued states of discrete-time dynamical systems comprising the networks considered here. Our main tool of investigation is the Vapnik-Chervonenkis dimension of neural networks. It is defined as follows (see also Anthony and Bartlett, 1999): A dichotomy of a set S ~ ~n is a partition of S into two disjoint subsets (So , Sd satisfying So U S1 = S. A class :F of functions mapping ~n to {O, I} is said to shatter S if for every dichotomy (So , Sd of S there is some f E :F that satisfies f(So) ~ {O} and f(S1) ~ {I}. The Vapnik-Chervonenkis (VC) dimension of :F is defined as 4"'+4",IL S~ 'I -1---Y-2----Y-5~1 Y5 output Y4 Figure 2: A recurrent neural network computing the functions fl in time 2l + 1. the largest number m such that there is a set of m elements shattered by F. A neural network given in terms of an architecture represents a class of functions obtained by assigning real numbers to all its adjustable parameters, that is, weights and thresholds or a subset thereof. The output of the network is assumed to be thresholded at some fixed constant so that the output values are binary. The VC dimension of a neural network is then defined as the VC dimension of the class of functions computed by this network. In deriving lower bounds in the next section, we make use of the following result on networks with product and sigmoidal units that has been previously established (Schmitt, 2002). We emphasize that the only constraint on the parameters of the product units is that they yield real-valued, that is, not complex-valued, functions. This means further that the statement holds for networks of arbitrary order, that is, it does not impose any restrictions on the magnitude of the weights of the product units. Proposition 1. (Schmitt, 2002, Theorem 2) Suppose N is a feedforward neural network consisting of sigmoidal, product, and linear units. Let k be its size and W the number of adjustable weights. The VC dimension of N restricted to real-valued functions is at most 4(Wk)2 + 20Wk log(36Wk). 3 Bounds on Computing Time We establish bounds on the time required by recurrent neural networks for computing a family of functions fl : JR -+ JR, l 2:: 1, where l can be considered as a measure of the complexity of fl. Specifically, fl is defined in terms of a dynamical system as the lth iterate of the logistic map ¢>(x) = 4x(1 - x), that is, fl(X) { ¢>(x) ¢>(fl- l (x)) l = 1, l > 2. We observe that there is a single recurrent network capable of computing every fl in time O(l). Lemma 2. There is a general recurrent neural network that computes fl in time 2l + 1 for every l. Proof. The network is shown in Figure 2. It consists of linear and second-order units. All computation nodes are initialized with 0, except Yl, which starts with 1 and outputs 0 during all following steps. The purpose of Yl is to let the input x output Figure 3: Network Nt. enter node Y2 at time 1 and keep it away at later times. Clearly, the value fl (x) results at node Y5 after 2l + 1 steps. D The network used for computing fl requires only linear and second-order units. The following result shows that the established upper bound is asymptotically almost tight, with a gap only of order four. Moreover, the lower bound holds for networks of unrestricted order and with sigmoidal units. Theorem 3. Every general recurrent neural network of size k requires at least time cl l / 4 j k to compute function fl' where c> 0 is some constant. Proof. The idea is to construct higher-order networks Nt of small size that have comparatively large VC dimension. Such a network will consist of linear and product units and hypothetical units that compute functions fJ for certain values of j. We shall derive a lower bound on the VC dimension of these networks. Assuming that the hypothetical units can be replaced by time-bounded general recurrent networks, we determine an upper bound on the VC dimension of the resulting networks in terms of size and computing time using an idea from Koiran and Sontag (1998) and Proposition 1. The comparison of the lower and upper VC dimension bounds will give an estimate of the time required for computing k Network Nt, shown in Figure 3, is a feedforward network composed of three networks • r(1) • r(2) .r(3) E h k • r(/1) 1 2 3 h l· d (/1) (/1) JVI ,JVI ,JVI . ac networ JVI ,J.L = , , , as lnput no es Xl' .. . , x I and 2l + 2 computation nodes yb/1), ... , Y~r~l (see Figure 4). There is only one adjustable parameter in Nt, denoted w, all other weights are fixed. The computation nodes are defined as follows (omitting time parameter t): for J.L = 3, for J.L = 1,2, y~/1) fll'--1 (Y~~)l) for i = 1, ... ,l and J.L = 1,2,3, y}~{ y~/1) . x~/1), for i = 1, .. . ,l and J.L = 1,2,3, (/1) (/1) (/1) c 1 2 3 Y21+l YIH + ... + Y21 lor J.L , , • The nodes Yb/1) can be considered as additional input nodes for N//1), where N;(3) gets this input from w, and N;(/1) from N;(/1+l) for J.L = 1,2. Node Y~r~l is the output node of N;(/1), and node Y~~~l is also the output node of Nt. Thus, the entire network has 3l + 6 nodes that are linear or product units and 3l nodes that compute functions h, fl' or f12. output 8 r------------' ..... L-----------, I I B B t t I x~p)1 ~ t input: w or output of N;(P+1) ----Figure 4: Network N;(p). We show that Ni shatters some set of cardinality [3, in particular, the set S = ({ ei : i = 1, . .. , [})3, where ei E {O, 1}1 is the unit vector with a 1 in position i and ° elsewhere. Every dichotomy of S can be programmed into the network parameter w using the following fact about the logistic function ¢ (see Koiran and Sontag, 1998, Lemma 2): For every binary vector b E {O, l}m, b = b1 .•. bm , there is some real number w E [0,1] such that for i = 1, ... , m E { [0,1/2) (1/2,1] if bi = 0, if bi = 1. Hence, for every dichotomy (So, Sd of S the parameter w can be chosen such that every (ei1' ei2 , ei3) E S satisfies 1/2 if (eillei2,eis) E So, 1/2 if (eillei2,eiJ E S1. Since h +i2 H i 3 .12 (w) = ¢i1 (¢i2'1 (¢i3 .12 (w))), this is the value computed by Ni on input (eill ei2' ei3), where ei" is the input given to network N;(p). (Input ei" selects the function li"'I,,-1 in N;(p).) Hence, S is shattered by Ni, implying that Ni has VC dimension at least [3. Assume now that Ii can be computed by a general recurrent neural network of size at most kj in time tj. Using an idea of Koiran and Sontag (1998), we unfold the network to obtain a feedforward network of size at most kjtj computing fj. Thus we can replace the nodes computing ft, ft, fl2 in Nz by networks of size k1t1, kltl, k12t12, respectively, such that we have a feedforward network '!J consisting of sigmoidal, product, and linear units. Since there are 3l units in Nl computing ft, ft, or fl2 and at most 3l + 6 product and linear units, the size of Nt is at most c1lkl2tl2 for some constant C1 > O. Using that Nt has one adjustable weight, we get from Proposition 1 that its VC dimension is at most c2l2kr2tr2 for some constant C2 > o. On the other hand, since Nz and Nt both shatter S, the VC dimension of Nt is at least l3. Hence, l3 ~ C2l2 kr2 tr2 holds, which implies that tl2 2: cl1/2 / kl2 for some c > 0, and hence tl 2: cl1/4 / kl. D Lemma 2 shows that a single recurrent network is capable of computing every function fl in time O(l). The following consequence of Theorem 3 establishes that this bound cannot be much improved. Corollary 4. Every general recurrent neural network requires at least time 0(ll/4 ) to compute the functions fl. 4 Conclusions and Perspectives We have established bounds on the computing time of analog recurrent neural networks. The result shows that for every network of given size there are functions of arbitrarily high time complexity. This fact does not rely on a bound on the magnitude of weights. We have derived upper and lower bounds that are rather tight- with a polynomial gap of order four- and hold for the computation of a specific family of real-valued functions in one variable. Interestingly, the upper bound is shown using second-order networks without sigmoidal units, whereas the lower bound is valid even for networks with sigmoidal units and arbitrary product units. This indicates that adding these units might decrease the computing time only marginally. The derivation made use of an upper bound on the VC dimension of higher-order sigmoidal networks. This bound is not known to be optimal. Any future improvement will therefore lead to a better lower bound on the computing time. We have focussed on product and sigmoidal units as nonlinear computing elements. However, the construction presented here is generic. Thus, it is possible to derive similar results for radial basis function units, models of spiking neurons, and other unit types that are known to yield networks with bounded VC dimension. The questions whether such results can be obtained for continuous-time networks and for networks operating in the domain of complex numbers, are challenging. A further assumption made here is that the networks compute the functions exactly. By a more detailed analysis and using the fact that the shattering of sets requires the outputs only to lie below or above some threshold, similar results can be obtained for networks that approximate the functions more or less closely and for networks that are subject to noise. Acknowledgment The author gratefully acknowledges funding from the Deutsche Forschungsgemeinschaft (DFG). This work was also supported in part by the ESPRIT Working Group in Neural and Computational Learning II, NeuroCOLT2, No. 27150. References Anthony, M. and Bartlett, P. L. (1999). Neural Network Learning: Theoretical Foundations. Cambridge University Press, Cambridge. Balcazar, J., Gavalda, R., and Siegelmann, H. T. (1997). Computational power of neural networks: A characterization in terms of Kolmogorov complexity. IEEE Transcations on Information Theory, 43: 1175- 1183. Blum, L., Cucker, F., Shub, M., and Smale, S. (1998). Complexity and Real Computation. Springer-Verlag, New York. Carrasco, R. C., Forcada, M. L., Valdes-Munoz, M. A., and Neco, R. P. (2000). Stable encoding of finite state machines in discrete-time recurrent neural nets with sigmoid units. Neural Computation, 12:2129- 2174. Durbin, R. and Rumelhart, D. (1989). Product units: A computationally powerful and biologically plausible extension to backpropagation networks. Neural Computation, 1:133- 142. Gavalda, R. and Siegelmann, H. T. (1999). Discontinuities in recurrent neural networks. Neural Computation, 11:715- 745. Haykin, S. (1999). Neural Networks: A Comprehensive Foundation. Prentice Hall, Upper Saddle River, NJ, second edition. Kilian, J. and Siegelmann, H. T. (1996). The dynamic universality of sigmoidal neural networks. Information and Computation, 128:48- 56. Koiran, P. and Sontag, E. D. (1998). Vapnik-Chervonenkis dimension of recurrent neural networks. Discrete Applied Mathematics, 86:63- 79. Maass, W., NatschUiger, T., and Markram, H. (2001). Real-time computing without stable states: A new framework for neural computation based on perturbations. Preprint. Maass, W. and Orponen, P. (1998). On the effect of analog noise in discrete-time analog computations. Neural Computation, 10:1071- 1095. Maass, W. and Sontag, E. D. (1999). Analog neural nets with Gaussian or other common noise distributions cannot recognize arbitrary regular languages. Neural Computation, 11:771- 782. amlin, C. W. and Giles, C. L. (1996). Constructing deterministic finite-state automata in recurrent neural networks. Journal of the Association for Computing Machinery, 43:937- 972. Schmitt, M. (2002). On the complexity of computing and learning with multiplicative neural networks. Neural Computation, 14. In press. Siegelmann, H. T . (1999). Neural Networks and Analog Computation: Beyond the Turing Limit. Progress in Theoretical Computer Science. Birkhiiuser, Boston. Siegelmann, H. T. and Sontag, E. D. (1995). On the computational power of neural nets. Journal of Computer and System Sciences, 50:132- 150. Sima, J. and Orponen, P. (2001). Exponential transients in continuous-time symmetric Hopfield nets. In Dorffner, G., Bischof, H., and Hornik, K. , editors, Proceedings of the International Conference on Artificial Neural Networks ICANN 2001, volume 2130 of Lecture Notes in Computer Science, pages 806- 813, Springer, Berlin.
|
2001
|
23
|
2,018
|
Playing is believing: The role of beliefs in multi-agent learning Yu-Han Chang Artificial Intelligence Laboratory Massachusetts Institute of Technology Cambridge, Massachusetts 02139 ychang@ai.mit.edu Leslie Pack Kaelbling Artificial Intelligence Laboratory Massachusetts Institute of Technology Cambridge, Massachusetts 02139 lpk@ai.mit.edu Abstract We propose a new classification for multi-agent learning algorithms, with each league of players characterized by both their possible strategies and possible beliefs. Using this classification, we review the optimality of existing algorithms, including the case of interleague play. We propose an incremental improvementto the existing algorithms that seems to achieve average payoffs that are at least the Nash equilibrium payoffs in the longrun against fair opponents. 1 Introduction The topic of learning in multi-agent environments has received increasing attention over the past several years. Game theorists have begun to examine learning models in their study of repeated games, and reinforcement learning researchers have begun to extend their singleagent learning models to the multiple-agent case. As traditional models and methods from these two fields are adapted to tackle the problem of multi-agent learning, the central issue of optimality is worth revisiting. What do we expect a successful learner to do? Matrix games and Nash equilibrium. From the game theory perspective, the repeated game is a generalization of the traditional one-shot game, or matrix game. The matrix game is defined as a reward matrix Ri for each player, Ri : A1 × A2 →R, where Ai is the set of actions available to player i. Purely competitive games are called zero-sum games and must satisfy R1 = −R2. Each player simultaneously chooses to play a particular action ai ∈Ai, or a mixed policy µi = PD(Ai), which is a probability distribution over the possible actions, and receives reward based on the joint action taken. Some common examples of single-shot matrix games are shown in Figure 1. The traditional assumption is that each player has no prior knowledge about the other player. As is standard in the game theory literature, it is thus reasonable to assume that the opponent is fully rational and chooses actions that are in its best interest. In return, we must play a best response to the opponent’s choice of action. A best response function for player i, BRi(µ−i), is defined to be the set of all optimal policies for player i, given that the other players are playing the joint policy µ−i: BRi(µ−i) = {µ∗ i ∈Mi|Ri(µ∗ i , µ−i) ≥Ri(µi, µ−i)∀µi ∈Mi}, where Mi is the set of all possible policies for agent i. If all players are playing best responses to the other players’ strategies, µi ∈BRi(µ−i)∀i, R1 = −1 1 1 −1 R1 = 0 −1 1 1 0 −1 −1 1 0 R1 = 0 3 1 2 R1 = 2 0 3 1 R2 = −R1 R2 = −R1 R2 = 0 1 3 2 R2 = 2 3 0 1 (a) Matching pennies (b) Rock-Paper-Scissors (c) Hawk-Dove (d) Prisoner’s Dilemna Figure 1: Some common examples of single-shot matrix games. then the game is said to be in Nash equilibrium. Once all players are playing by a Nash equilibrium, no single player has an incentive to unilaterally deviate from his equilibrium policy. Any game can be solved for its Nash equilibria using quadratic programming, and a player can choose an optimal strategy in this fashion, given prior knowledge of the game structure. The only problem arises when there are multiple Nash equilibria. If the players do not manage to coordinate on one equilibrium joint policy, then they may all end up worse off. The Hawk-Dove game shown in Figure 1(c) is a good example of this problem. The two Nash equilibria occur at (1,2) and (2,1), but if the players do not coordinate, they may end up playing a joint action (1,1) and receive 0 reward. Stochastic games and reinforcement learning. Despite these problems, there is general agreement that Nash equilibrium is an appropriate solution concept for one-shot games. In contrast, for repeated games there are a range of different perspectives. Repeated games generalize one-shot games by assuming that the players repeat the matrix game over many time periods. Researchers in reinforcement learning view repeated games as a special case of stochastic, or Markov, games. Researchers in game theory, on the other hand, view repeated games as an extension of their theory of one-shot matrix games. The resulting frameworks are similar, but with a key difference in their treatment of game history. Reinforcement learning researchers focus their attention on choosing a single stationary policy µ that will maximize the learner’s expected rewards in all future time periods given that we are in time t, maxµ Eµ hPT τ=t γτ−tRτ(µ) i , where T may be finite or infinite, and µ = PD(A). In the infinite time-horizon case, we often include the discount factor 0 < γ < 1. Littman [1] analyzes this framework for zero-sum games, proving convergence to the Nash equilibrium for his minimax-Q algorithm playing against another minimax-Q agent. Claus and Boutilier [2] examine cooperative games where R1 = R2, and Hu and Wellman [3] focus on general-sum games. These algorithms share the common goal of finding and playing a Nash equilibrium. Littman [4] and Hall and Greenwald [5] further extend this approach to consider variants of Nash equilibrium for which convergence can be guaranteed. Bowling and Veloso [6] and Nagayuki et al. [7] propose to relax the mutual optimality requirement of Nash equilibrium by considering rational agents, which always learn to play a stationary best-response to their opponent’s strategy, even if the opponent is not playing an equilibrium strategy. The motivation is that it allows our agents to act rationally even if the opponent is not acting rationally because of physical or computational limitations. Fictitious play [8] is a similar algorithm from game theory. Game theoretic perspective of repeated games. As alluded to in the previous section, game theorists often take a more general view of optimality in repeated games. The key difference is the treatment of the history of actions taken in the game. Recall that in the Table 1: Summary of multi-agent learning algorithms under our new classification. B0 B1 B∞ H0 minimax-Q, Nash-Q Bully H1 Godfather H∞ Q-learning (Q0), (WoLF-)PHC, fictitious play Q1 multiplicativeweight* * assumes public knowledge of the opponent’s policy at each period stochastic game model, we took µi = PD(Ai). Here we redefine µi : H →PD(Ai), where H = S t Ht and Ht is the set of all possible histories of length t. Histories are observations of joint actions, ht = (ai, a−i, ht−1). Player i’s strategy at time t is then expressed as µi(ht−1). In essence, we are endowing our agent with memory. Moreover, the agent ought to be able to form beliefs about the opponent’s strategy, and these beliefs ought to converge to the opponent’s actual strategy given sufficient learning time. Let βi : H →PD(A−i) be player i’s belief about the opponent’s strategy. Then a learning path is defined to be a sequence of histories, beliefs, and personal strategies. Now we can define the Nash equilibrium of a repeated game in terms of our personal strategy and our beliefs about the opponent. If our prediction about the opponent’s strategy is accurate, then we can choose an appropriate best-response strategy. If this holds for all players in the game, then we are guaranteed to be in Nash equilibrium. Proposition 1.1. A learning path {(ht, µi(ht−1), βi(ht−1))|t = 1, 2, . . .} converges to a Nash equilibrium iff the following two conditions hold: • Optimization: ∀t, µi(ht−1) ∈BRi(βi(ht−1)). We always play a best-response to our prediction of the opponent’s strategy. • Prediction: limt→∞|βi(ht−1) −µ−i(ht−1)| = 0. Over time, our belief about the opponent’s strategy converges to the opponent’s actual strategy. However, Nachbar and Zame [9] shows that this requirement of simultaneous prediction and optimization is impossible to achieve, given certain assumptions about our possible strategies and possible beliefs. We can never design an agent that will learn to both predict the opponent’s future strategy and optimize over those beliefs at the same time. Despite this fact, if we assume some extra knowledge about the opponent, we can design an algorithm that approximates the best-response stationary policy over time against any opponent. In the game theory literature, this concept is often called universal consistency. Fudenburg and Levine [8] and Freund and Schapire [10] independently show that a multiplicativeweight algorithm exhibits universal consistency from the game theory and machine learning perspectives. This give us a strong result, but requires the strong assumption that we know the opponent’s policy at each time period. This is typically not the case. 2 A new classification and a new algorithm We propose a general classification that categorizes algorithms by the cross-product of their possible strategies and their possible beliefs about the opponent’s strategy, H×B. An agent’s possible strategies can be classified based upon the amount of history it has in memory, from H0 to H∞. Given more memory, the agent can formulate more complex policies, since policies are maps from histories to action distributions. H0 agents are memoryless and can only play stationary policies. Agents that can recall the actions from the previous time period are classified as H1 and can execute reactive policies. At the other extreme, H∞agents have unbounded memory and can formulate ever more complex strategies as the game is played over time. An agent’s belief classification mirrors the strategy classification in the obvious way. Agents that believe their opponent is memoryless are classified as B0 players, Bt players believe that the opponent bases its strategy on the previous tperiods of play, and so forth. Although not explicitly stated, most existing algorithms make assumptions and thus hold beliefs about the types of possible opponents in the world. We can think of each Hs × Bt as a different league of players, with players in each league roughly equal to one another in terms of their capabilities. Clearly some leagues contain less capable players than others. We can thus define a fair opponent as an opponent from an equal or lesser league. The idea is that new learning algorithms should ideally be designed to beat any fair opponent. The key role of beliefs. Within each league, we assume that players are fully rational in the sense that they can fully use their available histories to construct their future policy. However, an important observation is that the definition of full rationality depends on their beliefs about the opponent. If we believe that our opponent is a memoryless player, then even if we are an H∞player, our fully rational strategy is to simply model the opponent’s stationary strategy and play our stationary best response. Thus, our belief capacity and our history capacity are inter-related. Without a rich set of possible beliefs about our opponent, we cannot make good use of our available history. Similarly, and perhaps more obviously, without a rich set of historical observations, we cannot hope to model complex opponents. Discussion of current algorithms. Many of the existing algorithms fall within the H∞× B0 league. As discussed in the previous section, the problem with these players is that even though they have full access to the history, their fully rational strategy is stationary due to their limited belief set. A general example of a H∞× B0 player is the policy hill climber (PHC). It maintains a policy and updates the policy based upon its history in an attempt to maximize its rewards. Originally PHC was created for stochastic games, and thus each policy also depends on the current state s. In our repeated games, there is only one state. For agent i, Policy Hill Climbing (PHC) proceeds as follows: 1. Let α and δ be the learning rates. Initialize Q(s, a) ←0, µi(s, a) ← 1 |Ai|∀s ∈S, a ∈Ai. 2. Repeat, a. From state s, select action a according to the mixed policy µi(s) with some exploration. b. Observing reward r and next state s′, update Q(s, a) ←(1 −α)Q(s, a) + α(r + γ max a′ Q(s′, a′)). c. Update µ(s, a) and constrain it to a legal probability distribution: µi(s, a) ←µi(s, a) + δ if a = argmaxa′ Q(s, a′) −δ |Ai|−1 otherwise . The basic idea of PHC is that the Q-values help us to define a gradient upon which we execute hill-climbing. Bowling and Veloso’s WoLF-PHC [6] modifies PHC by adjusting δ depending on whether the agent is “winning” or “losing.” True to their league, PHC players play well against stationary opponents. At the opposite end of the spectrum, Littman and Stone [11] propose algorithms in H0×B∞ and H1 × B∞that are leader strategies in the sense that they choose a fixed strategy and hope that their opponent will “follow” by learning a best response to that fixed strategy. Their “Bully” algorithm chooses a fixed memoryless stationary policy, while “Godfather” has memory of the last time period. Opponents included normal Q-learning and Q1 players, which are similar to Q-learners except that they explicitly learn using one period of memory because they believe that their opponent is also using memory to learn. The interesting result is that “Godfather” is able to achieve non-stationary equilibria against Q1 in the repeated prisoner’s dilemna game, with rewards for both players that are higher than the stationary Nash equilibrium rewards. This demonstrates the power of having belief models. However, because these algorithms do not have access to more than one period of history, they cannot begin to attempt to construct statistical models the opponent. “Godfather” works well because it has a built-in best response to Q1 learners rather than attempting to learn a best response from experience. Finally, Hu and Wellman’s Nash-Q and Littman’s minimax-Q are classified as H0 × B0 players, because even though they attempt to learn the Nash equilibrium through experience, their play is fixed once this equilibrium has been learned. Furthermore, they assume that the opponent also plays a fixed stationary Nash equilibrium, which they hope is the other half of their own equilibrium strategy. These algorithms are summarized in Table 1. A new class of players. As discussed above, most existing algorithms do not form beliefs about the opponent beyond B0. None of these approaches is able to capture the essence of game-playing, which is a world of threats, deceits, and generally out-witting the opponent. We wish to open the door to such possibilities by designing learners that can model the opponent and use that information to achieve better rewards. Ideally we would like to design an algorithm in H∞× B∞that is able to win or come to an equilibrium against any fair opponent. Since this is impossible [9], we start by proposing an algorithm in the league H∞× B∞that plays well against a restricted class of opponents. Since many of the current algorithms are best-response players, we choose an opponent class such as PHC, which is a good example of a best-response player in H∞× B0. We will demonstrate that our algorithm indeed beats its PHC opponents and in fact does well against most of the existing fair opponents. A new algorithm: PHC-Exploiter. Our algorithm is different from most previous work in that we are explicitly modeling the opponent’s learning algorithm and not simply his current policy. In particular, we would like to model players from H∞× B0. Since we are in H∞× B∞, it is rational for us to construct such models because we believe that the opponent is learning and adapting to us over time using its history. The idea is that we will “fool” our opponent into thinking that we are stupid by playing a decoy policy for a number of time periods and then switch to a different policy that takes advantage of their best response to our decoy policy. From a learning perspective, the idea is that we adapt much faster than the opponent; in fact, when we switch away from our decoy policy, our adjustment to the new policy is immediate. In contrast, the H∞× B0 opponent adjusts its policy by small increments and is furthermore unable to model our changing behavior. We can repeat this “bluff and bash” cycle ad infinitum, thereby achieving infinite total rewards as t →∞. The opponent never catches on to us because it believes that we only play stationary policies. A good example of a H∞× B0 player is PHC. Bowling and Veloso showed that in selfplay, a restricted version of WoLF-PHC always reaches a stationary Nash equilibrium in two-player two-action games, and that the general WoLF-PHC seems to do the same in experimental trials. Thus, in the long run, a WoLF-PHC player achieves its stationary Nash equilibrium payoff against any other PHC player. We wish to do better than that by exploiting our knowledge of the PHC opponent’s learning strategy. We can construct a PHC-Exploiter algorithm for agent i that proceeds like PHC in steps 1-2b, and then continues as follows: c. Observing action at −i at time t, update our history h and calculate an estimate of the opponent’s policy: ˆµt −i(s, a) = Pt τ=t−w #(h[τ] = a) w ∀a, where w is the window of estimation and #(h[τ] = a) = 1 if the opponent’s action at time τ is equal to a, and 0 otherwise. We estimate ˆµt−w −i (s) similarly. d. Update δ by estimating the learning rate of the PHC opponent: δ ← ˆµt −i(s) −ˆµt−w −i (s) w . e. Update µi(s, a). If we are winning, i.e. P a′ µi(s, a′)Q(s, a′) > Ri(ˆµ∗ i (s), ˆµ−i(s)), then update µi(s, a) ← 1 if a = argmaxa′ Q(s, a′) 0 otherwise , otherwise we are losing, then update µi(s, a) ←µi(s, a) + δ if a = argmaxa′ Q(s, a′) −δ |Ai|−1 otherwise . Note that we derive both the opponent’s learning rate δ and the opponent’s policy ˆµ−i(s) from estimates using the observable history of actions. If we assume the game matrix is public information, then we can solve for the equilibrium strategy ˆµ∗ i (s), otherwise we can run WoLF-PHC for some finite number of time periods to obtain an estimate this equilibrium strategy. The main idea of this algorithm is that we take full advantage of all time periods in which we are winning, that is, when P a′ µi(s, a′)Q(s, a′) > Ri(ˆµ∗ i (s), ˆµ−i(s)). Analysis. The PHC-Exploiter algorithm is based upon PHC and thus exhibits the same behavior as PHC in games with a single pure Nash equilibrium. Both agents generally converge to the single pure equilibrium point. The interesting case arises in competitive games where the only equilibria require mixed strategies, as discussed by Singh et al [12] and Bowling and Veloso [6]. Matching pennies, shown in Figure 1(a), is one such game. PHC-Exploiter is able to use its model of the opponent’s learning algorithm to choose better actions. In the full knowledge case where we know our opponent’s policy µ2 and learning rate δ2 at every time period, we can prove that a PHC-Exploiter learning algorithm will guarantee us unbounded reward in the long run playing games such as matching pennies. Proposition 2.1. In the zero-sum game of matching pennies, where the only Nash equilibrium requires the use of mixed strategies, PHC-Exploiter is able to achieve unbounded rewards as t →∞against any PHC opponent given that play follows the cycle C defined by the arrowed segments shown in Figure 2. Play proceeds along Cw, Cl, then jumps from (0.5, 0) to (1,0), follows the line segments to (0.5, 1), then jumps back to (0, 1). Given a point (x, y) = (µ1(H), µ2(H)) on the graph in Figure 2, where µi(H) is the probability by which player i plays Heads, we know that our expected reward is R1(x, y) = −1 × [(x)(y) + (1 −x)(1 −y)] + 1 × [(1 −x)(y) + (x)(1 −y)]. -0.5 0 0.5 1 1.5 -0.5 0 0.5 1 1.5 Player 2 probability choosing Heads Player 1 probability choosing Heads Action distribution of the two agent system Cl Cw -0.5 0 0.5 1 1.5 -0.5 0 0.5 1 1.5 Player 2 probability choosing Heads Player 1 probability choosing Heads Action distribution of the two agent system agent1 winning agent1 losing Figure 2: Theoretical (left), Empirical (right). The cyclic play is evident in our empirical results, where we play a PHC-Exploiter player 1 against a PHC player 2. -4000 -2000 0 2000 4000 6000 8000 0 20000 40000 60000 80000 100000 total reward time period Agent 1 total reward over time Figure 3: Total rewards for agent 1 increase as we gain reward through each cycle. We wish to show that Z C R1(x, y)dt = 2 × Z Cw R1(x, y)dt + Z Cl R1(x, y)dt > 0 . We consider each part separately. In the losing section, we let g(t) = x = t and h(t) = y = 1/2 −t, where 0 ≤t ≤1/2. Then Z Cl R1(x, y)dt = Z 1/2 0 R1(g(t), h(t))dt = −1 12 . Similarly, we can show that we receive 1/4 reward over Cw. Thus, R C R1(x, y)dt = 1/3 > 0, and we have shown that we receive a payoff greater than the Nash equilibrium payoff of zero over every cycle. It is easy to see that play will indeed follow the cycle C to a good approximation, depending on the size of δ2. In the next section, we demonstrate that we can estimate µ2 and δ2 sufficiently well from past observations, thus eliminating the full knowledge requirements that were used to ensure the cyclic nature of play above. Experimental results. We used the PHC-Exploiter algorithm described above to play against several PHC variants in different iterated matrix games, including matching pennies, prisoner’s dilemna, and rock-paper-scissors. Here we give the results for the matching pennies game analyzed above, playing against WoLF-PHC. We used a window of w = 5000 time periods to estimate the opponent’s current policy µ2 and the opponent’s learning rate δ2. As shown in Figure 2, the play exhibits the cyclic nature that we predicted. The two solid vertical lines indicate periods in which our PHC-Exploiter player is winning, and the dashed, roughly diagonal, lines indicate periods in which it is losing. In the analysis given in the previous section, we derived an upper bound for our total rewards over time, which was 1/6 for each time step. Since we have to estimate various parameters in our experimental run, we do not achieve this level of reward. We gain an average of 0.08 total reward for each time period. Figure 3 plots the total reward for our PHC-Exploiter agent over time. The periods of winning and losing are very clear from this graph. Further experiments tested the effectiveness of PHC-Exploiter against other fair opponents, including itself. Against all the existing fair opponents shown in Table 1, it achieved at least its average equilibrium payoff in the long-run. Not surprisingly, it also posted this score when it played against a multiplicative-weight learner. Conclusion and future work. In this paper, we have presented a new classification for multi-agent learning algorithms and suggested an algorithm that seems to dominate existing algorithms from the fair opponent leagues when playing certain games. Ideally, we would like to create an algorithm in the league H∞× B∞that provably dominates larger classes of fair opponents in any game. Moreover, all of the discussion contained within this paper dealt with the case of iterated matrix games. We would like to extend our framework to more general stochastic games with multiple states and multiple players. Finally, it would be interesting to find practical applications of these multi-agent learning algorithms. Acknowledgements. This work was supported in part by a Graduate Research Fellowship from the National Science Foundation. References [1] Michael L. Littman. Markov games as a framework for multi-agent reinforcement learning. In Proceedings of the 11th International Conference on Machine Learning (ICML-94), 1994. [2] Caroline Claus and Craig Boutilier. The dynamics of reinforcement learning in cooperative multiaent systems. In Proceedings of the 15th Natl. Conf. on Artificial Intelligence, 1998. [3] Junling Hu and Michael P. Wellman. Multiagent reinforcement learning: Theoretical framework and an algorithm. In Proceedings of the 15th Int. Conf. on Machine Learning (ICML-98), 1998. [4] Michael L. Littman. Friend-or-foe q-learning in general-sum games. In Proceedings of the 18th Int. Conf. on Machine Learning (ICML-01), 2001. [5] Keith Hall and Amy Greenwald. Correlated q-learning. In DIMACS Workshop on Computational Issues in Game Theory and Mechanism Design, 2001. [6] Michael Bowling and Manuela Veloso. Multiagent learning using a variable learning rate. Under submission. [7] Yasuo Nagayuki, Shin Ishii, and Kenji Doya. Multi-agent reinforcement learning: An approach based on the other agent’s internal model. In Proceedings of the International Conference on Multi-Agent Systems (ICMAS-00), 2000. [8] Drew Fudenburg and David K. Levine. Consistency and cautious fictitious play. Journal of Economic Dynamics and Control, 19:1065–1089, 1995. [9] J.H. Nachbar and W.R. Zame. Non-computable strategies and discounted repeated games. Economic Theory, 1996. [10] Yoav Freund and Robert E. Schapire. Adaptive game playing using multiplicative weights. Games and Economic Behavior, 29:79–103, 1999. [11] Michael Littman and Peter Stone. Leading best-response stratgies in repeated games. In 17th Int. Joint Conf. on Artificial Intelligence (IJCAI-2001) workshop on Economic Agents, Models, and Mechanisms, 2001. [12] S. Singh, M. Kearns, and Y. Mansour. Nash convergence of gradient dynamics in general-sum games. In Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence, 2000.
|
2001
|
24
|
2,019
|
Probabilistic principles in unsupervised learning of visual structure: human data and a model Shimon Edelman, Benjamin P. Hiles & Hwajin Yang Department of Psychology Cornell University, Ithaca, NY 14853 se37,bph7,hy56 @cornell.edu Nathan Intrator Institute for Brain and Neural Systems Box 1843, Brown University Providence, RI 02912 Nathan Intrator@brown.edu Abstract To find out how the representations of structured visual objects depend on the co-occurrence statistics of their constituents, we exposed subjects to a set of composite images with tight control exerted over (1) the conditional probabilities of the constituent fragments, and (2) the value of Barlow’s criterion of “suspicious coincidence” (the ratio of joint probability to the product of marginals). We then compared the part verification response times for various probe/target combinations before and after the exposure. For composite probes, the speedup was much larger for targets that contained pairs of fragments perfectly predictive of each other, compared to those that did not. This effect was modulated by the significance of their co-occurrence as estimated by Barlow’s criterion. For lone-fragment probes, the speedup in all conditions was generally lower than for composites. These results shed light on the brain’s strategies for unsupervised acquisition of structural information in vision. 1 Motivation How does the human visual system decide for which objects it should maintain distinct and persistent internal representations of the kind typically postulated by theories of object recognition? Consider, for example, the image shown in Figure 1, left. This image can be represented as a monolithic hieroglyph, a pair of Chinese characters (which we shall refer to as and ), a set of strokes, or, trivially, as a collection of pixels. Note that the second option is only available to a system previously exposed to various combinations of Chinese characters. Indeed, a principled decision whether to represent this image as , or otherwise can only be made on the basis of prior exposure to related images. According to Barlow’s [1] insight, one useful principle is tallying suspicious coincidences: two candidate fragments and should be combined into a composite object if the probability of their joint appearance
is much higher than , which is the probability expected in the case of their statistical independence. This criterion may be compared to the Minimum Description Length (MDL) principle, which has been previously discussed in the context of object representation [2, 3]. In a simplified form [4], MDL calls for representing explicitly as a whole if , just as the principle of suspicious coincidences does. While the Barlow/MDL criterion certainly indicates a suspicious coincidence, there are additional probabilistic considerations that may be used in setting the degree of association between and . One example is the possible perfect predictability of from and vice versa, as measured by
. If , then and are perfectly predictive of each other and should really be coded by a single symbol, whereas the MDL criterion may suggest merely that some association between the representation of and that of be established. In comparison, if and are not perfectly predictive of each other ( ), there is a case to be made in favor of coding them separately to allow for a maximally expressive representation, whereas MDL may actually suggest a high degree of association (if
). In this study we investigated whether the human visual system uses a criterion based on alongside MDL while learning (in an unsupervised manner) to represent composite objects. AB Figure 1: Left: how many objects are contained in image ? Without prior knowledge, a reasonable answer, which embodies a holistic bias, should be “one” (Gestalt effects, which would suggest two convex “blobs” [5], are beyond the scope of the present discussion). Right: in this set of ten images, appears five times as a whole; the other five times a fragment wholly contained in appears in isolation. This statistical fact provides grounds for considering to be composite, consisting of two fragments (call the upper one and the lower one ), because , but . To date, psychophysical explorations of the sensitivity of human subjects to stimulus statistics tended to concentrate on means (and sometimes variances) of the frequency of various stimuli (e.g., [6]. One recent and notable exception is the work of Saffran et al. [7], who showed that infants (and adults) can distinguish between “words” (stable pairs of syllables that recur in a continuous auditory stimulus stream) and non-words (syllables accidentally paired with each other, the first of which comes from one “word” and the second – from the following one). Thus, subjects can sense (and act upon) differences in transition probabilities between successive auditory stimuli. This finding has been recently replicated, with infants as young as 2 months, in the visual sequence domain, using successive presentation of simple geometric shapes with controlled transition probabilities [8]. Also in the visual domain, Fiser and Aslin [9] presented subjects with geometrical shapes in various spatial configurations, and found effects of conditional probabilities of shape co-occurrences, in a task that required the subjects to decide in each trial which of two simultaneously presented shapes was more familiar. The present study was undertaken to investigate the relevance of the various notions of statistical independence to the unsupervised learning of complex visual stimuli by human subjects. Our experimental approach differs from that of [9] in several respects. First, instead of explicitly judging shape familiarity, our subjects had to verify the presence of a probe shape embedded in a target. This objective task, which produces a pattern of response times, is arguably better suited to the investigation of internal representations involved in object recognition than subjective judgment. Second, the estimation of familiarity requires the subject to access in each trial the representations of all the objects seen in the experiment; in our task, each trial involved just two objects (the probe and the target), potentially sharpening the focus of the experimental approach. Third, our experiments tested the predictions of two distinct notions of stimulus independence: , and MDL, or Barlow’s ratio. 2 The psychophysical experiments In two experiments, we presented stimuli composed of characters such as those in Figure 1 to nearly 100 subjects unfamiliar with the Chinese script. The conditional probabilities of the appearance of individual characters were controlled. The experiments involved two types of probe conditions: PTYPE=Fragment, or (with as the reference condition), and PTYPE=Composite, or (with as reference). In this notation (see Figure 2, left), and are “familiar” fragments with controlled minimum conditional probability , and are novel (low-probability) fragments. Each of the two experiments consisted of a baseline phase, followed by training exposure (unsupervised learning), followed in turn by the test phase (Figure 2, right). In the baseline and test phases, the subjects had to indicate whether or not the probe was contained in the target (a task previously used by Palmer [5]). In the intervening training phase, the subjects merely watched the character triplets presented on the screen; to ensure their attention, the subjects were asked to note the order in which the characters appeared. probe target reference test Fragment Composite probe target V ABZ VW ABZ ABZ ABZ A AB probe target mask 1 2 3 4 baseline/test unsupervised training Figure 2: Left: illustration of the probe and target composition for the two levels of PTYPE (Fragment and Composite). For convenience, the various categories of characters that appeared in the experiment are annotated here by Latin letters: , stand for characters with controlled
, and stand for characters that appeared only once throughout an experiment. In experiment 1, the training set was constructed with for some pairs, and for others; in experiment 2, Barlow’s suspicious coincidence ratio was also controlled. Right top: the structure of a part verification trial (same for baseline and test phases). The probe stimulus was followed by the target (each presented for ; a mask was shown before and after the target). The subject had to indicate whether or not the former was contained in the latter (in this example, the correct answer is yes). A sequence consisting of 64 trials like this one was presented twice: before training (baseline phase) and after training (test phase). For “positive” trials (i.e., probe contained in target), we looked at the SPEEDUP following training, defined as
; negative trials were discarded. Right bottom: the structure of a training trial (the training phase, placed between baseline and test, consisted of 80 such trials). The three components of the stimulus appeared one by one for to make sure that the subject attended to each, then together for ! . The subject was required to note whether the sequence unfolded in a clockwise or counterclockwise order. The logic behind the psychophysical experiments rested on two premises. First, we knew from earlier work [5] that a probe is detected faster if it is represented monolithically (that is, considered to be a good “object” in the Gestalt sense). Second, we hypothesized that a composite stimulus would be treated as a monolithic object to the extent that its constituent characters are predictable from each other, as measured by a high conditional probability, , and/or by a high suspicious coincidence ratio, . The main prediction following from these premises is that the SPEEDUP (the difference in response time between baseline and test phases) for a composite probe should reflect the mutual predictability of the probe’s constituents in the training set. Thus, our hypothesis — that statistics of co-occurrence determine the constituents in terms of which structured objects are represented — would be supported if the SPEEDUP turns out to be larger for those composite probes whose constituents tend to appear together in the training set. The experiments, therefore, hinged on a comparison of the patterns of response times in the “positive” trials (in which the probe actually is embedded in the target; see Figure 2, left) before and after exposure to the training set. 0.4 0.6 0.8 1 0 100 200 300 400 speedup, ms minCP Composite Fragment 0.4 0.6 0.8 1 −0.2 −0.1 0 0.1 0.2 0.3 analog of speedup minCP Composite Fragment Figure 3: Left: unsupervised learning of statistically defined structure by human subjects, experiment 1 ( ). The dependent variable SPEED-UP is defined as the difference in
between baseline and test phases (least-squares estimates of means and standard errors, computed by the LSMEANS option of SAS procedure MIXED [10]). The SPEED-UP for composite probes (solid line) with exceeded that in the other conditions by about . Right: the results of a simulation of experiment 1 by a model derived from the one described in [4]. The model was exposed to the same 80 training images as the human subjects. The difference of reconstruction errors for probe and target served as the analog of RT; baseline measurements were conducted on half-trained networks. 2.1 Experiment 1 Fourteen subjects, none of them familiar with the Chinese writing system, participated in this experiment in exchange for course credit. Among the stimuli, two characters could be paired, in which case we had . Alternatively, could be unpaired, with , (in this experiment, we held the suspicious coincidence ratio
constant at ). For the paired the minimum conditional probability and the two characters were perfectly predictable from each other, whereas for the unpaired , and they were not. In the latter case probably should not be represented as a whole. As expected, we found the value of SPEED-UP to be strikingly different for composite probes with ( ) compared to the other three conditions (about ); see Figure 3, left. A mixed-effects repeated measures analysis of variance (SAS procedure MIXED [10]) for SPEED-UP revealed a marginal effect of PTYPE ( ! ) and a significant interaction PTYPE interaction ( ! ). This behavior conforms to the predictions of the principle: SPEEDUP was generally higher for composite probes, and disproportionately higher for composite probes with . The subjects in experiment 1 proved to be sensitive to the measure of independence in learning to associate object fragments together. Note that the suspicious coincidence ratio was the same in both cases,
. Thus, the visual system is sensitive to over and above the (constant-valued) MDLrelated criterion, according to which the propensity to form a unified representation of two fragments, and , should be determined by [1, 4]. 0.4 0.6 0.8 1 0 50 100 150 200 250 speedup, ms minCP r=1.13 0.4 0.6 0.8 1 0 50 100 150 200 250 speedup, ms minCP r=8.33 0 5 10 0 50 100 150 200 250 speedup, ms r minCP=0.5 0 5 10 0 50 100 150 200 250 speedup, ms r minCP=1.0 Figure 4: Human subjects, experiment 2 ( ). The effect of found in experiment 1 was modulated in a complicated fashion by the effect of the suspicious coincidence ratio (see text for discussion). 2.2 Experiment 2 In the second experiment, we studied the effects of varying both and together. Because these two quantities are related (through the Bayes theorem), they cannot be manipulated independently. To accommodate this constraint, some subjects saw two sets of stimuli, with and with , in the first session and other two sets, with and with , in the second session; for other subjects, the complementary combinations were used in each session. Eighty one subjects unfamiliar with the Chinese script participated in this experiment for course credit. The results (Figure 4) showed that SPEEDUP was consistently higher for composite probes. Thus, the association between probe constituents was strengthened by training in each of the four conditions. SPEEDUP was also generally higher for the high suspicious coincidence ratio case, , and disproportionately higher for composite probes in the , case, indicating a complicated synergy between the two measures of dependence, and . A mixed-effects repeated measures analysis of variance (SAS procedure MIXED [10]) for SPEED-UP revealed significant main effects of PTYPE ( ! ) and ( ! ), as well as two significant two-way interactions, ( ) and PTYPE ( ! ). There was also a marginal three-way interaction, PTYPE ( ). The findings of these two psychophysical experiments can be summarized as follows: (1) an individual complex visual shape (a Chinese character) is detected faster than a composite stimulus (a pair of such characters) when embedded in a 3-character scene, but this advantage is narrowed with practice; (2) a composite attains an “objecthood” status to the extent that its constituents are predictable from each other, as measured either by the conditional probability, , or by the suspicious coincidence ratio, ; (3) for composites, the strongest boost towards objecthood (measured by response speedup following unsupervised learning) is obtained when is high and is low, or vice versa. The nature of this latter interaction is unclear, and needs further study. 3 An unsupervised learning model and a simulated experiment The ability of our subjects to construct representations that reflect the probability of cooccurrence of complex shapes has been replicated by a pilot version of an unsupervised learning model, derived from the work of [4]. The model (Figure 5) is based on the following observation: an auto-association network fed with a sequence of composite images in which some fragment/location combinations are more likely than others develops a nonuniform spatial distribution of reconstruction errors. Specifically, smaller errors appear in those locations where the image fragments recur. This information can be used to form a spatial receptive field for the learning module, while the reconstruction error can signal its relevance to the current input [11, 12]. In the simplified pilot model, the spatial receptive field (labeled in Figure 5, left, as “relevance mask”) consists of four weights, one per quadrant: , . During the unsupervised training, the weights are updated by setting
, where is the reconstruction error in trial , and and are learning constants. In a simulation of experiment 1, a separate module with its own four-weight “receptive field” was trained for each of the composite stimuli shown to the human subjects.1 The Euclidean distance between probe and target representations at the output of the model served as the analog of response time, allowing us to compare the model’s performance with that of the humans. We found the same differential effects of for Fragment and Composite probes in the real and simulated experiments; compare Figure 3, left (humans) with Figure 3, right (model). 1The full-fledged model, currently under development, will have a more flexible receptive field structure, and will incorporate competitive learning among the modules. relevance mask (RF) error auto− associator adapt input input − reconstructed ensemble of modules erri Figure 5: Left: the functional architecture of a fragment module. The module consists of two adaptive components: a reconstruction network, and a relevance mask, which assigns different weights to different input pixels. The mask modulates the input multiplicatively, determining the module’s receptive field. Given a sequence of images, several such modules working in parallel learn to represent different categories of spatially localized patterns (fragments) that recur in those images. The reconstruction error serves as an estimate of the module’s ability to deal with the input ([11, 12]; in the error image, shown on the right, white corresponds to high values). Right: the Chorus of Fragments (CoF) is a bank of such fragment modules, each tuned to a particular shape category, appearing in a particular location [13, 4]. 4 Discussion Human subjects have been previously shown to be able to acquire, through unsupervised learning, sensitivity to transition probabilities between syllables of nonsense words [7] and between digits [14], and to co-occurrence statistics of simple geometrical figures [9]. Our results demonstrate that subjects can also learn (presumably without awareness; cf. [14]) to treat combinations of complex visual patterns differentially, depending on the conditional probabilities of the various combinations, accumulated during a short unsupervised training session. In our first experiment, the criterion of suspicious coincidence between the occurrences of and was met in both and conditions: in each case, we had . Yet, the subjects’ behavior indicated a significant holistic bias: the representation they form tends to be monolithic ( ), unless imperfect mutual predictability of the potential fragments ( and ) provides support for representing them separately. We note that a similar holistic bias, operating in a setting where a single encounter with a stimulus can make a difference, is found in language acquisition: an infant faced with an unfamiliar word will assume it refers to the entire shape of the most salient object [15]. In our second experiment, both the conditional probabilities as such, and the suspicious coincidence ratio were found to have the predicted effects, yet these two factors interacted in a complicated manner, which requires a further investigation. Our current research focuses on (1) the elucidation of the manner in which subjects process statistically structured data, (2) the development of the model of structure learning outlined in the preceding section, and (3) an exploration of the implications of this body of work for wider issues in vision, such as the computational phenomenology of scene perception [16]. References [1] H. B. Barlow. Unsupervised learning. Neural Computation, 1:295–311, 1989. [2] R. S. Zemel and G. E. Hinton. Developing population codes by minimizing description length. Neural Computation, 7:549–564, 1995. [3] E. Bienenstock, S. Geman, and D. Potter. Compositionality, MDL priors, and object recognition. In M. C. Mozer, M. I. Jordan, and T. Petsche, editors, Neural Information Processing Systems, volume 9. MIT Press, 1997. [4] S. Edelman and N. Intrator. A productive, systematic framework for the representation of visual structure. In T. K. Leen, T. G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems 13, pages 10–16. MIT Press, 2001. [5] S. E. Palmer. Hierarchical structure in perceptual representation. Cognitive Psychology, 9:441–474, 1977. [6] M. J. Flannagan, L. S. Fried, and K. J. Holyoak. Distributional expectations and the induction of category structure. Journal of Experimental Psychology: Learning, Memory and Cognition, 12:241–256, 1986. [7] J. R. Saffran, R. N. Aslin, and E. L. Newport. Statistical learning by 8-month-old infants. Science, 274:1926–1928, 1996. [8] N. Z. Kirkham, J. A. Slemmer, and S. P. Johnson. Visual statistical learning in infancy: Evidence for a domain general learning mechanism. Cognition, -:–, 2002. in press. [9] J. Fiser and R. N. Aslin. Unsupervised statistical learning of higher-order spatial structures from visual scenes. Psychological Science, 6:499–504, 2001. [10] SAS. User’s Guide, Version 8. SAS Institute Inc., Cary, NC, 1999. [11] D. Pomerleau. Input reconstruction reliability estimation. In C. L. Giles, S. J. Hanson, and J. D. Cowan, editors, Advances in Neural Information Processing Systems, volume 5, pages 279–286. Morgan Kaufmann Publishers, 1993. [12] I. Stainvas and N. Intrator. Blurred face recognition via a hybrid network architecture. In Proc. ICPR, volume 2, pages 809–812, 2000. [13] S. Edelman and N. Intrator. (Coarse Coding of Shape Fragments) + (Retinotopy) Representation of Structure. Spatial Vision, 13:255–264, 2000. [14] G. S. Berns, J. D. Cohen, and M. A. Mintun. Brain regions responsive to novelty in the absence of awareness. Science, 276:1272–1276, 1997. [15] B. Landau, L. B. Smith, and S. Jones. The importance of shape in early lexical learning. Cognitive Development, 3:299–321, 1988. [16] S. Edelman. Constraints on the nature of the neural representation of the visual world. Trends in Cognitive Sciences, 6:–, 2002. in press.
|
2001
|
25
|
2,020
|
Predictive Representations of State Michael L. Littman Richard S. Sutton AT&T Labs-Research, Florham Park, New Jersey {mlittman,sutton}~research.att.com Satinder Singh Syntek Capital, New York, New York baveja~cs.colorado.edu Abstract We show that states of a dynamical system can be usefully represented by multi-step, action-conditional predictions of future observations. State representations that are grounded in data in this way may be easier to learn, generalize better, and be less dependent on accurate prior models than, for example, POMDP state representations. Building on prior work by Jaeger and by Rivest and Schapire, in this paper we compare and contrast a linear specialization of the predictive approach with the state representations used in POMDPs and in k-order Markov models. Ours is the first specific formulation of the predictive idea that includes both stochasticity and actions (controls). We show that any system has a linear predictive state representation with number of predictions no greater than the number of states in its minimal POMDP model. In predicting or controlling a sequence of observations, the concepts of state and state estimation inevitably arise. There have been two dominant approaches. The generative-model approach, typified by research on partially observable Markov decision processes (POMDPs), hypothesizes a structure for generating observations and estimates its state and state dynamics. The history-based approach, typified by k-order Markov methods, uses simple functions of past observations as state, that is, as the immediate basis for prediction and control. (The data flow in these two approaches are diagrammed in Figure 1.) Of the two, the generative-model approach is more general. The model's internal state gives it temporally unlimited memorythe ability to remember an event that happened arbitrarily long ago--whereas a history-based approach can only remember as far back as its history extends. The bane of generative-model approaches is that they are often strongly dependent on a good model of the system's dynamics. Most uses of POMDPs, for example, assume a perfect dynamics model and attempt only to estimate state. There are algorithms for simultaneously estimating state and dynamics (e.g., Chrisman, 1992), analogous to the Baum-Welch algorithm for the uncontrolled case (Baum et al., 1970), but these are only effective at tuning parameters that are already approximately correct (e.g., Shatkay & Kaelbling, 1997). observations (and actions) (a) state 1-----1-----1..rep'n observations¢E (and actions) / state t/' rep'n 1-step --+ . delays (b) Figure 1: Data flow in a) POMDP and other recursive updating of state representation, and b) history-based state representation. In practice, history-based approaches are often much more effective. Here, the state representation is a relatively simple record of the stream of past actions and observations. It might record the occurrence of a specific subsequence or that one event has occurred more recently than another. Such representations are far more closely linked to the data than are POMDP representations. One way of saying this is that POMDP learning algorithms encounter many local minima and saddle points because all their states are equipotential. History-based systems immediately break symmetry, and their direct learning procedure makes them comparably simple. McCallum (1995) has shown in a number of examples that sophisticated history-based methods can be effective in large problems, and are often more practical than POMDP methods even in small ones. The predictive state representation (PSR) approach, which we develop in this paper, is like the generative-model approach in that it updates the state representation recursively, as in Figure l(a), rather than directly computing it from data. We show that this enables it to attain generality and compactness at least equal to that of the generative-model approach. However, the PSR approach is also like the history-based approach in that its representations are grounded in data. Whereas a history-based representation looks to the past and records what did happen, a PSR looks to the future and represents what will happen. In particular, a PSR is a vector of predictions for a specially selected set of action-observation sequences, called tests (after Rivest & Schapire, 1994). For example, consider the test U101U202, where U1 and U2 are specific actions and 01 and 02 are specific observations. The correct prediction for this test given the data stream up to time k is the probability of its observations occurring (in order) given that its actions are taken (in order) (i.e., Pr {Ok = 01, Ok+1 = 02 I A k = u1,Ak+ 1 = U2}). Each test is a kind of experiment that could be performed to tell us something about the system. If we knew the outcome of all possible tests, then we would know everything there is to know about the system. A PSR is a set of tests that is sufficient information to determine the prediction for all possible tests (a sufficient statistic). As an example of these points, consider the float/reset problem (Figure 2) consisting of a linear string of 5 states with a distinguished reset state on the far right. One action, f (float), causes the system to move uniformly at random to the right or left by one state, bounded at the two ends. The other action, r (reset), causes a jump to the reset state irrespective of the current state. The observation is always ounless the r action is taken when the system is already in the reset state, in which case the observation is 1. Thus, on an f action, the correct prediction is always 0, whereas on an r action, the correct prediction depends on how many fs there have been since the last r: for zero fS, it is 1; for one or two fS, it is 0.5; for three or four fS, it is 0.375; for five or six fs, it is 0.3125, and so on decreasing after every second f, asymptotically bottoming out at 0.2. No k-order Markov method can model this system exactly, because no limited-. .5 a) float action .5 b) reset action 1,0=1 Figure 2: Underlying dynamics of the float/reset problem for a) the float action and b) the reset action. The numbers on the arcs indicate transition probabilities. The observation is always 0 except on the reset action from the rightmost state, which produces an observation of 1. length history is a sufficient statistic. A POMDP approach can model it exactly by maintaining a belief-state representation over five or so states. A PSR, on the other hand, can exactly model the float/reset system using just two tests: rl and fOrI. Starting from the rightmost state, the correct predictions for these two tests are always two successive probabilities in the sequence given above (1, 0.5, 0.5, 0.375,...), which is always a sufficient statistic to predict the next pair in the sequence. Although this informational analysis indicates a solution is possible in principle, it would require a nonlinear updating process for the PSR. In this paper we restrict consideration to a linear special case of PSRs, for which we can guarantee that the number of tests needed does not exceed the number of states in the minimal POMDP representation (although we have not ruled out the possibility it can be considerably smaller). Of greater ultimate interest are the prospects for learning PSRs and their update functions, about which we can only speculate at this time. The difficulty of learning POMDP structures without good prior models are well known. To the extent that this difficulty is due to the indirect link between the POMDP states and the data, predictive representations may be able to do better. Jaeger (2000) introduced the idea of predictive representations as an alternative to belief states in hidden Markov models and provided a learning procedure for these models. We build on his work by treating the control case (with actions), which he did not significantly analyze. We have also been strongly influenced by the work of Rivest and Schapire (1994), who did consider tests including actions, but treated only the deterministic case, which is significantly different. They also explored construction and learning algorithms for discovering system structure. 1 Predictive State Representations We consider dynamical systems that accept actions from a discrete set A and generate observations from a discrete set O. We consider only predicting the system, not controlling it, so we do not designate an explicit reward observation. We refer to such a system as an environment. We use the term history to denote a test forming an initial stream of experience and characterize an environment by a probability distribution over all possible histories, P : {OIA}* H- [0,1], where P(Ol··· Otla1··· at) is the probability of observations 01, ... , O£ being generated, in that order, given that actions aI, ... ,at are taken, in that order. The probability of a test t conditional on a history h is defined as P(tlh) = P(ht)/P(h). Given a set of q tests Q = {til, we define their (1 x q) prediction vector, p(h) = [P(t1Ih),P(t2Ih), ... ,P(tqlh)], as a predictive state representation (PSR) if and only if it forms a sufficient statistic for the environment, Le., if and only if P(tlh) = ft(P(h)), (1) for any test t and history h, and for some projection junction ft : [0, l]q ~ [0,1]. In this paper we focus on linear PSRs, for which the projection functions are linear, that is, for which there exist a (1 x q) projection vector mt, for every test t, such that P(tlh) == ft(P(h)) =7 p(h)mf, (2) for all histories h. Let Pi(h) denote the ith component of the prediction vector for some PSR. This can be updated recursively, given a new action-observation pair a,o, by (3) .(h ) == P(t.lh ) == P(otilha) == faati(P(h)) == p(h)m'{;ati P2 ao 2 ao P(olha) faa (P(h)) p(h)mro ' where the last step is specific to linear PSRs. We can now state our main result: Theorem 1 For any environment that can be represented by a finite POMDP model, there exists a linear PSR with number of tests no larger than the number of states in the minimal POMDP model. 2 Proof of Theorem 1: Constructing a PSR from a POMDP We prove Theorem 1 by showing that for any POMDP model of the environment, we can construct in polynomial time a linear PSR for that POMDP of lesser or equal complexity that produces the same probability distribution over histories as the POMDP model. We proceed in three steps. First, we review POMDP models and how they assign probabilities to tests. Next, we define an algorithm that takes an n-state POMDP model and produces a set of n or fewer tests, each of length less than or equal to n. Finally, we show that the set of tests constitute a PSR for the POMDP, that is, that there are projection vectors that, together with the tests' predictions, produce the same probability distribution over histories as the POMDP. A POMDP (Lovejoy, 1991; Kaelbling et al., 1998) is defined by a sextuple (8, A, 0, bo,T, 0). Here, 8 is a set of n underlying (hidden) states, A is a discrete set of actions, and 0 is a discrete set of observations. The (1 x n) vector bo is an initial state distribution. The set T consists of (n x n) transition matrices Ta, one for each action a, where Tlj is the probability of a transition from state i to j when action a is chosen. The set 0 consists of diagonal (n x n) observation matrices oa,o, one for each pair of observation 0 and action a, where o~'o is the probability of observation 0 when action a is selected and state i is reached. l The state representation in a POMDP (Figure l(a)) is the belief state-the (1 x n) vector of the state-occupation probabilities given the history h. It can be computed recursively given a new action a and observation 0 by b(h)Taoa,o b(hao) = b(h)Taoa,oe;' where en is the (1 x n)-vector of all Is. Finally, a POMDP defines a probability distribution over tests (and thus histories) by P(Ol ... otlhal ... at) == b(h)Ta1oal,Ol ... Taloa£,Ole~. (4) IThere are many equivalent formulations and the conversion procedure described here can be easily modified to accommodate other POMDP definitions. We now present our algorithm for constructing a PSR for a given POMDP. It uses a function u mapping tests to (1 x n) vectors defined recursively by u(c) == en and u(aot) == (Taoa,ou(t)T)T, where c represents the null test. Conceptually, the components of u(t) are the probabilities of the test t when applied from each underlying state of the POMDP; we call u(t) the outcome vector for test t. We say a test t is linearly independent of a set of tests S if its outcome vector is linearly independent of the set of outcome vectors of the tests in S. Our algorithm search is used and defined as Q -<- search(c, {}) search(t, S): for each a E A, 0 E 0 if aot is linearly independent of S then S -<- search(aot, S U {aot}) return S The algorithm maintains a set of tests and searches for new tests that are linearly independent of those already found. It is a form of depth-first search. The algorithm halts when it checks all the one-step extensions of its tests and finds none that are linearly independent. Because the set of tests Q returned by search have linearly independent outcome vectors, the cardinality of Q is bounded by n, ensuring that the algorithm halts after a polynomial number of iterations. Because each test in Q is formed by a one-step extension to some other test in Q, no test is longer than n action-observation pairs. The check for linear independence can be performed in many ways, including Gaussian elimination, implying that search terminates in polynomial time. By construction, all one-step extensions to the set of tests Q returned by search are linearly dependent on those in Q. We now show that this is true for any test. Lemma 1 The outcome vectors of the tests in Q can be linearly combined to produce the outcome vector for any test. Proof: Let U be the (n x q) matrix formed by concatenating the outcome vectors for all tests in Q. Since, for all combinations of a and 0, the columns of Taoa,ou are linearly dependent on the columns of U, we can write Taoa,ou == UWT for some q x q matrix of weights W. If t is a test that is linearly dependent on Q, then anyone-step extension of t, aot, is linearly dependent on Q. This is because we can write the outcome vector for t as u(t) == (UwT)T for some (1 x q) weight vector w and the outcome vector for aot as u(aot) == (Taoa,ou(t)T)T == (Taoa,oUwT)T == (UWTwT)T. Thus, aot is linearly dependent on Q. Now, note that all one-step tests are linearly dependent on Q by the structure of the search algorithm. Using the previous paragraph as an inductive argument, this implies that all tests are linearly dependent on Q. 0 Returning to the float/reset example POMDP, search begins with by enumerating the 4 extensions to the null test (fO, fl, rO, and rl). Of these, only fa and rO are are linearly independent. Of the extensions of these, fOrO is the only one that is linearly independent of the other two. The remaining two tests added to Q by search are fOfOrO and fOfOfOrO. No extensions of the 5 tests in Q are linearly independent of the 5 tests in Q, so the procedure halts. We now show that the set of tests Q constitute a PSR for the POMDP by constructing projection vectors that, together with the tests' predictions, produce the same probability distribution over histories as the POMDP. For each combination of a and 0, define a q x q matrix Mao == (U+Taoa,ou)T and a 1 x q vector mao == (U+Taoa,oe;;JT, where U is the matrix of outcome vectors defined in the previous section and U+ is its pseudoinverse2 • The ith row of Mao is maoti. The probability distribution on histories implied by these projection vectors is p(h)m~101 alOl p(h)M~ol M~_10l_1 m~Ol b(h)UU+ra1 oa1,01U ... U+Tal-10al-1,Ol-1 UU+Taloal,ole; b(h)Ta10 a1,01 ... ral-l0al-t,ol-lTaloal,Ole~, Le., it is the same as that of the POMDP, as in Equation 4. Here, the last step uses the fact that UU+vT == vT for vT linearly dependent on the columns of U. This holds by construction of U in the previous section. This completes the proof of Theorem 1. Completing the float/reset example, consider the Mf,o matrix found by the process defined in this section. It derives predictions for each test in Qafter taking action f. Most of these are quite simple because the tests are so similar: the new prediction for rO is exactly the old prediction for fOrO, for example. The only non trivial test is fOfOfOrO. Its outcome can be computed from 0.250 p(rOlh) - 0.0625 p(fOrOlh) + 0.750 p(fOfOrOlh). This example illustrates that the projection vectors need not contain only positive entries. 3 Conclusion We have introduced a predictive state representation for dynamical systems that is grounded in actions and observations and shown that, even in its linear form, it is at least as general and compact as POMDPs. In essence, we have established PSRs as a non-inferior alternative to POMDPs, and suggested that they might have important advantages, while leaving demonstration of those advantages to future work. We conclude by summarizing the potential advantages (to be explored in future work): Learnability. The k-order Markov model is similar to PSRs in that it is entirely based on actions and observations. Such models can be learned trivially from data by counting-it is an open question whether something similar can be done with a PSR. Jaeger (2000) showed how to learn such a model in the uncontrolled setting, but the situation is more complex in the multiple action case since outcomes are conditioned on behavior, violating some required independence assumptions. Compactness. We have shown that there exist linear PSRs no more complex that the minimal POMDP for an environment, but in some cases the minimal linear PSR seems to be much smaller. For example, a POMDP extension of factored MDPs explored by Singh and Cohn (1998) would be cross-products of separate POMDPs and have linear PSRs that increase linearly with the number and size of the component POMDPs, whereas their minimal POMDP representation would grow as the size 2If U = A~BT is the singular value decomposition of U, then B:E+AT is the pseudoinverse. The pseudoinverse of the diagonal matrix }J replaces each non-zero element with its reciprocal. of the state space, Le., exponential in the number of component POMDPs. This (apparent) advantage stems from the PSR's combinatorial or factored structure. As a vector of state variables, capable of taking on diverse values, a PSR may be inherently more powerful than the distribution over discrete states (the belief state) of a POMDP. We have already seen that general PSRs can be more compact than POMDPs; they are also capable of efficiently capturing environments in the diversity representation used by Rivest and Schapire (1994), which is known to provide an extremely compact representation for some environments. Generalization. There are reasons to think that state variables that are themselves predictions may be particularly useful in learning to make other predictions. With so many things to predict, we have in effect a set or sequence of learning problems, all due to the same environment. In many such cases the solutions to earlier problems have been shown to provide features that generalize particularly well to subsequent problems (e.g., Baxter, 2000; Thrun & Pratt, 1998). Powerful, extensible representations. PSRs that predict tests could be generalized to predict the outcomes of multi-step options (e.g., Sutton et al., 1999). In this case, particularly, they would constitute a powerful language for representing the state of complex environments. AcknowledgIllents: We thank Peter Dayan, Lawrence Saul, Fernando Pereira and Rob Schapire for many helpful discussions of these and related ideas. References Baum, L. E., Petrie, T., Soules, G., & Weiss, N. (1970). A maximization technique occurring in the statistical analysis of probabilistic functions of Markov chains. Annals of Mathematical Statistics, 41, 164-171. Baxter, J. (2000). A model of inductive bias learning. Journal of Artificial Intelligence Research, 12, 149-198. Chrisman, L. (1992). Reinforcement learning with perceptual aliasing: The perceptual distinctions approach. Proceedings of the Tenth National Conference on Artificial Intelligence (pp. 183-188). San Jose, California: AAAI Press. Jaeger, H. (2000). Observable operator models for discrete stochastic time series. Neural Computation, 12, 1371-1398. Kaelbling, L. P., Littman, M. L., & Cassandra, A. R. (1998). Planning and acting in ' partially observable stochastic domains. Artificial Intelligence, 101, 99-134. Lovejoy, W. S. (1991). A survey of algorithmic methods for partially observable Markov decision processes. Annals of Operations Research, 28, 47-65. McCallum, A. K. (1995). Reinforcement learning with selective perception and hidden state. Doctoral diss.ertation, Department of Computer Science, University of Rochester. Rivest, R. L., & Schapire, R. E. (1994). Diversity-based inference of finite automata. Journal of the ACM, 41, 555-589. Shatkay, H., & Kaelbling, L. P. (1997). Learning topological maps with weak local odometric information~ Proceedings of Fifteenth International Joint Conference on Artificial Intelligence (IJCAI-91) (pp. 920-929). Singh, S., & Cohn, D. (1998). How to dynamically merge Markov decision processes. Advances in Neural and Information Processing Systems 10 (pp. 1057-1063). Sutton, R. S., Precup, D., & Singh, S. (1999). Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning. Artificial Intelligence, 181-211. Thrun, S., & Pratt, L. (Eds.). (1998). Learning to learn. Kluwer Academic Publishers.
|
2001
|
26
|
2,021
|
Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering applications, detecting and understanding differences between two groups of examples can be reduced to a classical problem of training a classifier for labeling new examples while making as few mistakes as possible. In the traditional classification setting, the resulting classifier is rarely analyzed in terms of the properties of the input data captured by the discriminative model. However, such analysis is crucial if we want to understand and visualize the detected differences. We propose an approach to interpretation of the statistical model in the original feature space that allows us to argue about the model in terms of the relevant changes to the input vectors. For each point in the input space, we define a discriminative direction to be the direction that moves the point towards the other class while introducing as little irrelevant change as possible with respect to the classifier function. We derive the discriminative direction for kernel-based classifiers, demonstrate the technique on several examples and briefly discuss its use in the statistical shape analysis, an application that originally motivated this work. 1 Introduction Once a classifier is estimated from the training data, it can be used to label new examples, and in many application domains, such as character recognition, text classification and others, this constitutes the final goal of the learning stage. The statistical learning algorithms are also used in scientific studies to detect and analyze differences between the two classes when the “correct answer” is unknown, and the information we have on the differences is represented implicitly by the training set. Example applications include morphological analysis of anatomical organs (comparing organ shape in patients vs. normal controls), molecular design (identifying complex molecules that satisfy certain requirements), etc. In such applications, interpretation of the resulting classifier in terms of the original feature vectors can provide an insight into the nature of the differences detected by the learning algorithm and is therefore a crucial step in the analysis. Furthermore, we would argue that studying the spatial structure of the data captured by the classification function is important in any application, as it leads to a better understanding of the data and can potentially help in improving the technique. This paper addresses the problem of translating a classifier into a different representation that allows us to visualize and study the differences between the classes. We introduce and derive a so called discriminative direction at every point in the original feature space with respect to a given classifier. Informally speaking, the discriminative direction tells us how to change any input example to make it look more like an example from another class without introducing any irrelevant changes that possibly make it more similar to other examples from the same class. It allows us to characterize differences captured by the classifier and to express them as changes in the original input examples. This paper is organized as follows. We start with a brief background section on kernelbased classification, stating without proof the main facts on kernel-based SVMs necessary for derivation of the discriminative direction. We follow the notation used in [3, 8, 9]. In Section 3, we provide a formal definition of the discriminative direction and explain how it can be estimated from the classification function. We then present some special cases, in which the computation can be simplified significantly due to a particular structure of the kernel. Section 4 demonstrates the discriminative direction for different kernels, followed by an example from the problem of statistical analysis of shape differences that originally motivated this work. 2 Basic Notation Given a training set of l pairs {(xk, yk)}l k=1, where xk ∈Rn are observations and yk ∈{−1, 1} are corresponding labels, and a kernel function K : Rn × Rn 7→R, (with its implied mapping function ΦK : Rn 7→F), the Support Vector Machines (SVMs) algorithm [8] constructs a classifier by implicitly mapping the training data into a higher dimensional space and estimating a linear classifier in that space that maximizes the margin between the classes (Fig. 1a). The normal to the resulting separating hyperplane is a linear combination of the training data: w = X k αkykΦK(xk), (1) where the coefficients αk are computed by solving a constrained quadratic optimization problem. The resulting classifier fK(x) = ⟨x · w⟩+b = X k αkyk ⟨ΦK(x) · ΦK(xk)⟩+b = X k αkykK(x, xk)+b (2) defines a nonlinear separating boundary in the original feature space. 3 Discriminative Direction Equations (1) and (2) imply that the classification function fK(x) is directly proportional to the signed distance from the input point to the separating boundary computed in the higher dimensional space defined by the mapping ΦK. In other words, the function output depends only on the projection of vector ΦK(x) onto w and completely ignores the component of ΦK(x) that is perpendicular to w. This suggests that in order to create a displacement of ΦK(x) that corresponds to the differences between the two classes, one should change the vector’s projection onto w while keeping its perpendicular component the same. In the linear case, we can easily perform this operation, since we have access to the image vectors, ΦK(x) = x. This is similar to visualization techniques typically used in linear generative modeling, where the data variation is captured using PCA, and new samples are generated by changing a single principal component at a time. However, this approach is infeasible in the non-linear case, because we do not have access to the image vectors ΦK(x). Furthermore, the resulting image vector might not even have a source in the original feature space, i.e., there might be no vector in the original space Rn that maps into the resulting vector in the space F. Our solution is to search for the direction around (a) Φ w (b) dx z x d z p e w Φ Figure 1: Kernel-based classification (a) and the discriminative direction (b). the feature vector x in the original space that minimizes the divergence of its image ΦK(x) from the direction of the projection vector w1. We call it a discriminative direction, as it represents the direction that affects the output of the classifier while introducing as little irrelevant change as possible into the input vector. Formally, as we move from x to x + dx in Rn, the image vector in the space F changes by dz = ΦK(x + dx) −ΦK(x) (Fig. 1b). This displacement can be thought of as a vector sum of its projection onto w and its deviation from w: p = ⟨dz · w⟩ ⟨w · w⟩w and e = dz −p = dz −⟨dz · w⟩ ⟨w · w⟩w. (3) The discriminative direction minimizes the divergence component e, leading to the following optimization problem: minimize E(dx) = ∥e∥2 = ⟨dz · dz⟩−⟨dz · w⟩2 ⟨w · w⟩ (4) s.t. ∥dx∥2 = ϵ. (5) Since the cost function depends only on dot products of vectors in the space F, it can be computed using the kernel function K: ⟨w · w⟩ = X k,m αkαmykymK(xk, xm), (6) ⟨dz · w⟩ = ∇fK(x)dx, (7) ⟨dz · dz⟩ = dxT HK(x)dx, (8) where ∇fK(x) is the gradient of the classifier function fK evaluated at x and represented by a row-vector and matrix HK(x) is one of the (equivalent) off-diagonal quarters of the Hessian of K, evaluated at (x, x): HK(x)[i, j] = ∂2K(u, v) ∂ui∂vj (u=x,v=x) . (9) Substituting into Equation (4), we obtain minimize E(dx) = dxT HK(x) −∥w∥−2∇f T K(x)∇fK(x) dx (10) s.t. ∥dx∥2 = ϵ. (11) 1A similar complication arises in kernel-based generative modeling, e.g., kernel PCA [7]. Constructing linear combinations of vectors in the space F leads to a global search in the original space [6, 7]. Since we are interested in the direction that best approximates w, we use infinitesimal analysis that results in a different optimization problem. The solution to this problem is the smallest eigenvector of matrix QK(x) = HK(x) −∥w∥−2∇f T K(x)∇fK(x). (12) Note that in general, the matrix QK(x) and its smallest eigenvector are not the same for different points in the original space and must be estimated separately for every input vector x. Furthermore, each solution defines two opposite directions in the input space, corresponding to the positive and the negative projections onto w. We want to move the input example towards the opposite class and therefore assign the direction of increasing function values to the examples with label −1 and the direction of decreasing function values to the examples with label 1. Obtaining a closed-form solution of this minimization problem could be desired, or even necessary, if the dimensionality of the input space is high and computing the smallest eigenvector is computationally expensive and numerically challenging. In the next section, we demonstrate how a particular form of the matrix HK(x) leads to an analytical solution for a large family of kernel functions2. 3.1 Analytical Solution for Discriminative Direction It is easy to see that if HK(x) is a multiple of the identity matrix, HK(x) = cI, then the smallest eigenvector of the matrix QK(x) is equal to the largest eigenvector of the matrix ∇f T K(x)∇fK(x), namely the gradient of the classifier function ∇f T K(x). We will show in this section that both for the linear kernel and, more surprisingly, for RBF kernels, the matrix HK(x) is of the right form to yield an analytical solution of this form. It is well known that to achieve the fastest change in the value of a function, one should move along its gradient. In the case of the linear and the RBF kernels, the gradient also corresponds to the direction that distinguishes between the two classes while ignoring inter-class variability. Dot product kernels, K(u, v) = k(⟨u · v⟩). For any dot product kernel, ∂2K(u, v) ∂ui∂vj (u=x,v=x) = k′(∥x∥2)δij + k′′(∥x∥2)xixj, (13) and therefore HK(x) = cI for all x if and only if k′′(∥x∥2) ≡0, i.e., when k is a linear function. Thus the linear kernel is the only dot product kernel for which this simplification is relevant. In the linear case, HK(x) = I, and the discriminative direction is defined as dx∗= ∇f T K(x) = w = X αkykxk; E(dx∗) = 0. (14) This is not entirely surprising, as the classifier is a linear function in the original space and we can move precisely along w. Polynomial kernels are a special case of dot product kernels. For polynomials of degree d ≥2, ∂2K(u, v) ∂ui∂vj (u=x,v=x) = d(1 + ∥x∥2)d−1δij + d(d −1)(1 + ∥x∥2)d−2xixj. (15) HK(x) is not necessarily diagonal for all x, and we have to solve the general eigenvector problem to identify the discriminative direction. 2While a very specialized structure of HK(x) in the next section is sufficient for simplifying the solution significantly, it is by no means necessary, and other kernel families might exist for which estimating the discriminative direction does not require solving the full eigenvector problem. Distance kernels, K(u, v) = k(∥u −v∥2). For a distance kernel, ∂2K(u, v) ∂ui∂vj (u=x,v=x) = −2k′(0)δij, (16) and therefore the discriminative direction can be determined analytically: dx∗= ∇f T K(x); E(dx∗) = −2k′(0) −∥w∥−2∥∇f T K(x)∥ 2. (17) The Gaussian kernels are a special case of the distance kernel family, and yield a closed form solution for the discriminative direction: dx∗= −2/γ X k αkyke−∥x−xk∥2 γ (x−xk); E(dx∗) = 2/γ −∥∇f T K(x)∥ 2/∥w∥2. (18) Unlike the linear case, we cannot achieve zero error, and the discriminative direction is only an approximation. The exact solution is unattainable in this case, as it has no corresponding direction in the original space. 3.2 Geometric Interpretation We start by noting that the image vectors ΦK(x)’s do not populate the entire space F, but rather form a manifold of lower dimensionality whose geometry is fully defined by the kernel function K (Fig. 1). We will refer to this manifold as the target manifold in this discussion. We cannot explicitly manipulate elements of the space F, but can only explore the target manifold through search in the original space. We perform the search in the original space by considering all points on an infinitesimally small sphere centered at the original input vector x. In the range space of the mapping function ΦK, the images of points x + dx form an ellipsoid defined by the quadratic form dzT dz = dxT HK(x)dx. For HK(x) ∼I, the ellipsoid becomes a sphere, all dz’s are of the same length, and the minimum of error in the displacement vector dz corresponds to the maximum of the projection of dz onto w. Therefore, the discriminative direction is parallel to the gradient of the classifier function. If HK(x) is of any other form, the length of the displacement vector dz changes as we vary dx, and the minimum of the error in the displacement is not necessarily aligned with the direction that maximizes the projection. As a side note, our sufficient condition, HK(x) ∼I, implies that the target manifold is locally flat, i.e., its Riemannian curvature is zero. Curvature and other properties of target manifolds have been studied extensively for different kernel functions [1, 4]. In particular, one can show that the kernel function implies a metric on the original space. Similarly to the natural gradient [2] that maximizes the change in the function value under an arbitrary metric, we minimize the changes that do not affect the function under the metric implied by the kernel. 3.3 Selecting Inputs Given any input example, we can compute the discriminative direction that represents the differences between the two classes captured by the classifier in the neighborhood of the example. But how should we choose the input examples for which to compute the discriminative direction? We argue that in order to study the differences between the classes, one has to examine the input vectors that are close to the separating boundary, namely, the support vectors. Note that this approach is significantly different from the generative modeling, where a “typical” representative, often constructed by computing the mean of the training data, is used for analysis and visualization. In the discriminative framework, we are more interested in the examples that lie close to the opposite class, as they define the differences between the two classes and the optimal separating boundary. (a) −3 −2 −1 0 1 2 3 4 5 6 7 8 −5 −4 −3 −2 −1 0 1 2 3 4 5 6 (b) −3 −2 −1 0 1 2 3 4 5 6 7 8 −5 −4 −3 −2 −1 0 1 2 3 4 5 6 (c) −3 −2 −1 0 1 2 3 4 5 6 7 8 −5 −4 −3 −2 −1 0 1 2 3 4 5 6 Figure 2: Discriminative direction for linear (a), quadratic (b) and Gaussian RBF (c) classifiers. The background is colored using the values of the classifier function. The black solid line is the separating boundary, the dotted lines indicate the margin corridor. Support vectors are indicated using solid markers. The length of the vectors is proportional to the magnitude of the classifier gradient. Support vectors define a margin corridor whose shape is determined by the kernel type used for training. We can estimate the distance from any support vector to the separating boundary by examining the gradient of the classification function for that vector. Large gradient indicates that the support vector is close to the separating boundary and therefore can provide more information on the spatial structure of the boundary. This provides a natural heuristic for assigning importance weighting to different support vectors in the analysis of the discriminative direction. 4 Simple Example We first demonstrate the the proposed approach on a simple example. Fig. 2 shows three different classifiers, linear, quadratic and Gaussian RBF, for the same example training set that was generated using two Gaussian densities with different means and covariance matrices. We show the estimated discriminative direction for all points that are close to the separating boundary, not just support vectors. While the magnitude of discriminative direction vector is irrelevant in our infinitesimal analysis, we scaled the vectors in the figure according to the magnitude of the classifier gradient to illustrate importance ranking. Note that for the RBF support vectors far away from the boundary (Fig. 2c), the magnitude of the gradient is so small (tenth of the magnitude at the boundary), it renders the vectors Normal Control Patient Figure 3: Right hippocampus in schizophrenia study. First support vector from each group is shown, four views per shape (front, medial, back, lateral). The color coding is used to visualize the amount and the direction of the deformation that corresponds to the discriminative direction, changing from blue (moving inwards) to green (zero deformation) to red (moving outwards). too short to be visible in the figure. We can see that in the areas where there is enough evidence to estimate the boundary reliably, all three classifiers agree on the boundary and the discriminative direction (lower cluster of arrows). However, if the boundary location is reconstructed based on the regularization defined by the kernel, the classifiers suggest different answers (the upper cluster of arrows), stressing the importance of model selection for classification. The classifiers also provide an indication of the reliability of the differences represented by each arrow, which was repeatedly demonstrated in other experiments we performed. 5 Morphological Studies Morphological studies of anatomical organs motivated the analysis presented in this paper. Here, we show the results for the hippocampus study in schizophrenia. In this study, MRI scans of the brain were acquired for schizophrenia patients and a matched group of normal control subjects. The hippocampus structure was segmented (outlined) in all of the scans. Using the shape information (positions of the outline points), we trained a Gaussian RBF classifier to discriminate between schizophrenia patients and normal controls. However, the classifier in its original form does not provide the medical researchers with information on how the hippocampal shape varies between the two groups. Our goal was to translate the information captured by the classifier into anatomically meaningful terms of organ development and deformation. In this application, the coordinates in the input space correspond to the surface point locations for any particular example shape. The discriminative direction vector corresponds to displacements of the surface points and can be conveniently represented by a deformation of the original shape, yielding an intuitive description of shape differences for visualization and further analysis. We show the deformation that corresponds to the discriminative direction, omitting the details of shape extraction (see [5] for more information). Fig. 3 displays the first support vector from each group with the discriminative direction “painted” on it. Each row shows four snapshots of the same shape form different viewpoints3. The color at every node of the surface encodes the corresponding component of the discriminative direction. Note that the deformation represented by the two vectors is very similar in nature, but of opposite signs, as expected from the analysis in Section 3.3. We can see that the main deformation represented by this pair of vectors is localized in the bulbous “head” of 3An alternative way to visualize the same information is to actually generate the animation of the example shape undergoing the detected deformation. the structure. The next four support vectors in each group represent a virtually identical deformation to the one shown here. Starting with such visualization, the medical researchers can explore the organ deformation and interaction caused by the disease. 6 Conclusions We presented an approach to quantifying the classifier’s behavior with respect to small changes in the input vectors, trying to answer the following question: what changes would make the original input look more like an example from the other class without introducing irrelevant changes? We introduced the notion of the discriminative direction, which corresponds to the maximum changes in the classifier’s response while minimizing irrelevant changes in the input. For kernel-based classifiers the discriminative directions is determined by minimizing the divergence of the infinitesimal displacement vector and the normal to the separating hyperplane in the higher dimensional kernel space. The classifier interpretation in terms of the original features in general, and the discriminative direction in particular, is an important component of the data analysis in many applications where the statistical learning techniques are used to discover and study structural differences in the data. Acknowledgments. Quadratic optimization was performed using PR LOQO optimizer written by Alex Smola. This research was supported in part by NSF IIS 9610249 grant. References [1] S. Amari and S. Wu. Improving Support Vector Machines by Modifying Kernel Functions. Neural Networks, 783-789, 1999. [2] S. Amari. Natural Gradient Works Efficiently in Learning. Neural Comp., 10:251-276, 1998. [3] C. J. C. Burges. A Tutorial on Support Vector Machines for Pattern Recognition. Data Mining and Knowledge Discovery, 2(2):121-167, 1998. [4] C. J. C. Burges. Geometry and Invariance in Kernel Based Methods. In Adv. in Kernel Methods: Support Vector Learning, Eds. Sch¨olkopf, Burges and Smola, MIT Press, 89-116, 1999. [5] P. Golland et al. Small Sample Size Learning for Shape Analysis of Anatomical Structures. In Proc. of MICCAI’2000, LNCS 1935:72-82, 2000. [6] B. Sch¨olkopf et al. Input Space vs. Feature Space in Kernel-Based Methods. IEEE Trans. on Neural Networks, 10(5):1000-1017, 1999. [7] B. Sch¨olkopf, A. Smola, and K.-R. M¨uller. Nonlinear Component Analysis as a Kernel Eigenvalue Problem. Neural Comp., 10:1299-1319, 1998. [8] V. N. Vapnik. The Nature of Statistical Learning Theory. Springer, 1995. [9] V. N. Vapnik. Statistical Learning Theory. John Wiley & Sons, 1998.
|
2001
|
27
|
2,022
|
Contextual Modulation of Target Saliency Antonio Torralba Dept. of Brain and Cognitive Sciences MIT, Cambridge, MA 02139 torralba@ai.mit. edu Abstract The most popular algorithms for object detection require the use of exhaustive spatial and scale search procedures. In such approaches, an object is defined by means of local features. fu this paper we show that including contextual information in object detection procedures provides an efficient way of cutting down the need for exhaustive search. We present results with real images showing that the proposed scheme is able to accurately predict likely object classes, locations and sizes. 1 Introduction Although there is growing evidence of the role of contextual information in human perception [1], research in computational vision is dominated by object-based representations [5,9,10,15]. In real-world scenes, intrinsic object information is often degraded due to occlusion, low contrast, and poor resolution. In such situations, the object recognition problem based on intrinsic object representations is ill-posed. A more comprehensive representation of an object should include contextual information [11,13]: Obj. representatian == {intrisic obj. model, contextual obj. model}. In this representation, an object is defined by 1) a model of the intrinsic properties of the object and 2) a model of the typical contexts in which the object is immersed. Here we show how incorporating contextual models can enhance target object saliency and provide an estimate of its likelihood and intrinsic properties. 2 Target saliency and object likelihood Image information can be partitioned into two sets of features: local features, VL, that are intrinsic to an object, and contextual features, rUe which encode structural properties of the background. In a statistical framework, object detection requires evaluation of the likelihood function (target saliency function): P(O IVL, va) which provides the probability of presence of the object 0 given a set of local and contextual measurements. 0 is the set of parameters that define an object immersed in a scene: 0 == {on, x, y, i} with on==object class, (x,y)==location in image coordinates (1) and bobject appearance parameters. By applying Bayes rule we can write: P(O IVL, va) = P(vL11 va) P(VL 10, va)P(O Iva) Those three factors provide a simplified framework for representing three levels of attention guidance when looking for a target: The normalization factor, 1/P(VL Iva), does not depend on the target or task constraints, and therefore is a bottom-up factor. It provides a measure of how unlikely it is to find a set of local measurements VL within the context va. We can define local saliency as S(x,y) == l/P(vL(x,y) Iva). Saliency is large for unlikely features ina' scene. The second factor, P(VL 10, va), gives the likelihood of the local measurements VL when the object is present at such location in a particular context. We can write P(VL 10, va) ~ P(VL 10), which is a convenient approximation when the aspect of the target object is fully determined by the parameters given by the description O. This factor represents the top-down knowledge of the target· appearance and how it contributes to the search. Regions of the image with features unlikely to belong to the target object are vetoed. and regions with attended features are enhanced. The third factor, the PDF P(O Iva), provides context-based priors on object class, location and scale. It is of capital importance for insuring reliable inferences in situations where the local image measurements VL produce ambiguous interpretations. This factor does not depend on local measurements and target models [8,13]. Therefore, the term P(O Iva) modulates the saliency of local image properties when looking for an object of the class On. Contextual priors become more evident if we apply Bayes rule successively in order to split the PDF P(0 Iva) into three factors that model three kinds of context priming on object search: (2) According to this decomposition of the PDF, the contextual modulation of target saliency is a function of three main factors: Object likelihood: P(on Iva) provides the probability of presence of the object class On in the scene. If P(On Iva) is very small, then object search need not be initiated (we do not need to look for cars in a living room). Contextual control of focus of attention: P(x, y I On, va)· This PDDF gives the most likely locations for the presence of object On given context information, and it allocates computational resources into relevant scene regions. Contextual selection of local target appearance: P(tl.va, on). This gives the likely (prototypical) shapes (point of views, size, aspect ratio, object aspect) of the object On in the context Va- Here t == {a, p}, with a==scale and p==aspect ratio. Other parameters describing the appearance of an object in an image can be added. The image features most commonly used for describing local structures are the energy outputs of oriented band-pass filters, as they have been shown to be relevant for the task of object detection [9,10] and scene recognition [2,4,8,12]~ Therefore, the local image representation at the spatial location (x) is given by the vector VL(X) == {v(X,k)}k==l,N with: (3) 1 1 1 ",.-..... '0 ",.-..... u u ;> ;> ;> -a -a -a 0 0 0 '0: P: 0: 2 3 4 o 1 2 3 4 o 1 2 3 4 Figure 1: Contextual object prImIng of four objects categories (I-people, 2furniture, 3-vehicles and 4-trees) where i(x) is the input image and gk(X) are oriented band-pass filters defined by gk(i) == e-llxI12/u~e27fj<f~,x>. In such a representation [8], v(i,k) is the output magnitude- at the location i of a complex Gabor filter tuned to the spatial frequency f~. The variable k indexes filters tuned to different spatial frequencies and orientations. On the other ,hand, contextual features have to summarize the structure of the whole image. It has been shown that a holistic low-dimensional encoding of the local image features conveys enough information for a semantic categorization of the scene/context [8] and can be used for contextual priming in object recognition tasks [13]. Such a representation can be achieved by decomposing the image features into the basis functions provided by PCA: an == L L v{x, k) 1/ln{x, k) x k N v(x, k) ~ L an1/ln(x, k) n=l (4) We propose to use the decomposition coefficients vc == {an}n=l,N as context features. The functions 1/ln are the eigenfunctions of the covariance operator given by v(x, k). By using only a reduced set of components (N == 60 for the rest of the paper), the coefficients {an}n=l,N encode the main spectral characteristics of the scene with a coarse description of their spatial arrangement. In essence, {an}n=l,N is a holistic representation as all the regions of the image contribute to all the coefficients, and objects are not encoded individually [8]. In the rest of the paper we show the efficacy of this set of features in context modeling for object detection tasks. 3 Contextual object priming The PDF P(On Iva) gives the probability of presence of the object class On given contextual information. In other words, the PDF P{on Ive) evaluates the consistency of the object On with the context vc. For instance, a car has a high probability of presence in a highway scene but it is inconsistent with an indoor environment. The goal of P(on Ive) is to cut down the number of possible object categories to deal with before- expending computational resources in the object recognition process. The learning of the PDF P(on Ive) == P(ve IOn)P(on)/p(ve) with p(vo) == P(vc IOn)P{on) + P(vc l-,on)P(-,on) is done by approximating the in-class and out-of-class PDFs by a mixture of Gaussians: L P(ve IOn) == L bi,nG(VC;Vi,n, Vi,n) i=l (5) Figure 2: Contextual control of focus of attention when the algorithm is looking for cars (upper row) or heads (bottom row). The model parameters (bi,n, Vi,n, Vi,n) for the object class On are obtained using the EM algorithm [3]. The learning requires the use of few Gaussian clusters (L == 2 provides very good performances). For the learning, the system is trained with a set of examples manually annotated with the .presence/absence of four objects categories (i-people, 2-furniture, 3-vehicles and 4-trees). Fig. 1 shows some typical results from the priming model on the four superordinate categories of objects defined. Note. that the probability function P(on Ive) provides information about the probable presence of one object without scanning the picture. IfP(On Ive) > 1th then we can predict that the target is present. On the other hand, if P(On Ive) < th we can predict that the object is likely to be absent before exploring the image. The number of scenes in which the system may be able to take high confidence decisions will depend on different factors such as: the strength of the relationship between the target object and its context and the ability of ve for efficiently characterizing the context. Figure 1 shows some typical results from the priming model for a set of super-ordinate categories of objects. When forcing the model to take binary decisions in all the images (by selecting an acceptance threshold of th == 0.5) the presence/absence of the objects was correctly predicted by the model on 81%of the scenes of the test set. For each object category, high confidence predictions (th == .1) were made in at least 50% of the tested scene pictures and the presence/absence of each object class was correctly predicted by the model on 95% of those images. Therefore, for those images, we do not need to use local image analysis to decide about the presence/absence of the object. 4 Contextual control of focus of attention One of the strategies that biological visual systems use to deal with the analysis of real-world scenes is to focus attention (and, therefore, computational resources) onto the important image regions while neglecting others. Current computational models of visual attention (saliency maps anQ target detection) rely exclusively on local information or intrinsic object models [6,7,9,14,16]. The control of the focus of attention by contextual information that we propose. here is both task driven (looking for object on) and context driven (given global context information: ve). However, it does riot include any model of the target object at this stage. In our framework, the problem of contextual control of the focus of attention involves the CARS. 1 ~ CARS o ••• I P; •• "'0 0: • ~ :.S: _ filii •• \ • .~\.:.. ~ .\. fill tI':,._.: •• ••\ ~ .tto II 0.4 tre. • • 1,---~_"""""""-----R_eal_sc_al--..Je .: oReal pose 1 10 pixels 100 0.4 1 100 Q.) ~ ~ Q.) HEADS •••• 11 1.8 100 ] •• ~ "'0 .~.~. E t. ,.,:-.,,, • S•• , -: • 10 .~ ... ,••-=- •• ~ •• fIlIIe·': ":I':·.? 1 ~ , Real scale 1 10 pixels 100 Figure 3: Estimation results of object scale and pose based on contextual features. (6) evaluation of the PDF P(xlon,vo). For the learning, the joint PDF is modeled as a sum of gaussian clusters. Each cluster is decomposed into the product of two gaussians modeling respectively the distribution of object locations and the distribution of contextual features for each cluster: L P(x, vol on) == L bi,n G(x; Xi,n, Xi,n)G(VO; Vi,n, Vi,n) i==l The training set used for the learning of the PDF P(x, vol on) is a subset of'the pictures that contain the object On. The training data is {Vt}t==l,Nt and {Xt}t==l,Nt where Vt are the contextual features of the picture t of the training set and Xt is the location of object On in the image. The model parameters are obtained using the EM algorithm [3,13]. We used 1200 pictures for training and a separate set of 1200 pictures for testing. The success of the PDF in narrowing the region of the focus of attention will depend on the consistency of the relationship between the object and the context. Fig. 2 shows several examples of images and the selected regions based on contextual features when looking for cars and faces. From the PDF P(x, Vo IOn) we selected the region with the highest probability (33% of the image size on average). 87% of the heads present in the test pictures were inside the selected regions. 5 Contextual selection of object appearance models One major problem for computational approaches to object detection is the large variability in object appearance. The classical solution is to explore the space of possible shapes looking for the best match. The main sources of variability in object appearance are size, pose and intra-class shape variability (deformations, style, etc.). We show here that including contextual information can reduce at le.ast the first two sources of variability. For instance, the expected size of people in an image differs greatly between an indoor environment and a perspective view of a street. Both environments produce different patterns of contextual features vo [8]. For the second factor, pose, in the case of cars, there is a strong relationship between the possible orientations of the object and the scene configuration. For instance, looking down a highway, we expect to see the back of the cars, however, in a street view, looking towards the buildings, lateral views of cars are more likely. The expected scale and pose of the target object can be estimated by a regression procedure. The training database used for building the regression is a set of 1000 images in which the target object On is present. For each training image the target Figure 4: Selection of prototypical object appearances based on contextual cues. object was selected by cropping a rectangular window. For faces and cars we define the u == scale as the height of the selected window and the P == pose as the ratio between the horizontal and vertical dimensions of the window (~y/~x). On average, this definition of pose provides a good estimation of the orientation for cars but not for heads. Here we used regression using a mixture of gaussians for estimating the conditional PDFs between scale, pose and contextual features: P(u IVa, on) and PCP Iva, on). This yields the next regression procedures [3]: (j == Ei Ui,nbi,nG(Va; Vi,n, Vi,n) Ei bi,nG(vO; Vi,n, Vi,n) _ EiPi,nbi,nG(VO;Vi,n, Vi,n) P == Ei bi,nG(VC;Vi,n, Vi,n) (7) The results summarized in fig. 3 show that context is a strong cue for scale selection for the face detection task but less important for the car detection task. On the other hand, context introduces strong constraints on the prototypical point of views of cars but not at all for heads. Once the two parameters (pose and scale) have been estimated, we can build a prototypical model of the target object. In the case of a view-based object representation, the model of the object will consist of a collection of templates that correspond to the possible aspects of the target. For each image the system produces a collection of views, selected among a database of target examples that have the scale and pose given by eqs. (7). Fig. 4 shows some results from this procedure. In the statistical framework, the object detection requires the evaluation of the function P(VL 10, va). We can approximate Input image (target = cars) Object priming and Contextual control Target model selection of focus of attention 1 Integration of local features Target saliency Figure 5: Schematic layout of the model for object detection (here cars) by integration of contextual and local information. The bottom example is an error in detection due to incorrect context identification. P(VL 10, va) ~ P(VL IOn' (J", p). Fig. 5 and 6 show the complete chain of operations and some detection results using a simple correlation technique between the image and the generated object models (100 exemplars) at only one scale. The last image of each row shows the total object likelihood obtained by multiplying the object saliency maps (obtained by the correlation) and the contextual control of the focus of attention. The result shows how the use of context helps reduce false alarms. This results in good detection performances despite the simplicity of the matching procedure used. 6 Conclusion The contextual schema of a scene provides the likelihood of presence, typical locations and appearances of objects within the scene. We have proposed a model for incorporating such contextual cues in the task of object detection. The main aspects of our approach are: 1) Progressive reduction of the window of focus of attention: the system reduces the size of the focus of attention by first integrating contextual information and then local information. 2) Inhibition of target like patterns that are in inconsistent locations. 3) Faster detection of correctly scaled targets that have a pose in agreement with the context. 4) No requirement of parsing a scene into individual objects. Furthermore, once one object has been detected, it can introduce new contextual information for analyzing the rest of the scene. Acknowledglllents The author wishes to thank Dr. Pawan Sinha, Dr. Aude Oliva and Prof. Whitman Richards for fruitful discussions. References [1] Biederman, I., Mezzanotte, R.J., & Rabinowitz, J.C. (1982). Scene perception: detecting and judging objects undergoing relational violations. Cognitive Psychology, 14:143177. Feature maps \ I V t---HXJ---+l . . . . ~ Figure 6: Schema for object detection (e.g. cars) integrating local and giobal information. [2] Carson, C., Belongie, S., Greenspan, H., and Malik, J. (1997). Region-based image querying. Proc. IEEE W. on Content-Based Access of Image and Video Libraries, pp: 42-49. [3] Gershnfeld, N. The nature of mathematical modeling. Cambridge university press, 1999. [4] Gorkani, M. M., Picard, R. W. (1994). Texture orientation for sorting photos 'at a glance'. Proc. Int. Conf. Pat. Rec., Jerusalem, Vol. I: 459-464. [5] Heisle, B., T. Serre, S. Mukherjee and T. Poggio. (2001) Feature Reduction and Hierarchy of Classifiers for Fast Object Detection in Video Images. In: Proceedings of 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE Computer Society Press, Jauai, Hawaii. [6] Itti, L., Koch, C., & Niebur, E. (1998). A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Analysis and Machine Vision, 20(11):1254. [7] Moghaddam, B., & Pentland, A. (1997). Probabilistic Visual Learning for Object Representation. IEEE Trans. Pattern Analysis and Machine Vision, 19(7):696-710. [8] Oliva, A., & Torralba, A. (2001). Modeling the Shape of the Scene: A holistic representation of the spatial envelope. Int. Journal of Computer Vision, 42(3):145-175. [9] Rao, R.P.N., Zelinsky, G.J., Hayhoe, M.M., & Ballard, D.H. (1996). Modeling saccadic targeting in visual search. NIPS 8. Cambridge, MA: MIT Press. [10] Schiele, B., Crowley, J. L. (2000) Recognition without Correspondence using Multidimensional Receptive Field Histograms, Int. Journal of Computer Vision, Vol. 36(1):31-50. [11] Strat, T. M., & Fischler, M. A. (1991). Context-based vision: recognizing objects using information from both 2-D and 3-D imagery. IEEE trans. on Pattern Analysis and Machine Intelligence, 13(10): 1050-1065. [12] Szummer, M., and Picard, R. W. (1998). Indoor-outdoor image classification. In IEEE intl. workshop on Content-based Access of Image and Video Databases, 1998. [13] Torralba, A., & Sinha, P. (2001). Statistical context priming for object detection. IEEE Proc. Of Int. Conf in Compo Vision. [14] Treisman, A., & Gelade, G. (1980). A feature integration theory of attention. Cognitive Psychology, Vol. 12:97-136. [15] Viola, P. and Jones, M. (2001). Rapid object detection using a boosted cascade of simple features. In: Proceedings of 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2001), IEEE Computer Society Press, Jauai, Hawaii. [16] Wolfe, J. M. (1994). Guided search 2.0. A revised model of visual search. Psychonomic Bulletin and Review, 1:202-228
|
2001
|
28
|
2,023
|
On Kernel-Target Alignment N ello Cristianini BIOwulf Technologies nello@support-vector. net Andre Elisseeff BIOwulf Technologies andre@barnhilltechnologies.com John Shawe-Taylor Royal Holloway, University of London john@cs.rhul.ac.uk Jaz Kandola Royal Holloway, University of London jaz@cs.rhul.ac.uk Abstract We introduce the notion of kernel-alignment, a measure of similarity between two kernel functions or between a kernel and a target function. This quantity captures the degree of agreement between a kernel and a given learning task, and has very natural interpretations in machine learning, leading also to simple algorithms for model selection and learning. We analyse its theoretical properties, proving that it is sharply concentrated around its expected value, and we discuss its relation with other standard measures of performance. Finally we describe some of the algorithms that can be obtained within this framework, giving experimental results showing that adapting the kernel to improve alignment on the labelled data significantly increases the alignment on the test set, giving improved classification accuracy. Hence, the approach provides a principled method of performing transduction. Keywords: Kernels, alignment, eigenvectors, eigenvalues, transduction 1 Introduction Kernel based learning algorithms [1] are modular systems formed by a generalpurpose learning element and by a problem specific kernel function. It is crucial for the performance of the system that the kernel function somehow fits the learning target, that is that in the feature space the data distribution is somehow correlated to the label distribution. Several results exist showing that generalization takes place only when such correlation exists (nofreelunch; luckiness), and many classic estimators of performance (eg the margin) can be understood as estimating this relation. In other words, selecting a kernel in this class of systems amounts to the classic feature and model selection problems in machine learning. Measuring the similarity between two kernels, or the degree of agreement between a kernel and a given target function, is hence an important problem both for conceptual and for practical reasons. As an example, it is well known that one can obtain complex kernels by combining or manipulating simpler ones, but how can one predict whether the resulting kernel is better or worse than its components? What a kernel does is to virtually map data into a feature space so that their relative positions in that space are what matters. The degree of clustering achieved in that space, and the relation between the clusters and the labeling to be learned, should be captured by such an estimator. Alternatively, one could regard kernels as 'oracles' or 'experts' giving their opinion on whether two given points belong to the same class or not. In this case, the correlation between experts (seen as random variables) should provide an indication of their similarity. We will argue that - if one were in possess of this information - the ideal kernel for a classification target y(x) would be K(x, z) = y(x)y(z). One way of estimating the extent to which the kernel achieves the right clustering is to compare the sum of the within class distances with the sum of the between class distances. This will correspond to the alignment between the kernel and the ideal kernel y(x)y(z). By measuring the similarity of this kernel with the kernel at hand - on the training set - one can assess the degree of fitness of such kernel. The measure of similarity that we propose, 'kernel alignment' would give in this way a reliable estimate of its expected value, since it is sharply concentrated around its mean. In this paper we will motivate and introduce the notion of Alignment (Section 2); prove its concentration (Section 3); discuss its implications for the generalisation of a simple classifier (Section 4) and deduce some simple algorithms (Section 5) to optimize it and finally report on some experiments (Section 6). 2 Alignment Given an (unlabelled) sample 8 = {Xl, ... ,xm }, we use the following inner product between Gram matrices, (K1,K2)F = 2:7,'j=l K 1(Xi,Xj)K2(Xi,Xj) Definition 1 Alignment The (empirical) alignment of a kernel k1 with a kernel k2 with respect to the sample 8 is the quantity A(8 k k) = (K1,K2 )F , 1, 2 J(K1,K1)F(K2, K2)F' where Ki is the kernel matrix for the sample 8 using kernel ki. This can also be viewed as the cosine of the angle between two bi-dimensional vectors K1 and K2, representing the Gram matrices. If we consider K2 = yyl, where y is the vector of { -1, + I} labels for the sample, then A(8 K I) (K, yyl)F (K, yyl)F . / 1 I) 2 , ,yy =. / / K K) / 1 I) . / / K K) , smce \yy ,yy F = m y \, F\YY ,yy F my \, F We will occasionally omit the arguments K or y when these are understood from the context or when y forms part of the sample. In the next section we will see how this definition provides with a method for selecting kernel parameters and also for combining kernels. 3 Concentration The following theorem shows that the alignment is not too dependent on the training set 8. This result is expressed in terms of 'concentration'. Concentration means that the probability of an empirical estimate deviating from its mean can be bounded as an exponentially decaying function of that deviation. This will have a number of implications for the application and optimisation of the alignment. For example if we optimise the alignment on a random sample we can expect it to remain high on a second sample. Furthermore we will show in the next section that if the expected value of the alignment is high, then there exist functions that generalise well. Hence, the result suggests that we can optimise the alignment on a training set and expect to keep high alignment and hence good performance on a test set. Our experiments will demonstrate that this is indeed the case. The theorem makes use of the following result due to McDiarmid. Note that lEs is the expectation operator under the selection of the sample. TheoreIll 2 (McDiarmid!4}) Let Xl, ... ,Xn be independent random variables taking values in a set A, and assume that f : An -+ m. satisfies for 1 ::::; i ::::; n then for all f > 0, TheoreIll 3 The sample based estimate of the alignment is concentrated around its expected value. For a kernel with feature vectors of norm 1, we have that pm{s: 1.4(S) - A(y)1 ::::: €} ::::; 8 where € = C(S)V8ln(2/8)/m, (1) for a non-trivial function C (S) and value A(y). Proof: Let A 1 ~ A 1 ~ 2 lEs[.41(S)] A1(S) = m 2 .~ Yiyjk(Xi,Xj),A2(S) = m 2 .~ k(xi,Xj) , and A(y) = / A • ',J=l ',J=l ylES [A2(S)] First note that .4(S) = .41(S)/) .42(S). Define Al = lES[A1(S)] and A2 = lES[A2(S)], First we make use of McDiarmid's theorem to show that Ai(S) are concentrated for i = 1,2. Consider the training set S' = S \ {(Xi, Yi)} U {(X~, y~)}. We must bound the difference A A 1 4 IAj(S) - Aj(S')1 ::::; -2 (2(m - 1)2) < -, m m for j = 1,2. Hence, we have Ci = 4/m for all i and we obtain from an application of McDiarmid's Theorem for j = 1 and 2, < 2exp ( f;m) Setting f = V8ln(2/8)/m, the right hand sides are less than or equal to 8/2. Hence, with probability at least 1 - 8, we have for j = 1, 2 1 Aj (S) - Aj 1 < f. But whenever these two inequalities hold, we have < < Remark. We could also define the true Alignment, based on the input distribution P, as follows: given functions f,g : X 2 --+ JR, we define (j,g)p = IX2 f(x, z)g(x, z)dP(x)dP(z), Then the alignment of a kernel k1 with a kernel k2 is the quantity A(k1' k2) = J (kl,k2)P . (kl ,kl) P (k2 ,k2) P Then it is possible to prove that asymptotically as m tends to infinity the empirical alignment as defined above converges to the true alignment. However if one wants to obtain unbiased convergence it is necessary to slightly modify its definition by removing the diagonal, since for finite samples it biases the expectation by receiving too large a weight. With this modification A(y) in the statement of the theorem becomes the true alignment. We prefer not to pursue this avenue further for simplicity in this short article, we just note that the change is not significant. 4 Generalization In this section we consider the implications of high alignment for the generalisation of a classifier. By generalisation we mean the test error err(h) = P(h(x) ¥- y). Our next observation relates the generalisation of a simple classification function to the value of the alignment. The function we consider is the expected Parzen window estimator hex) = sign(f(x)) = sign (lE(XI ,v') [y'k(x', x)]). This corresponds to thresholding a linear function f in the feature space. We will show that if there is high alignment then this function will have good generalisation. Hence, by optimising the alignment we may expect Parzen window estimators to perform well. We will demonstrate that this prediction does indeed hold good in experiments. Theorem 4 Given any 8 > O. With probability 1 - 8 over a randomly drawn training set S, the generalisation accuracy of the expected Parzen window estimator h(x) = sign (lE(XI ,yl) [y' k(X', x)]) is bounded from above by err(h(x)) ::::: 1- A(S) + E + (mJ A2(S)) - 1, where E = C(S)V! ln~. Proof: (sketch) We assume throughout that the kernel has been normalised so that k(x,x) = 1 for all x. First observe that by Theorem 3 with probability greater than 1- 8/2, IA(y) - A(S)I ::::: E. The result will follow if we show that with probability greater than 1- 8/2 the generalisation error of hS\(xl,y,) can be upper bounded by 1 - A(y) + ~. Consider the quantity A(y) from Theorem 3. m A2(S) A(y) But lEs [~L:Z;= 1 Yiyjk(xi,xj)] lEs [~2 L:Z;=1 k(Xi,Xj)2] I mC-ml f(x) I IlE [2] < V (x,y) y lEs [~L:#j Yiyjk(xi,xj)] + ~ C (m -1)2 I 2 C2m 2 lE(XI,yl) [k(x, x ) ] < 1 Hence, if E P(f(x) ¥y) and a P(f(x) y), we have lEs [C~2 L:#j YiYj k(Xi' Xj)] ::::: 1 x a + 0 x E = a and E = 1 - a ::::: 1 - A(y) + c~, D An empirical estimate of the function f would be the Parzen window function. The expected margin of the empirical function is concentrated around the expected margin of the expected Parzen window. Hence, with high probability we can bound the error of j in terms of the empirically estimated alignment A(S). This is omitted due to lack of space. The concentration of j is considered in [3]. 5 Algorithms The concentration of the alignment can be directly used for tuning a kernel family to the particular task, or for selecting a kernel from a set, with no need for training. The probability that the level of alignment observed on the training set will be out by more than € from its expectation for one of the kernels is bounded by 6, where € is given by equation (1) for E = J ~ (InINI + lnj), where INI is the size of the set from which the kernel has been chosen. In fact we will select from an infinite family of kernels. Providing a uniform bound for such a class would require covering numbers and is beyond the scope of this paper. One of the main consequences of the definition of kernel alignment is in providing a practical criterion for combining kernels. We will justify the intuitively appealing idea that two kernels with a certain alignment with a target that are not aligned to each other, will give rise to a more aligned kernel combination. In particular we have that This shows that if two kernels with equal alignment to a given target yare also completely aligned to each other, then IIKI + K211F = IIKlllF + IIK211F and the alignment of the combined kernel remains the same. If on the other hand the kernels are not completely aligned, then the alignment of the combined kernel is correspondingly increased. To illustrate the approach we will take to optimising the kernel, consider a kernel that can be written in the form k(x, Xl) = l:.k I-tk(yk(x)yk(xl)) , where all the yk are orthogonal with respect to the inner product defined on the training set S, (y, yl)S = l:.:l YiYj. Assume further that one of them yt is the true label vector. We can now evaluate the alignment as A(y) ~ I-tt/v'l:.kl-t% . In terms of the Gram matrix this is written as Kij = l:.k I-tkyfyj where yf is the i-th label of the k-th classification. This special case is approximated by the decomposition into eigenvectors of the kernel matrix K = l:. Aiviv~, where Vi denotes the transpose of v and Vi is the i-th eigenvector with eigenvalue Ai. In other words, the more peaked the spectrum the more aligned (specific) the kernel can be. If by chance the eigenvector of the largest eigenvalue Al corresponds to the target labeling, then we will give to that labeling a fraction Ad v'l:.i AT of the weight that we can allocate to different possible labelings. The larger the emphasis of the kernel on a given target, the higher its alignment. In the previous subsection we observed that combining non-aligned kernels that are aligned with the target yields a kernel that is more aligned to the target. Consider the base kernels Ki = ViV~ where Vi are the eigenvectors of K, the kernel matrix for both labeled and unlabeled data. Instead of choosing only the most aligned ones, one could use a linear combination, with the weights proportional to their alignment (to the available labels): k = l:.i f(ai)viv~ where ai is the alignment of the kernel K i , and f(a) is a monotonically increasing function (eg. the identity or an exponential). Note that a recombination of these rank 1 kernels was made in so-called latent semantic kernels [2]. The overall alignment of the new kernel with the labeled data should be increased, and the new kernel matrix is expected also to be more aligned to the unseen test labels (because of the concentration, and the assumption that the split was random). Moreover, in general one can set up an optimization problem, aimed at finding the optimal a, that is the parameters that maximize the alignment of the combined kernel with the available labels. Given K = Li aiviv~ , using the orthonormality of the Vi and that (v v' ,uu') F = (v, u)}, the alignment can be written as A.(y) = (K, yy')F Li ai(vi, y)} mJLij aiaj(viv~, VjVj)F J(yy', yY')FJLi a;· Hence we have the following optimization problem: maximise W (a) (2) Setting derivatives to zero we obtain ~:. (Vi,Y)} - A2ai = 0 and hence ai (X (Vi,Y)}, giving the overall alignment A.(y) = JL,i~i'Y)j". This analysis suggests the following transduction algorithm. Given a partially labelled set of examples optimise its alignment by adapting the full kernel matrix by recombining its rank one eigenmatrices ViV~ using the coefficients ai determined by measuring the alignment between Vi and y on the labelled examples. Our results suggest that we should see a corresponding increase in the alignment on the unlabelled part of the set, and hence a reduction in test error when using a Parzen window estimator. Results of experiments testing these predictions are given in the next section. 6 Experiments We applied the transduction algorithm designed to take advantage of our results by optimizing alignment with the labeled part of the dataset using the spectral method described above. All of the results are averaged over 20 random splits with the standard deviation given in brackets. Table 1 shows the alignments of the Train Align Test Align Train Align Test Align 0.076 (0.007) 0.092 (0.029) 0.207 (0.020) 0.240 (0.083 0.228 ~0.012) 0.219 ~0.041) 0.240 ~0.016) 0.257 ~0.059) K50 0.075 ~0.016) 0.084 ~0.017) 0.210 ~0.031) 0.216 ~0.033) G50 0.242 (0.023) 0.181 (0.043) 0.257 (0.023) 0.202 (0.015) K 20 0.072 ~0.022) 0.081 ~0.006) 0.227 ~0.057) 0.210 ~0.015) G20 0.273 ~0.037) 0.034 ~0.046) 0.326 ~0.023) 0.118 ~0.017) Table 1: Mean and associated standard deviation alignment values using a linear kernel on the Breast (left two columns) and Ionosphere (right two columns). Gram matrices to the label matrix for different sizes of training set. The index indicates the percentage of training points. The K matrices are before adaptation, while the G matrices are after optimisation of the alignment using equation (2). The results on the left are for Breast Cancer data using a linear kernel, while the results on the right are for Ionosphere data. The left two columns of Table 2 shows the alignment values for Breast Cancer data using a Gaussian kernel together with the performance of an SVM classifier trained Table 2: Breast alignment (cols 1,2) and SVM error for a Gaussian kernel (sigma = 6) (col 3), Parzen window error for Breast (col 4) and Ionosphere (col 5) with the given gram matrix in the third column. The right two columns show the performance of the Parzen window classifier on the test set for Breast linear kernel (left column) and Ionosphere (right column). The results clearly show that optimising the alignment on the training set does indeed increase its value in all but one case by more than the sum of the standard deviations. Furthermore, as predicted by the concentration this improvement is maintained in the alignment measured on the test set with both linear and Gaussian kernels in all but one case (20% train with the linear kernel). The results for Ionosphere are less conclusive. Again as predicted by the theory the larger the alignment the better the performance that is obtained using the Parzen window estimator. The results of applying an SVM to the Breast Cancer data using a Gaussian kernel show a very slight improvement in the test error for both 80% and 50% training sets. 7 Conclusions We have introduced a measure of performance of a kernel machine that is much easier to analyse than standard measures (eg the margin) and that provides much simpler algorithms. We have discussed its statistical and geometrical properties, demonstrating that it is a well motivated and formally useful quantity. By identifying that the ideal kernel matrix has a structure of the type yy', we have been able to transform a measure of similarity between kernels into a measure of fitness of a given kernel. The ease and reliability with which this quantity can be estimated using only training set information prior to training makes it an ideal tool for practical model selection. We have given preliminary experimental results that largely confirm the theoretical analysis and augur well for the use of this tool in more sophisticated model (kernel) selection applications. References [1] N. Cristianini and J. Shawe-Taylor. An Introduction to Support Vector Machines. Cambridge University Press, 2000. See also the web site www.supportvector.net. [2] Nello Cristianini, Huma Lodhi, and John Shawe-Taylor. Latent semantic kernels for feature selection. Technical Report NC-TR-00-080, NeuroCOLT Working Group, http://www.neurocolt.org, 2000. [3] L. Devroye, L. Gyorfi, and G. Lugosi. A Probabilistic Theory of Pattern Recognition. Number 31 in Applications of mathematics. Springer, 1996. [4] C. McDiarmid. On the method of bounded differences. In Surveys in Combinatorics 1989, pages 148-188. Cambridge University Press, 1989.
|
2001
|
29
|
2,024
|
Eye movements and the maturation of cortical orientation selectivity Michele Rucci and Antonino Casile Department of Cognitive and Neural Systems, Boston University, Boston, MA 02215. Scuola Superiore S. Anna, Pisa, Italy Abstract Neural activity appears to be a crucial component for shaping the receptive fields of cortical simple cells into adjacent, oriented subregions alternately receiving ON- and OFF-center excitatory geniculate inputs. It is known that the orientation selective responses of V1 neurons are refined by visual experience. After eye opening, the spatiotemporal structure of neural activity in the early stages of the visual pathway depends both on the visual environment and on how the environment is scanned. We have used computational modeling to investigate how eye movements might affect the refinement of the orientation tuning of simple cells in the presence of a Hebbian scheme of synaptic plasticity. Levels of correlation between the activity of simulated cells were examined while natural scenes were scanned so as to model sequences of saccades and fixational eye movements, such as microsaccades, tremor and ocular drift. The specific patterns of activity required for a quantitatively accurate development of simple cell receptive fields with segregated ON and OFF subregions were observed during fixational eye movements, but not in the presence of saccades or with static presentation of natural visual input. These results suggest an important role for the eye movements occurring during visual fixation in the refinement of orientation selectivity. 1 Introduction Cortical orientation selectivity, i.e. the preference to edges with specific orientations exhibited by most cells in the primary visual cortex of different mammal species, is one of the most investigated characteristics of neural responses. Although the essential elements of cortical orientation selectivity seem to develop before the exposure to patterned visual input, visual experience appears essential both for refining orientation selectivity, and maintaining the normal response properties of cortical neurons. The precise mechanisms by which visually-induced activity contribute to the maturation of neural responses are not known. A number of experimental findings support the hypothesis that the development of orientation selective responses relies on Hebbian/covariance mechanisms of plasticity. According to this hypothesis, the stabilization of synchronously firing afferents onto common postsynaptic neurons may account for the segregation of neural inputs observed in the receptive fields of simple cells, where the adjacent oriented excitatory and inhibitory subregions receive selective input from geniculate ON- and OFF-center cells in the same retinotopic positions. Modeling studies [10, 9] have shown the feasibility of this proposal assuming suitable patterns of spontaneous activity in the LGN before eye opening. After eye opening, the spatiotemporal structure of LGN activity depends not only on the characteristics of the visual input, but also on the movements performed by the animal while exploring its environment. It may be expected that changes in the visual input induced by these movements play an important role in shaping the responses of neurons in the visual system. In this paper we focus on how visual experience and eye movements might jointly influence the refinement of orientation selectivity under the assumption of a Hebbian mechanism of synaptic plasticity. As illustrated in Fig. 1, a necessary requirement of the Hebbian hypothesis is a consistency between the correlated activity of thalamic afferents and the organization of simple cell receptive fields. Synchronous activation is required among geniculate cells of the same type (ON- or OFF-center) with receptive fields located at distances smaller than the width of a simple cell subregion, and among cells of opposite polarity with receptive fields at distances comparable to the separation between adjacent subregions. We have analyzed the second order statistical structure of neural activity in a model of cat LGN when natural visual input was scanned so as to replicate the oculomotor behavior of the cat. Patterns of correlated activity were compared to the structure of simple cell receptive fields at different visual eccentricities. 2 The model Modeling the activity of LGN cells LGN cells were modeled as linear elements with quasi-separable spatial and temporal components as proposed by [3]. This model, derived using the reverse-correlation technique, has been shown to produce accurate estimates of the activity of different types of LGN cells. Changes in the instantaneous firing rates with respect to the level of spontaneous activity were generated by evaluating the spatiotemporal convolution of the input image with the receptive field kernel
(1) where is the symbol for convolution, and are the spatial and temporal variables, and the operator indicates rectification ( !
"$#&% if &'(%!) otherwise). For each cell, the kernel consisted of two additive components, representing the center ( * ) and the periphery ( + ) of the receptive field respectively. Each of these two contributions was separable in its spatial ( , ) and temporal ( - ) elements: .
,0/ ! -1/ 0# ,02 ! -12 The spatial receptive fields of both center and surround were modeled as two-dimensional Gaussians, with a common space constant for both dimensions. Spatial parameters varied with eccentricity following neurophysiological measurements. As in [3], the temporal profile of the response was given by the difference of two gamma functions, with the temporal function for the periphery equal to that for the center and delayed by 3 ms. Modeling eye movements Modeled eye movements included saccades (both large-scale saccades and microsaccades), ocular drift, and tremor. Saccades— Voluntary saccadic eye movements, the fast shifts of gaze among fixation points, were modeled by assuming a generalized exponential distribution of fixation times. The amplitude and direction of a saccade were randomly selected among all possible saccades that would keep the point of fixation on the image. Following data described in the literature, the duration of each saccade was proportional to its amplitude. A modulation of geniculate activity was present in correspondence of each saccade [7]. Neural activity around the time of a saccade was multiplied by a gain function so that an initial suppression of activity with a peak of 10%, gradually reversed to a 20% facilitation with peak occurring 100 ms after the end of the saccade. Fixational eye movements— Small eye movements included fixational saccades, ocular drift and tremor. Microsaccades were modeled in a similar way to voluntary saccades, with amplitude randomly selected from a uniform distribution between 1 and 10 minutes of arc. No modulation of LGN activity was present in the case of microsaccades. Ocular drift and tremor were modeled together by approximating their power spectrum by means of a Poisson process filtered by a second order eye plant transfer function over the frequency range 0-40 Hz where the power declines as . This term represents the irregular discharge rate of motor units for frequency less than 40 Hz. Parameters were adjusted so as to give a mean amplitude of and a mean velocity equal to /s, which are the values measured in the cat [11]. 3 Results We simulated the activity of geniculate cells with receptive fields in different positions of the visual field, while receiving visual input in the presence of different types of eye movements. The relative level of correlation between units of the same and different types at positions and
in the LGN was measured by means of the correlation difference, D
ONON # ONOFF , where the two terms are the correlation coefficients evaluated between the two ON units at positions and
, and between the ON unit at position and the OFF unit at position
respectively. D is positive when the activity of units of the same type covary more strongly than that of units of different types, and is negative when the opposite occurs. The average relative levels of correlation between units with receptive fields at different distances in the visual field were examined by means of the function D
D ' , which evaluates the average correlation difference D among all pairs of cells at positions and
at distance from each other. For simplicity, in the following we refer to D as the correlation difference, implicitly assuming that a spatial averaging has taken place. The correlation difference is a useful tool for predicting the emerging patterns of connectivity in the presence of a Hebbian mechanism of synaptic plasticity. The average separation at which D changes sign is a key element in determining the spatial extent of the different subfields within the receptive fields of simple cells. Fig. 1 ( ) provides an example of application of the correlation difference function to quantify the correlated activity of LGN cells. In this example we have measured the level of correlation between pairs of cells with receptive fields at different separations when a spot of light was presented as input. An important element in the resulting level of correlation is the polarity of the two cells (i.e. whether they are ON- or OFF-center). As shown in Fig. 1 ( ), since geniculate cells tend to be coactive when the ON and OFF subregions of their receptive fields overlap, the correlation between pairs of cells of the same type decreases when the separation between their receptive fields is increased, while pairs of cells of opposite types tend to become more correlated. As a consequence, the correlation difference function, D , is positive at small separations, and negative at large ones. Fig. 2 shows the measured correlated activity for LGN cells located around 17 deg. of visual eccentricity in the presence of two types of visual input: retinal spontaneous activity and natural visual stimulation. Spontaneous activity was simulated on the basis of Matronarde’s data on the correlated firing of ganglion cells in the cat retina [8]. As illustrated by the graph, a close correspondence is present between the measured D and the response profile of an average cortical simple cell at this eccentricity, indicating that a ... ...
LGN OFF LGN ON V1 V1 RF (a) OFF ON 0 20 40 60 80 100 distance (min.) −1 0 1 2 normalized correlation ON−ON, OFF−OFF ON−OFF, OFF−ON difference (b) Figure 1: ( ) Patterns of correlated activity required by a Hebbian mechanism of synaptic plasticity to produce a segregation of geniculate afferents. On average ON- and OFF-center LGN cells overlapping excitatory and inhibitory subregions in the receptive field of a simple cell must be simultaneously active. ( ) Example of application of the correlation difference function, D . The icons on the top of each graph represent the positions of the receptive fields of the two cells at the corresponding separations along the axis. The bright dot marks the center of the spot of light. The three curves represent the correlation coefficients for pairs of units of the same type 2 (continuous thin line), units of opposite types (dashed line), and the correlation difference function D
2 # (bold line). Positive (negative) values of D indicate that the activity of LGN cells of the same (opposite) type covary more closely than the activity of cells of opposite (same) types. Hebbian mechanism of synaptic plasticity can well account for the structure of simple cell receptive fields before eye opening. What happens in the presence of natural visual input? We evaluated the correlation difference function on a database of 30 images of natural scenes. The mean power spectrum of our database was best approximated by !
"! $# %'& , which is consistent with the results of several studies investigating the power spectrum of natural images. The mean correlation difference function measured when the input images were analyzed statically is marked by dark triangles in the left panel of Fig. 2. Due to the wide spatial correlations of natural visual input, the estimated correlation difference did not change sign within the receptive field of a typical simple cell. That is, LGN cells of the same type were found to covary more closely than cells of opposite types at all separations within the receptive field of a simple cell. This result is not consistent with the putative role of a direct Hebbian/covariance model in the refinement of orientation selectivity after eye opening. A second series of simulations was dedicated to analyze the effects of eye movements on the structure of correlated activity. In these simulations the images of natural scenes were scanned so as to replicate cat oculomotor behavior. As shown in right panel of Fig. 2, significantly different patterns of correlated neural activity were found in the LGN in the presence of different types of eye movements. In the presence of large saccades, levels of correlations among the activity of geniculate cells were similar to the case of static presentation of natural visual input, and they did not match the structure of simple cell receptive fields. The dark triangles in Fig. 2 represent the correlation difference function evaluated over a window of observation of 100 ms in the presence of both large saccades and fixation eye movements. In contrast, when our analysis was restricted to the periods of visual fixation during which microscopic eye movements occurred, strong covariances were measured between cells of the same type located nearby and between cells of opposite types at distances compatible with the separation between different subregions in the receptive fields of simple cells. 0.0 0.5 1.0 1.5 2.0 2.5 3.0 distance (deg.) −0.3 0.2 0.7 normalized correlation cortical RF spontaneous activity natural visual input 0.0 0.5 1.0 1.5 2.0 distance (deg.) −0.3 0.2 0.7 normalized Cd Saccade + Fixation Cortical RF Fixation Figure 2: Analysis of the correlated activity of LGN units in different experimental conditions. In both graphs, the curve marked by white circles is the average receptive field of a simple cell, as measured by Jones and Palmer (1987) shown here for comparison. (Left) Static analysis: patterns of correlated activity in the presence of spontaneous activity and when natural visual input was analyzed statically. (Right) Effects of eye movements: correlation difference functions measured when natural images were scanned with sequence or saccades or fixational eye movements. Fig. 3 shows the results of a similar analysis for LGN cells at different visual eccentricities. The white circles in the panels of Fig. 3 represent the width of the largest subfield in the receptive field of cortical simple cells as measured by [13]. The other curves on the left panel represent the widths of the central lobe of the correlation difference functions (the spatial separation over which cells of the same type possess correlated activity, measured as the double of the point in which the correlation difference function intersects the zero axis) in the cases of spontaneous activity and static presentation of natural visual input. As in Fig. 2, (1) a close correspondence was present between the experimental data and the subregion widths predicted by the correlation difference function in the case of spontaneous activity; and (2) a significant deviation between the two measurements was present in the case of static examination of natural visual input. The right panel in Fig. 3 shows the correlation difference functions obtained at different visual eccentricities in the presence of fixational eye movements. The minimum separation between receptive fields necessary for observing strong levels of covariance between cells with opposite polarity increased with eccentricity, as illustrated by the increase in the central lobe of the estimated correlation functions at the different visual eccentricities. As for the case of spontaneous activity, a close correspondence is now present between the spatiotemporal characteristics of LGN activity and the organization of simple cell receptive fields. 4 Discussion In this paper we have used computer modeling to study the correlated activity of LGN cells when images of natural scenes were scanned so as to replicate cat eye movements. In the absence of eye movements, when a natural visual environment was observed statically, similar to the way it is examined by animals with their eyes paralyzed, we found that the simulated responses of geniculate cells of the same type at any separation smaller than the receptive field of a simple cell were strongly correlated. These spatial patterns of covarying geniculate activity did not match the structure of simple cell receptive fields. A similar result was obtained when natural scenes were scanned through saccades. Conversely, in 0 10 20 30 eccentricity (deg.) 0 2 5 8 10 central width (deg) Wilson & Sherman, 1976 spontaneous activity natural input (static) 0 10 20 30 eccentricity (deg.) 0 2 5 8 10 central width (deg) Wilson & Sherman, 1976 visual fixation Figure 3: Analysis of the correlated activity of LGN units at different visual eccentricities. The width of the larger subfield in the receptive field of simple cells at different eccentricities as measured by Wilson and Sherman (1976) (white circles) is compared to the width of the central lobe of the correlation difference functions measured in different conditions (Left) Static analysis: results obtained in the presence of spontaneous activity and when natural visual input was analyzed statically. (Right) Case of fixational eye movements and natural visual input. the case of micromovements, including both microsaccades and the combination of ocular drift and tremor, strong correlations were measured among cells of the same type located nearby and among cells of opposite types at distances compatible with the separation between different subregions in the receptive fields of simple cells. These results suggest a developmental role for the small eye movements that occur during visual fixation. Although the role of visual experience in the development of orientation selectivity has been extensively investigated, relatively few studies have focused on whether eye movements contribute to the development of the responses of cortical cells. Yet, experiments in which kittens were raised with their eyes paralyzed have shown basic deficiencies in the development of visually-guided behavior [6], as well as impairments in ocular dominance plasticity [4, 12]. In addition, it has been shown that eye movements are necessary for the reestablishment of cortical orientation selectivity in dark-reared kittens exposed to visual experience within the critical period [2, 5]. This indicates that simultaneous experience of visual input and eye movements (and/or eye movement proprioception) may be necessary for the refinement of orientation selectivity [1]. Our finding that the patterns of LGN activity with static presentation of natural images did not match the spatial structure of the receptive fields of simple cells is in agreement with the hypothesis that exposure to pattern vision per se is not sufficient to account for a normal visual development. A main assumption of this study is that the refinement and maintenance of orientation selectivity after eye opening is mediated by a Hebbian/covariance process of synaptic plasticity. The term Hebbian is used here with a generalized meaning to indicate the family of algorithms in which modifications of synaptic efficacies occur on the basis of the patterns of input covariances. While no previous theoretical study has investigated the influence of eye movements on the development of orientation selectivity, some models have shown that schemes of synaptic modifications based on the correlated activity of thalamic afferents can account well for the segregation of ON- and OFF-center inputs before eye opening in the presence of suitable patterns of spontaneous activity [10, 9]. By showing that, during fixation, the spatiotemporal structure of visually-driven geniculate activity is compatible with the structure of simple cell receptive fields, the results of the present study extend the plausibility of such schemes to the period after eye opening in which exposure to pattern vision occurs. Ocular movements are a common feature of the visual system of different species. It should not come as a surprise that a trace of their existence can be found even in some of the most basic properties of neurons in the early stages of the visual system, such as orientation selectivity. Further studies are needed to investigate whether similar traces can be found in other features of visual neural responses. References [1] P. Buisseret. Influence of extraocular muscle proprioception on vision. Physiol. Rev., 75(2):323–338, 1995. [2] P. Buisseret, E. Gary-Bobo, and M. Imbert. Ocular motility and recovery of orientational properties of visual cortical neurons in dark-reared kittens. Nature, 272:816– 817, 1978. [3] D. Cai, G. C. DeAngelis, and R. D. Freeman. Spatiotemporal receptive field organization in the lateral geniculate nucleus of cats and kitten. J. Neurophysiol., 78(2):1045– 61, 1997. [4] R. D. Freeman and A. B. Bonds. Cortical plasticity in monocularly deprived immobilized kittens depends on eye movement. Science, 206:1093–1095, 1979. [5] E. Gary-Bobo, C. Milleret, and P. Buisseret. Role of eye movements in developmental process of orientation selectivity in the kitten visual cortex. Vision Res., 26(4):557– 567, 1986. [6] A. Hein, F. Vital-Durand, W. Salinger, and R. Diamond. Eye movements initiate visual-motor development in the cat. Science, 204:1321–1322, 1979. [7] D. Lee and J. G. Malpeli. Effect of saccades on the activity of neurons in the cat lateral geniculate nucleus. J. Neurophysiol., 79:922–936, 1998. [8] D. N. Mastronarde. Correlated firing of cat retinal ganglion cells. I spontaneously active inputs to X and Y cells. J. Neurophysiol., 49(2):303–323, 1983. [9] K. D. Miller. A model of the development of simple cell receptive fields and the ordered arrangement of orientation columns through activity-dependent competition between ON- and OFF- center inputs. J. Neurosci., 14(1):409–441, 1994. [10] M. Miyashita and S. Tanaka. A mathematical model for the self-organization of orientation columns in visual cortex. Neuroreport, 3:69–72, 1992. [11] E. Olivier, A. Grantyn, M. Chat, and A. Berthoz. The control of slow orienting eye movements by tectoreticulospinal neurons in the cat: behavior, discharge patterns and underlying connections. Exp. Brain Res., 93:435–449, 1993. [12] W. Singer and J. Raushecker. Central-core control of developmental plasticity in the kitten visual cortex II. Electrical activation of mesencephalic and diencephalic projections. Exp. Brain Res., 47:22–233, 1982. [13] J. R. Wilson and S. M. Sherman. Receptive-field characteristics of neurons in the cat striate cortex: changes with visual field eccentricity. J. Neurophysiol., 39(3):512–531, 1976.
|
2001
|
3
|
2,025
|
Escaping the Convex Hull with Extrapolated Vector Machines. Patrick Haffner AT&T Labs-Research, 200 Laurel Ave, Middletown, NJ 07748 haffner@research.att.com Abstract Maximum margin classifiers such as Support Vector Machines (SVMs) critically depends upon the convex hulls of the training samples of each class, as they implicitly search for the minimum distance between the convex hulls. We propose Extrapolated Vector Machines (XVMs) which rely on extrapolations outside these convex hulls. XVMs improve SVM generalization very significantly on the MNIST [7] OCR data. They share similarities with the Fisher discriminant: maximize the inter-class margin while minimizing the intra-class disparity. 1 Introduction Both intuition and theory [9] seem to support that the best linear separation between two classes is the one that maximizes the margin. But is this always true? In the example shown in Fig.(l), the maximum margin hyperplane is Wo; however, most observers would say that the separating hyperplane WI has better chances to generalize, as it takes into account the expected location of additional training sam••••••••••• f\J:- •••••••••• •• .., --.Q. •• ~".' ~, _ .. ................ ~x~... W 1 --------------~--------------·K~ ··········· '''-,,- /0- 0 00 00 00 0'\ "OW-0_ o __ oo- o::-o . .. . ....... (} ............ . Figure 1: Example of separation where the large margin is undesirable. The convex hull and the separation that corresponds to the standard SVM use plain lines while the extrapolated convex hulls and XVMs use dotted lines. pIes. Traditionally, to take this into account, one would estimate the distribution of the data. In this paper, we just use a very elementary form of extrapolation ("the poor man variance") and show that it can be implemented into a new extension to SVMs that we call Extrapolated Vector Machines (XVMs). 2 Adding Extrapolation to Maximum Margin Constraints This section states extrapolation as a constrained optimization problem and computes a simpler dual form. Take two classes C+ and C_ with Y+ = +1 and Y_ = -1 1 as respective targets. The N training samples {(Xi, Yi); 1 ::::; i ::::; N} are separated with a margin p if there exists a set of weights W such that Ilwll = 1 and Vk E {+, -}, Vi E Ck, Yk(w,xi+b) 2: p (1) SVMs offer techniques to find the weights W which maximize the margin p. Now, instead of imposing the margin constraint on each training point, suppose that for two points in the same class Ck, we require any possible extrapolation within a range factor 17k 2: 0 to be larger than the margin: Vi,j E Ck, V)" E [-17k, l+17k], Yk (W.()"Xi + (l-)")Xj) + b) 2: P (2) It is sufficient to enforce the constraints at the end of the extrapolation segments, and (3) Keeping the constraint over each pair of points would result in N 2 Lagrange multipliers. But we can reduce it to a double constraint applied to each single point. If follows from Eq.(3) that: (4) (5) We consider J.Lk = max (Yk(W.Xj)) and Vk = min (Yk(W.Xj)) as optimization varilEC. lEC. abIes. By adding Eq.(4) and (5), the margin becomes 2p = L ((17k+1)vk- 17kJ.Lk) = L (Vk -17dJ.Lk - Vk)) (6) k k Our problem is to maximize the margin under the double constraint: Vi E Ck, Vk ::::; Yk(W.Xi) ::::; J.Lk In other words, the extrapolated margin maximization is equivalent to squeezing the points belonging to a given class between two hyperplanes. Eq.(6) shows that p is maximized when Vk is maximized while J.Lk - Vk is minimized. Maximizing the margin over J.Lk , Vk and W with Lagrangian techniques gives us the following dual problem: (7) lIn this paper, it is necessary to index the outputs y with the class k rather than the more traditional sample index i, as extrapolation constraints require two examples to belong to the same class. The resulting equations are more concise, but harder to read. Compared to the standard SVM formulation, we have two sets of support vectors. Moreover, the Lagrange multipliers that we chose are normalized differently from the traditional SVM multipliers (note that this is one possible choice of notation, see Section.6 for an alternative choice). They sum to 1 and allow and interesting geometric interpretation developed in the next section. 3 Geometric Interpretation and Iterative Algorithm For each class k, we define the nearest point to the other class convex hull along the direction of w: Nk = I:iECk f3iXi. Nk is a combination of the internal support vectors that belong to class k with f3i > O. At the minimum of (7), because they correspond to non zero Lagrange multipliers, they fallon the internal margin Yk(W,Xi) = Vk; therefore, we obtain Vk = Ykw.Nk· Similarly, we define the furthest point Fk = I:i ECk ~i Xi' Fk is a combination of the external support vectors, and we have flk = Ykw.Fk. The dual problem is equivalent to the distance minimization problem min IILYk ((1Jk+I)Nk _1Jk F k)11 2 Nk ,Fk EHk k where 1{k is the convex hull containing the examples of class k. It is possible to solve this optimization problem using an iterative Extrapolated Convex Hull Distance Minimization (XCHDM) algorithm. It is an extension of the Nearest Point [5] or Maximal Margin Percept ron [6] algorithms. An interesting geometric interpretation is also offered in [3]. All the aforementioned algorithms search for the points in the convex hulls of each class that are the nearest to each other (Nt and No on Fig.I), the maximal margin weight vector w = Nt - No-' XCHDM look for nearest points in the extrapolated convex hulls (X+ I and X-I on Fig.I). The extrapolated nearest points are X k = 1JkNk - 1JkFk' Note that they can be outside the convex hull because we allow negative contribution from external support vectors. Here again, the weight vector can be expressed as a difference between two points w = X+ - X - . When the data is non-separable, the solution is trivial with w = O. With the double set of Lagrange multipliers, the description of the XCHDM algorithm is beyond the scope of this paper. XCHDM with 1Jk = 0 are simple SVMs trained by the same algorithm as in [6]. An interesting way to follow the convergence of the XCHDM algorithm is the following. Define the extrapolated primal margin 1'; = 2p = L ((1Jk+ I )vk- 1Jkflk) k and the dual margin 1'; = IIX+ - X-II Convergence consists in reducing the duality gap 1'~ -1'; down to zero. In the rest of the paper, we will measure convergence with the duality ratio r = 1'~ . 1'2 To determine the threshold to compute the classifier output class sign(w.x+b) leaves us with two choices. We can require the separation to happen at the center of the primal margin, with the primal threshold (subtract Eq.(5) from Eq.(4)) 1 bl = -2" LYk ((1Jk+ I )vk-1JkJ.lk) k or at the center of the dual margin, with the dual threshold b2 = - ~w. 2:)(T}k+1)Nk - T}kFk) = - ~ (IIx+ 112 -lix-in k Again, at the minimum, it is easy to verify that b1 = b2 . When we did not let the XCHDM algorithm converge to the minimum, we found that b1 gave better generalization results. Our standard stopping heuristic is numerical: stop when the duality ratio gets over a fixed value (typically between 0.5 and 0.9). The only other stopping heuristic we have tried so far is based on the following idea. Define the set of extrapolated pairs as {(T}k+1)Xi -T}kXj; 1 :S i,j :S N}. Convergence means that we find extrapolated support pairs that contain every extrapolated pair on the correct side of the margin. We can relax this constraint and stop when the extrapolated support pairs contain every vector. This means that 12 must be lower than the primal true margin along w (measured on the non-extrapolated data) 11 = y+ + Y -. This causes the XCHDM algorithm to stop long before 12 reaches Ii and is called the hybrid stopping heuristic. 4 Beyond SVMs and discriminant approaches. Kernel Machines consist of any classifier of the type f(x) = L:i Yi(XiK(x, Xi). SVMs offer one solution among many others, with the constraint (Xi > O. XVMs look for solutions that no longer bear this constraint. While the algorithm described in Section 2 converges toward a solution where vectors act as support of margins (internal and external), experiments show that the performance of XVMs can be significantly improved if we stopped before full convergence. In this case, the vectors with (Xi =/: 0 do not line up onto any type of margin, and should not be called support vectors. The extrapolated margin contains terms which are caused by the extrapolation and are proportional to the width of each class along the direction of w. We would observe the same phenomenon if we had trained the classifier using Maximum Likelihood Estimation (MLE) (replace class width with variance). In both MLE and XVMs, examples which are the furthest from the decision surface play an important role. XVMs suggest an explanation why. Note also that like the Fisher discriminant, XVMs look for the projection that maximizes the inter-class variance while minimizing the intra-class variances. 5 Experiments on MNIST The MNIST OCR database contains 60,000 handwritten digits for training and 10,000 for testing (the testing data can be extended to 60,000 but we prefer to keep unseen test data for final testing and comparisons). This database has been extensively studied on a large variety of learning approaches [7]. It lead to the first SVM "success story" [2], and results have been improved since then by using knowledge about the invariance of the data [4]. The input vector is a list of 28x28 pixels ranging from 0 to 255. Before computing the kernels, the input vectors are normalized to 1: x = II~II' Good polynomial kernels are easy to define as Kp(x, y) = (x.y)P. We found these normalized kernels to outperform the unnormalized kernels Kp(x, y) = (a(x.y)+b)P that have been traditionally used for the MNIST data significantly. For instance, the baseline error rate with K4 is below 1.2%, whereas it hovers around 1.5% for K4 (after choosing optimal values for a and b)2. We also define normalized Gaussian kernels: Kp(x, y) = exp ( - ~ Ilx - y112) = [exp (x.y- 1)JP. (8) Eq.(8) shows how they relate to normalized polynomial kernels: when x.y « 1, Kp and Kp have the same asymptotic behavior. We observed that on MNIST, the performance with Kp is very similar to what is obtained with unnormalized Gaussian kernels Ku(x, y) = exp _(X~Y)2. However, they are easier to analyze and compare to polynomial kernels. MNIST contains 1 class per digit, so the total number of classes is M=10. To combine binary classifiers to perform multiclass classifications, the two most common approaches were considered . • In the one-vs-others case (lvsR) , we have one classifier per class c, with the positive examples taken from class c and negative examples form the other classes. Class c is recognized when the corresponding classifier yields the largest output . • In the one-vs-one case (lvs1), each classifier only discriminates one class from another: we need a total of (MU:;-l) = 45 classifiers. Despite the effort we spent on optimizing the recombination of the classifiers [8] 3, 1 vsR SVMs (Table 1) perform significantly better than 1 vs1 SVMs (Table 2). 4 For each trial, the number of errors over the 10,000 test samples (#err) and the total number of support vectors( #SV) are reported. As we only count SV s which are shared by different classes once, this predicts the test time. For instance, 12,000 support vectors mean that 20% of the 60,000 vectors are used as support. Preliminary experiments to choose the value of rJk with the hybrid criterion show that the results for rJk = 1 are better than rJk = 1.5 in a statistically significant way, and slightly better than rJk = 0.5. We did not consider configurations where rJ+ f; rJ-; however, this would make sense for the assymetrical 1 vsR classifiers. XVM gain in performance over SVMs for a given configuration ranges from 15% (1 vsR in Table 3) to 25% (1 vs1 in Table 2). 2This may partly explain a nagging mystery among researchers working on MNIST: how did Cortes and Vapnik [2] obtain 1.1% error with a degree 4 polynomial ? 3We compared the Max Wins voting algorithm with the DAGSVM decision tree algorithm and found them to perform equally, and worse than 1 vsR SVMs. This is is surprising in the light of results published on other tasks [8], and would require further investigations beyond the scope of this paper. 4Slightly better performance was obtained with a new algorithm that uses the incremental properties of our training procedure (this is be the performance reported in the tables). In a transductive inference framework, treat the test example as a training example: for each of the M possible labels, retrain the M among (M(":-l) classifiers that use examples with such label. The best label will be the one that causes the smallest increase in the multiclass margin p such that it combines the classifier margins pc in the following manner ~= ,,~ 2 ~ 2 P c~M Pc The fact that this margin predicts generalization is "justified" by Theorem 1 in [8]. Duality Ratio stop Kernel 0.40 0.75 0.99 #err #SV #err #SV # err #SV K3 136 8367 136 11132 132 13762 K4 127 8331 117 11807 119 15746 K5 125 8834 119 12786 119 17868 Kg 136 13002 137 18784 141 25953 [(2 147 9014 128 11663 131 13918 [(4 125 8668 119 12222 117 16604 K5 125 8944 125 12852 125 18085 Table 1: SVMs on MNIST with 10 1vsR classifiers Kernel SVM/ratio at 0.99 XVM/Hybrid # err #SV # err #SV K3 138 11952 117 17020 K4 135 13526 110 16066 K5 191 13526 114 15775 Table 2: SVMjXVM on MNIST with 45 1 vs1 classifiers The 103 errors obtained with K4 and r = 0.5 in Table 3 represent only about 1% error: this is the lowest error ever reported for any learning technique without a priori knowledge about the fact that the input data corresponds to a pixel map (the lowest reproducible error previously reported was 1.2% with SVMs and polynomials of degree 9 [4], it could be reduced to 0.6% by using invariance properties of the pixel map). The downside is that XVMs require 4 times as many support vectors as standards SVMs. Table 3 compares stopping according to the duality ratio and the hybrid criterion. With the duality ratio, the best performance is most often reached with r = 0.50 (if this happens to be consistently true, validation data to decide when to stop could be spared). The hybrid criterion does not require validation data and yields errors that, while higher than the best XVM, are lower than SVMs and only require a few more support vectors. It takes fewer iterations to train than SVMs. One way to interpret this hybrid stopping criterion is that we stop when interpolation in some (but not all) directions account for all non-interpolated vectors. This suggest that interpolation is only desirable in a few directions. XVM gain is stronger in the 1 vs 1 case (Table 2). This suggests that extrapolating on a convex hull that contains several different classes (in the 1 vsR case) may be undesirable. Duality Ratio stop Hybrid. Kernel 0.40 0.50 0.75 Stop Crit. # err #SV # err #SV # err #SV # err #SV K3 118 46662 111 43819 116 50216 125 20604 K4 112 40274 103 43132 110 52861 107 18002 K5 109 36912 106 44226 110 49383 107 17322 Kg 128 35809 126 39462 131 50233 125 19218 K2 114 43909 114 46905 114 53676 119 20152 [(4 108 36980 111 40329 114 51088 108 16895 Table 3: XVMs on MNIST with 10 1 vsR classifiers 6 The Soft Margin Case MNIST is characterized by the quasi-absence of outliers, so to assume that the data is fully separable does not impair performance at all. To extend XVMs to non-separable data, we first considered the traditional approaches of adding slack variables to allow margin constraints to be violated. The most commonly used approach with SVMs adds linear slack variables to the unitary margin. Its application to the XVM requires to give up the weight normalization constraint, so that the usual unitary margin can be used in the constraints [9] . Compared to standard SVMs, a new issue to tackle is the fact that each constraint corresponds to a pair of vectors: ideally, we should handle N 2 slack variables ~ij. To have linear constraints that can be solved with KKT, we need to have the decomposition ~ij = ('T}k+1)~i+'T}k~; (factors ('T}k+1) and 'T}k are added here to ease later simplifications). Similarly to Eq.(3), the constraint on the extrapolation from any pair of points is Vi,j E Ck, Yk (w. (('T}k+1)xi - 'T}kXj) +b) 2: 1 - ('T}k+1)~i - 'T}k~; with ~i'~; 2: 0 (9) Introducing J.tk = max (Yk(w,xj+b) ~;) and Vk = min (Yk(W,Xi+b) + ~i)' we obJECk .ECk tain the simpler double constraint Vi E Ck, Vk -~i ~ Yk(W,Xi+b) ~ J.tk+~; with ~i'~; 2: 0 (10) It follows from Eq.(9) that J.tk and Vk are tied through (l+'T}k)vk = l+'T}kJ.tk If we fix J.tk (and thus Vk) instead of treating it as an optimization variable, it would amount to a standard SVM regression problem with {-I, + I} outputs, the width of the asymmetric f-insensitive tube being J.tk-Vk = (~~~;)' This remark makes it possible for the reader to verify the results we reported on MNIST. Vsing the publicly available SVM software SVMtorch [1] with C = 10 and f = 0.1 as the width of the f-tube yields a 10-class error rate of 1.15% while the best performance using SVMtorch in classification mode is 1.3% (in both cases, we use Gaussian kernels with parameter (J = 1650). An explicit minimization on J.tk requires to add to the standard SVM regression problem the following constraint over the Lagrange multipliers (we use the same notation as in [9]): Yi=l Yi=- l Yi=l Yi=- l Note that we still have the standard regression constraint I: ai = I: ai This has not been implemented yet, as we question the pertinence of the ~; slack variables for XVMs. Experiments with SVMtorch on a variety of tasks where non-zero slacks are required to achieve optimal performance (Reuters, VCI/Forest, VCI/Breast cancer) have not shown significant improvement using the regression mode while we vary the width of the f-tube. Many experiments on SVMs have reported that removing the outliers often gives efficient and sparse solutions. The early stopping heuristics that we have presented for XVMs suggest strategies to avoid learning (or to unlearn) the outliers, and this is the approach we are currently exploring. 7 Concluding Remarks This paper shows that large margin classification on extrapolated data is equivalent to the addition of the minimization of a second external margin to the standard SVM approach. The associated optimization problem is solved efficiently with convex hull distance minimization algorithms. A 1 % error rate is obtained on the MNIST dataset: it is the lowest ever obtained without a-priori knowledge about the data. We are currently trying to identify what other types of dataset show similar gains over SVMs, to determine how dependent XVM performance is on the facts that the data is separable or has invariance properties. We have only explored a few among the many variations the XVM models and algorithms allow, and a justification of why and when they generalize would help model selection. Geometry-based algorithms that handle potential outliers are also under investigation. Learning Theory bounds that would be a function of both the margin and some form of variance of the data would be necessary to predict XVM generalization and allow us to also consider the extrapolation factor 'TJ as an optimization variable. References [1] R. Collobert and S. Bengio. Support vector machines for large-scale regression problems. Technical Report IDIAP-RR-00-17, IDIAP, 2000. [2] C. Cortes and V. Vapnik. Support vector networks. Machine Learning, 20:1- 25, 1995. [3] D. Crisp and C.J.C. Burges. A geometric interpretation of v-SVM classifiers. In Advances in Neural Information Processing Systems 12, S. A. Solla, T. K. Leen, K.-R. Mller, eds, Cambridge, MA, 2000. MIT Press. [4] D. DeCoste and B. Schoelkopf. Training invariant support vector machines. Machine Learning, special issue on Support Vector Machines and Methods, 200l. [5] S.S. Keerthi, S.K. Shevade, C. Bhattacharyya, and K.R.K. Murthy. A fast iterative nearest point algorithm for support vector machine classifier design. IEEE transactions on neural networks, 11(1):124 - 136, jan 2000. [6] A. Kowalczyk. Maximal margin perceptron. In Advances in Large Margin Classifiers, Smola, Bartlett, Schlkopf, and Schuurmans, editors, Cambridge, MA, 2000. MIT Press. [7] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition. proceedings of the IEEE, 86(11), 1998. [8] J. Platt, N. Christianini, and J. Shawe-Taylor. Large margin dags for multiclass classification. In Advances in Neural Information Processing Systems 12, S. A. Solla, T. K. Leen, K.-R. Mller, eds, Cambridge, MA, 2000. MIT Press. [9] V. N. Vapnik. Statistical Learning Theory. John Wiley & Sons, New-York, 1998.
|
2001
|
30
|
2,026
|
Algorithmic Luckiness Ralf Herbrich Microsoft Research Ltd. CB3 OFB Cambridge United Kingdom rherb@microsoft·com Robert C. Williamson Australian National University Canberra 0200 Australia Bob. Williamson@anu.edu.au Abstract In contrast to standard statistical learning theory which studies uniform bounds on the expected error we present a framework that exploits the specific learning algorithm used. Motivated by the luckiness framework [8] we are also able to exploit the serendipity of the training sample. The main difference to previous approaches lies in the complexity measure; rather than covering all hypotheses in a given hypothesis space it is only necessary to cover the functions which could have been learned using the fixed learning algorithm. We show how the resulting framework relates to the VC, luckiness and compression frameworks. Finally, we present an application of this framework to the maximum margin algorithm for linear classifiers which results in a bound that exploits both the margin and the distribution of the data in feature space. 1 Introduction Statistical learning theory is mainly concerned with the study of uniform bounds on the expected error of hypotheses from a given hypothesis space [9, 1]. Such bounds have the appealing feature that they provide performance guarantees for classifiers found by any learning algorithm. However, it has been observed that these bounds tend to be overly pessimistic. One explanation is that only in the case of learning algorithms which minimise the training error it has been proven that uniformity of the bounds is equivalent to studying the learning algorithm's generalisation performance directly. In this paper we present a theoretical framework which aims at directly studying the generalisation error of a learning algorithm rather than taking the detour via the uniform convergence of training errors to expected errors in a given hypothesis space. In addition, our new model of learning allows the exploitation of the fact that we serendipitously observe a training sample which is easy to learn by a given learning algorithm. In that sense, our framework is a descendant of the luckiness framework of Shawe-Taylor et al. [8]. In the present case, the luckiness is a function of a given learning algorithm and a given training sample and characterises the diversity of the algorithms solutions. The notion of luckiness allows us to study given learning algorithms at many different perspectives. For example, the maximum margin algorithm [9] can either been studied via the number of dimensions in feature space, the margin of the classifier learned or the sparsity of the resulting classifier. Our main results are two generalisation error bounds for learning algorithms: one for the zero training error scenario and one agnostic bound (Section 2). We shall demonstrate the usefulness of our new framework by studying its relation to the VC framework, the original luckiness framework and the compression framework of Littlestone and Warmuth [6] (Section 3). Finally, we present an application of the new framework to the maximum margin algorithm for linear classifiers (Section 4). The detailed proofs of our main results can be found in [5]. We denote vectors using bold face, e.g. x = (Xl, ... ,xm ) and the length of this vector by lxi, i.e. Ixl = m. In order to unburden notation we use the shorthand notation Z[i:jJ := (Zi,"" Zj) for i :::; j. Random variables are typeset in sans-serif font. The symbols Px, Ex [f (X)] and IT denote a probability measure over X, the expectation of f (.) over the random draw of its argument X and the indicator function, respectively. The shorthand notation Z(oo) := U;;'=l zm denotes the union of all m- fold Cartesian products of the set Z. For any mEN we define 1m C {I, ... , m } m as the set of all permutations of the numbers 1, ... ,m, 1m := {(il , ... ,im) E {I, ... ,m}m I'v'j f:- k: ij f:- id . Given a 2m- vector i E hm and a sample z E z2m we define Wi : {I, ... , 2m} -+ {I, ... , 2m} by Wi (j) := ij and IIdz) by IIi (z) := (Z7ri(l), ... , Z7ri(2m))' 2 Algorithmic Luckiness Suppose we are given a training sample z = (x, y) E (X x y)m = zm of size mEN independently drawn (iid) from some unknown but fixed distribution PXy = Pz together with a learning algorithm A : Z( 00) -+ yX . For a predefined loss l : y x y -+ [0,1] we would like to investigate the generalisation error Gl [A, z] := Rl [A (z)] - infhEYx Rl [h] of the algorithm where the expected error Rl [h] of his defined by Rl [h] := Exy [l (h (X) ,Y)] . Since infhEYx Rl [h] (which is also known as the Bayes error) is independent of A it suffices to bound Rl [A (z)]. Although we know that for any fixed hypothesis h the training error ~ 1 Rdh,z]:=~ L l(h(xi),Yi) (X i ,Yi) Ez is with high probability (over the random draw of the training sample z E Z(oo)) close to Rl [h], this might no longer be true for the random hypothesis A (z). Hence we would like to state that with only small probability (at most 8), the expected error Rl [A (z)] is larger than the training error HI [A (z), z] plus some sample and algorithm dependent complexity c (A, z, 8), Pzm (Rl [A (Z)] > HI [A (Z), Z] + c (A, Z,8)) < 8. (1) In order to derive such a bound we utilise a modified version of the basic lemma of Vapnik and Chervonenkis [10]. Lemma 1. For all loss functions l : y x y -+ [0,1], all probability measures Pz, all algorithms A and all measurable formulas Y : zm -+ {true, false}, if mc2 > 2 then Pzm (( RdA (Z)] > HdA (Z) , Z] + c) /\ Y (Z)) < 2PZ2m ((HI [A (Z[l:m]) ,Z[(m+l):2mJJ > HI [A (Z[l:mJ) ,Z[l:mJJ + ~) /\ Y (Z[l:m])) . , .I V J(Z) Proof (Sketch). The probability on the r.h.s. is lower bounded by the probability of the conjunction of event on the l.h.s. and Q (z) = Rl [A (Z[l:mj)] Rl [A (Z[l:mj) ,Z(m+1):2m] < ~. Note that this probability is over z E z2m. If we now condition on the first m examples, A (Z[l:mj) is fixed and therefore by an application of Hoeffding's inequality (see, e.g. [1]) and since m€2 > 2 the additional event Q has probability of at least ~ over the random draw of (Zm+1, ... , Z2m). 0 Use of Lemma 1 which is similar to the approach of classical VC analysis reduces the original problem (1) to the problem of studying the deviation of the training errors on the first and second half of a double sample z E z2m of size 2m. It is of utmost importance that the hypothesis A (Z[l:mj) is always learned from the first m examples. Now, in order to fully exploit our assumptions of the mutual independence of the double sample Z E z2m we use a technique known as symmetrisation by permutation: since PZ2~ is a product measure, it has the property that PZ2»> (J (Z)) = PZ2~ (J (ITi (Z))) for any i E hm. Hence, it suffices to bound the probability of permutations Jri such that J (ITi (z)) is true for a given and fixed double sample z. As a consequence thereof, we only need to count the number of different hypotheses that can be learned by A from the first m examples when permuting the double sample. Definition 1 (Algorithmic luckiness). Any function L that maps an algorithm A : Z(oo) -+ yX and a training sample z E Z(oo) to a real value is called an algorithmic luckiness. For all mEN, for any z E z2m , the lucky set HA (L , z) ~ yX is the set of all hypotheses that are learned from the first m examples (Z7ri(1),···, Z7ri(m)) when permuting the whole sample z whilst not decreasing the luckiness, i.e. (2) where Given a fixed loss function 1 : y x y -+ [0,1] the induced loss function set £1 (HA (L,z)) is defined by £1 (HA (L,z)) := {(x,y) r-+ 1 (h(x) ,y) I h E HA (L,z)} . For any luckiness function L and any learning algorithm A , the complexity of the double sample z is the minimal number N1 (T, £1 (HA (L, z)) ,z) of hypotheses h E yX needed to cover £1 (HA (L , z)) at some predefined scale T, i.e. for any hypothesis hE HA (L, z) there exists a h E yX such that (4) To see this note that whenever J (ITi (z)) is true (over the random draw of permutations) then there exists a function h which has a difference in the training errors on the double sample of at least ~ + 2T. By an application of the union bound we see that the number N 1 (T, £1 (HA (L, z)) , z) is of central importance. Hence, if we are able to bound this number over the random draw of the double sample z only using the luckiness on the first m examples we can use this bound in place of the worst case complexity SUPzEZ2~ N1 (T, £1 (HA (L, z)) ,z) as usually done in the VC framework (see [9]). Definition 2 (w- smallness of L). Given an algorithm A : Z ( 00) -+ yX and a loss l : y x y -+ [a, 1] the algorithmic luckiness function Lis w- small at scale T E jR+ if for all mEN, all J E (a, 1] and all Pz PZ2~ (Nl (T, £"1 (1iA (L, Z)), Z) > w (L (A, Z[l:ml) ,l, m, J,T)) < J. , " v S(Z) Note that if the range of l is {a, I} then N 1 (2~ ' £"1 (1iA (L, z)) , z) equals the number of dichotomies on z incurred by £"1 (1iA (L,z)). Theorem 1 (Algorithmic luckiness bounds). Suppose we have a learning algorithm A : Z( oo) -+ yX and an algorithmic luckiness L that is w-small at scale T for a loss function l : y X Y -+ [a, 1]. For any probability measure Pz, any dEN and any J E (a, 1], with probability at least 1 - J over the random draw of the training sample z E zm of size m, if w (L (A, z) ,l, m, J/4, T) :::; 2d then Rz[A (z)] :::; Rz[A (z), z] + ! (d + 10g2 (~) ) + 4T. (5) Furthermore, under the above conditions if the algorithmic luckiness L is wsmall at scale 2~ for a binary loss function l (".) E {a, I} and Rl [A (z), z] = a then (6) Proof (Compressed Sketch). We will only sketch the proof of equation (5) ; the proof of (6) is similar and can be found in [5]. First, we apply Lemma 1 with Y (z) == w (L (A,z) ,l,m,J/4,T) :::; 2d. We now exploit the fact that PZ2~ (J (Z)) :Z 2~ (J (Z) 1\ S (Z) ), +PZ 2~ (J (Z) 1\ ...,S (Z)) v :::: P Z 2~ (S(Z)) J < 4 + PZ2~ (J (Z) I\...,S (Z)) , which follows from Definition 2. Following the above-mentioned argument it suffices to bound the probability of a random permutation III (z) that J (III (z)) 1\ ...,S (III (z)) is true for a fixed double sample z. Noticing that Y (z) 1\ ...,S (z) => Nl (T,£"l (1iA (L,z)) ,z) :::; 2d we see that we only consider swappings Jri for which Nl (T,£"l (1iA (L,IIi (z))) ,IIi (z)) :::; 2d. Thus let us consider such a cover of size not more than 2. By (4) we know that whenever J (IIi (z)) 1\ ...,S (IIi (z)) is true for a swapping i then there exists a hypothesis h E yX in the cover such that Rl [h, (III (z))[(m+1):2ml] - Rl [h, (III (z))[l:ml] > ~ + 2T. Using the union bound and Hoeffding's inequality for a particular choice of PI shows that PI (J (III (z)) 1\ ...,S (III (z))) :::; £ which finalises the proof. D A closer look at (5) and (6) reveals that the essential difference to uniform bounds on the expected error is within the definition of the covering number: rather than covering all hypotheses h in a given hypothesis space 1i ~ yX for a given double sample it suffices to cover all hypotheses that can be learned by a given learning algorithm from the first half when permuting the double sample. Note that the usage of permutations in the definition of (2) is not only a technical matter; it fully exploits all the assumptions made for the training sample, namely the training sample is drawn iid. 3 Relationship to Other Learning Frameworks In this section we present the relationship of algorithmic luckiness to other learning frameworks (see [9, 8, 6] for further details of these frameworks). VC Framework If we consider a binary loss function l (".) E {a, I} and assume that the algorithm A selects functions from a given hypothesis space H ~ yX then L (A, z) = - VCDim (H) is a w- smallluckiness function where ( 1) (2em) -Lo w Lo,l,m,8, 2m :S -Lo . (7) This can easily be seen by noticing that the latter term is an upper bound on maxzEZ2", I{ (l (h (Xl) ,yI) , ... ,l (h (X2m), Y2m)) : h E H}I (see also [9]). Note that this luckiness function neither exploits the particular training sample observed nor the learning algorithm used. Luckiness Framework Firstly, the luckiness framework of Shawe-Taylor et al. [8] only considered binary loss functions l and the zero training error case. In this work, the luckiness £ is a function of hypothesis and training samples and is called wsmall if the probability over the random draw of a 2m sample z that there exists a hypothesis h with w(£(h, (Zl, ... ,zm)), 8) < J'--h (2;" {(X, y) t--+ l (g (x) ,y) 1£ (g, z) ::::: £ (h, Z)}, z), is smaller than 8. Although similar in spirit, the classical luckiness framework does not allow exploitation of the learning algorithm used to the same extent as our new luckiness. In fact, in this framework not only the covering number must be estimable but also the variation of the luckiness £ itself. These differences make it very difficult to formally relate the two frameworks. Compression Framework In the compression framework of Littlestone and Warmuth [6] one considers learning algorithms A which are compression schemes, i.e. A (z) = :R (e (z)) where e (z) selects a subsample z ~ z and :R : Z(oo) -+ yX is a permutation invariant reconstruction function. For this class of learning algorithms, the luckiness L(A,z) = -le(z)1 is w- small where w is given by (7). In order to see this we note that (3) ensures that we only consider permutations 7ri where e (IIi (z)) :S Ie (z)l, i.e. we use not more than -L training examples from z E z2m. As there are exactly e;;) distinct choices of d training examples from 2m examples the result follows by application of Sauer's lemma [9]. Disregarding constants, Theorem 1 gives exactly the same bound as in [6]. 4 A New Margin Bound For Support Vector Machines In this section we study the maximum margin algorithm for linear classifiers, i.e. A : Z(oo) -+ Hcp where Hcp := {x t--+ (¢ (x), w) I wE }C} and ¢ : X -+ }C ~ £~ is known as the feature mapping. Let us assume that l (h (x) ,y) = lO-l (h (x) ,y) := lIyh(x)::;o, Classical VC generalisation error bounds exploit the fact that VCDim (Hcp) = nand (7). In the luckiness framework of Shawe-Taylor et al. [8] it has been shown that we can use fat1i.p h'z (w)) :S h'z (W))-2 (at the price of an extra 10g2 (32m) factor) in place of VCDim (Hcp) where "(z (w) = min(xi,Yi)Ez Yi (¢ (Xi) , w) / Ilwll is known as the margin. Now, the maximum margin algorithm finds the weight vector WMM that maximises "(z (w). It is known that WMM can be written as a linear combination of the ¢ (Xi). For notational convenience, we shall assume that A: Z(oo) -+ 1R(00) maps to the expansion coefficients 0: such that Ilwall = 1 where Wa := 2:1~ 1 (XicfJ(Xi). Our new margin bound follows from the following theorem together with (6). Theorem 2. Let fi (x) be the smallest 10 > 0 such that {cfJ (Xl) , ... , cfJ (Xm) } can be covered by at most i balls of radius less than or equal f. Let f z (w) be d fi d b f ( ) .. Yi (4)(Xi),W) D th l l d e ne y z W . mm(Xi,Yi)Ez 114>(Xi)II.llwll. ror e zero-one oss 0-1 an the maximum margin algorithm A , the luckiness function L(A ) =_ . {. ",,-T .> (fi (X)2:7=1 IA(Z)jl) 2} ,Z mIn ~ E 1'1 ~ _ ( ) , fz W A(z) (8) is w-small at scale 112m w.r.t. the function ( 1) (2em)- 2LO w Lo,l,m,8, 2m = -Lo (9) Proof (Sketch). First we note that by a slight refinement of a theorem of Makovoz [7] we know that for any Z E zm there exists a weight vector w = 2::1 iiicfJ (Xi) such that (10) and a E ]Rm has no more than - L (A, z) non-zero components. Although only WA(z) is of unit length, one can show that (10) implies that (WA(z), wi IIwll) ~ )1- f; (WA(z»). Using equation (10) of [4] this implies that w correctly classifies Z E zm. Consider a fixed double sample Z E z2m and let ko := L (A, (Zl , ... , zm)). By virtue of (3) and the aforementioned argument we only need to consider permutations tri such that there exists a weight vector w = 2:;:1 iijcfJ (Xj) with no more than ko non-zero iij. As there are exactly (2;;) distinct choices of dE {I, ... , ko} training examples from the 2m examples Z there are no more than (2emlko)kO different subsamples to be used in w. For each particular subsample z ~ Z the weight vector w is a member of the class of linear classifiers in a ko (or less) dimensional space. Thus, from (7) it follows that for the given subsample z there are no more (2emlko)kO different dichotomies induced on the double sample Z E z2m. As this holds for any double sample, the theorem is proven. D There are several interesting features about this margin bound. Firstly, observe that 2:;:1 IA (Z)j I is a measure of sparsity of the solution found by the maximum margin algorithm which, in the present case, is combined with margin. Note that for normalised data, i.e. IlcfJ Oil = constant, the two notion of margins coincide, i.e. f z (w) = I Z (w). Secondly, the quantity fi (x) can be considered as a measure of the distribution of the mapped data points in feature space. Note that for all i E N, fi (x) :S 101 (x) :S maxjE{l, ... ,m} IlcfJ (xj)ll. Supposing that the two classconditional probabilities PX1Y=y are highly clustered, 102 (x) will be very small. An extension of this reasoning is useful in the multi-class case; binary maximum margin classifiers are often used to solve multi-class problems [9]. There appears to be also a close relationship of fi (x) with the notion of kernel alignment recently introduced in [3]. Finally, one can use standard entropy number techniques to bound fi (x) in terms of eigenvalues of the inner product matrix or its centred variants. It is worth mentioning that although our aim was to study the maximum margin algorithm the above theorem actually holds for any algorithm whose solution can be represented as a linear combination of the data points. 5 Conclusions In this paper we have introduced a new theoretical framework to study the generalisation error of learning algorithms. In contrast to previous approaches, we considered specific learning algorithms rather than specific hypothesis spaces. We introduced the notion of algorithmic luckiness which allowed us to devise data dependent generalisation error bounds. Thus we were able to relate the compression framework of Littlestone and Warmuth with the VC framework. Furthermore, we presented a new bound for the maximum margin algorithm which not only exploits the margin but also the distribution of the actual training data in feature space. Perhaps the most appealing feature of our margin based bound is that it naturally combines the three factors considered important for generalisation with linear classifiers: margin, sparsity and the distribution of the data. Further research is concentrated on studying Bayesian algorithms and the relation of algorithmic luckiness to the recent findings for stable learning algorithms [2]. Acknowledgements This work was done while RCW was visiting Microsoft Research Cambridge. This work was also partly supported by the Australian Research Council. RH would like to thank Olivier Bousquet for stimulating discussions. References [1) M. Anthony and P. Bartlett. A Theory of Learning in Artificial Neural Networks. Cambridge University Press, 1999. [2) O. Bousquet and A. Elisseeff. Algorithmic stability and generalization performance. In T. K. Leen, T. G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems 13, pages 196- 202. MIT Press, 2001. [3) N. Cristianini, A. Elisseeff, and J. Shawe-Taylor. On optimizing kernel alignment. Technical Report NC2-TR-2001-087, NeuroCOLT, http://www.neurocolt.com. 2001. [4) R. Herbrich and T . Graepel. A PAC-Bayesian margin bound for linear classifiers: Why SVMs work. In T. K. Leen, T. G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems 13, pages 224- 230, Cambridge, MA, 2001. MIT Press. [5) R. Herbrich and R. C. Williamson. Algorithmic luckiness. Technical report, Microsoft Research, 2002. [6) N. Littlestone and M. Warmuth. Relating data compression and learnability. Technical report, University of California Santa Cruz, 1986. [7) Y. Makovoz. Random approximants and neural networks. Journal of Approximation Theory, 85:98- 109, 1996. [8) J. Shawe-Taylor, P. L. Bartlett, R. C. Williamson, and M. Anthony. Structural risk minimization over data-dependent hierarchies. IEEE Transactions on Information Theory, 44(5):1926- 1940, 1998. [9) V. Vapnik. Statistical Learning Theory. John Wiley and Sons, New York, 1998. [10) V. N. Vapnik and A. Y. Chervonenkis. On the uniform convergence of relative frequencies of events to their probabilities. Theory of Probability and its Applications, 16(2):264- 281, 1971.
|
2001
|
31
|
2,027
|
Optimising Synchronisation Times for Mobile Devices Neil D. Lawrence Department of Computer Science, Regent Court, 211 Portobello Road, Sheffield, Sl 4DP, U.K. neil~dcs.shef . ac.uk Antony 1. T. Rowstron Christopher M . Bishop Michael J. Taylor Microsoft Research 7 J. J. Thomson A venue Cambridge, CB3 OFB, U.K. {antr,cmbishop,mitaylor}~microsoft.com Abstract With the increasing number of users of mobile computing devices (e.g. personal digital assistants) and the advent of third generation mobile phones, wireless communications are becoming increasingly important. Many applications rely on the device maintaining a replica of a data-structure which is stored on a server, for example news databases, calendars and e-mail. ill this paper we explore the question of the optimal strategy for synchronising such replicas. We utilise probabilistic models to represent how the data-structures evolve and to model user behaviour. We then formulate objective functions which can be minimised with respect to the synchronisation timings. We demonstrate, using two real world data-sets, that a user can obtain more up-to-date information using our approach. 1 Introduction As the available bandwidth for wireless devices increases, new challenges are presented in the utilisation of such bandwidth. Given that always up connections are generally considered infeasible an important area of research in mobile devices is the development of intelligent strategies for communicating between mobile devices and servers. ill this paper we consider the scenario where we are interested in maintaining, on a personal digital assistant (PDA) with wireless access, an up-to-date replica of some, perhaps disparate, data-structures which are evolving in time. The objective is to make sure our replica is not 'stale'. We will consider a limited number of connections or synchronisations. Each synchronisation involves a reconciliation between the replica on the mobile device and the data-structures of interest on the server. Later in the paper we shall examine two typical examples of such an application,an internet news database and a user's e-mail messages. Currently the typical strategy! for performing such reconciliations is to synchronise every M minutes, lSee, for example, AvantGo http://vvv. avantgo. com. where M is a constant, we will call this strategy the uniformly-spaced strategy. We will make the timings of the synchronisations adaptable by developing a cost function that can be optimised with respect to the timings, thereby improving system performance. 2 Cost Function We wish to minimise the staleness of the replica, where we define staleness as the time between an update of a portion of the data-structure on the server and the time of the synchronisation of that update with the PDA. For simplicity we shall assume that each time the PDA synchronises all the outstanding updates are transferred. Thus, after synchronisation the replica on the mobile device is consistent with the master copy on the server. Therefore, if skis the time of the kth synchronisation in a day, and updates to the data-structure occur at times Uj then the average staleness of the updates transferred during synchronisation Sk will be (1) As well as staleness, we may be interested in optimising other criteria. For example, mobile phone companies may seek to equalise demand across the network by introducing time varying costs for the synchronisations, c(t). Additionally one could argue that there is little point in keeping the replica fresh during periods when the user is unlikely to check his PDA, for example when he or she is sleeping. We might therefore want to minimise the time between the user's examination of the PDA and the last synchronisation. If the user looks at the PDA at times ai then we can express this as (2) Given the timings Uj and ai, the call cost schedule c(t) and K synchronisations, the total cost function may now be written K C = L (-aFk + fJSk + C(Sk)) ' (3) k=l where a and fJ are constants with units of ~~~:y which express the relative importance of the separate parts of the cost function. Unfortunately, of course, whilst we are likely to have knowledge of the call cost schedule, c(t), we won't know the true timings {Uj} and {ai} and the cost function will be a priori incomputable. If, though, we have historic data2 relating to these times, we can seek to make progress by modelling these timings probabilistically. Then, rather than minimising the actual cost function, we can look to minimise the expectation of the cost function under these probabilistic models. 3 Expected Cost There are several different possibilities for our modelling strategy. A sensible assumption is that there is independence between different parts of the data-structure (i.e. e-mail and business news can be modelled separately), however, there may be dependencies between update times which occur within the same part. The 2When modelling user ru:cess times, if historic data is not available, models could also be constructed by querying the user about their likely ru:tivities. periodicity of the data may be something we can take advantage of, but any nonstationarity in the data may cause problems. There are various model classes we could consider; for this work however, we restrict ourselves to stationary models, and ones in which updates arrive independently and in a periodic fashion . . Let us take T to be the largest period of oscillation in the data arrivals, for a particular portion of a data-structure. We model this portion with a probability distribution, Pu(t). Naturally more than one update may occur in that interval, therefore our probability distribution really specifies a distribution over time given one that one update (or user access) has occurred. To fully specify the model we also are required to store the expected number of updates, Ju , (or accesses, Ja ) that occur in that interval. The expected value of Sk may now be written, (4) where Op(:v) is an expectation under the distribution p(x), Au(t) = JuPu(t) can be viewed as the rate at which updates are occurring and So = SK - T. We can model the user access times, ai, in a similar manner, which leads us to the expected value of the freshness, (Fk)Pa(t) = J:k k +l Aa(t)(t - sk)dt, where Aa(t) = JaPa(t) The overall expected cost, which we will utilise as our objective function, may therefore be written K (C) = L (Sk)p" - (Fk)Pa + C(Sk)) . (5) k=l 3.1 Probabilistic Models. We now have an objective function which is a function of the variables we wish to optimise, the synchronisation times, but whilst we have mentioned some characteristics of the models Pu(t) and Pa(t) we have not yet fully specified their form. We have decreed that the models should be periodic and that they may consider each datum to occur independently. In effect we are modelling data which is mapped to a circle. Various options are available for handling such models; for this work, we constrain our investigations to kernel density estimates (KDE). In order to maintain periodicity, we must select a basis function for our KDE which represents a distribution on a circle, one simple way of achieving this aim is to wrap a distribution that is defined along a line to the circle (Mardia, 1972). A traditional density which represents a distribution on the line, p(t), may be wrapped around a circle of circumference T to give us a distribution defined on the circle, p( 0), where 0 = t mod T. This means a basis function with its centre at T - 8, that will typically have probability mass when u > T, wraps around to maintain a continuous density at T. The wrapped Gaussian distribution3 that we make use of takes the form (6) The final kernel density estimate thus consists of mapping the data points tn -t On 3In practice we must approximate the wrapped distribution by restricting the number of terms in the sum. Thu Fri Sat Sun Thu Fri Sat Sun "-' ~60 .§ ~40 "-' .S 1)520 ~ OJ H Co> OJ ""0 ~ -20 Figure 1: Left: part of the KDE developed for the business category together with a piecewise constant approximation. Middle: the same portion of the KDE for the FA Carling Premiership data. Right: percent decrease in staleness vs number of synchronisations per day for e-mail data. and obtaining a distribution 1 N p(()) = N L W N(()I()n, (}"2), n=l (7) where N is the number of data-points and the width parameters, (}", can be set through cross validation. Models of this type may be made use of for both Pu(t) and Pa(t). 3.2 Incorporating Prior Knowledge. The underlying component frequencies of the data will clearly be more complex than simply a weekly or daily basis. Ideally we should be looking to incorporate as much of our prior knowledge about these component frequencies as possible. IT we were modelling financial market's news, for example, we would expect weekdays to have similar characteristics to each other, but differing characteristics from the weekend. For this work, we considered four different scenarios of this type. For the first scenario, we took T = 1 day and placed no other constraints on the model. For the second we considered the longest period to be one week, T = 1 week, and placed no further constraints on the model. For the remaining two though we also considered T to be one week, but we implemented further assumptions about the nature of the data. Firstly we split the data into weekdays and weekends. We then modelled these two categories separately, making sure that we maintained a continuous function for the whole week by wrapping basis functions between weekdays and weekends. Secondly we split the datainto weekdays, Saturdays and Sundays, modelling each category separately and again wrapping basis functions across the days. 3.3 Model Selection. To select the basis function widths, and to determine which periodicity assumption best matched the data, we · utilised ten fold cross validation. For each different periodicity we used cross validation to first select the basis function width. We then compared the average likelihood across the ten validation sets, selecting the periodicity with the highest associated value. 4 Optimising the Synchronisation Times Given that our user model, Pa(t), and our data model, Pu(t) will be a KDE based on wrapped Gaussians, we should be in a position to compute the required integrals '" C/J t{360 t{360 60 >=1 >=1 Q) Q) x <il X ]40 + + + .....,40 X !I 40 j C/J + )I( C/J X ~ C/J .S .S C/J ! i Q) ~ I~! >=1 S520 S520 .$20 ~ ~ ~ Q) Q) ....., !-< !-< C/J U U .S X Q) Q) "0 4 "0 2 2 12 24 X Q) ~ ~ C/J X X ~ X -20 -20 ~20 X X Figure 2: Results from the news database tests. Left: u xx Q) X February /March based models tested on April. Middle: "0 ~40 X X xx March/ April testing on May. Right: April/May testing on X X June. The results are in the form of box plots. The lower xXx line of the box represents the 25th percentile of the data, the -60 upper line the 75th percentile and the central line the median. The 'whiskers' represent the maximum extent of the data up to 1.5 x (75th percentile - 25th percentile). Data which lies outside the whiskers is marked with crosses. in (5) and evaluate our objective function and derivatives thereof. First though, we must give some attention to the target application for the algorithm. A known disadvantage of the standard kernel density estimate is the high storage requirements of the end model. The model requires that N floating point numbers must be stored, where N is the quantity of training data. Secondly, integrating across the cost function results in an objective function which is dependent on a large number of evaluations of the cumulative Gaussian distribution. Given that we envisage that such optimisations could be occurring within a PDA or mobile phone, it would seem prudent to seek a simpler approach to the required minimisation. An alternative approach that we explored is to approximate the given distributions with a functional form which is more amenable to the integration. For example, a piecewise constant approximation to the KDE simplifies the integral considerably. It leads to a piecewise constant approximation for Aa(t) and Au(t). Integration over which simply leads to a piecewise linear function which may be computed in a straightforward manner. Gradients may also be computed. We chose to reduce the optimisation to a series of one-dimensional line minimisations. This can be achieved in the following manner. First, note that the objective function, as a function of a particular synchronisation time Sk, may be written: (8) In other words, each synchronisation is only dependent on that of its neighbours. We may therefore perform the optimisation by visiting each synchronisation time, Sk, in a random order and optimising its position between its neighbours, which involves a one dimensional line minimisation of (8). This process, which is guaranteed to find a (local) minimum in our objective function, may be repeated until convergence. 5 Results In this section we mainly explore the effectiveness of modelling the data-structures of interest. We will briefly touch upon the utility of modelling the cost evolution and user accesses in Section 5.2 but we leave a more detailed exploration of this area to later works. 5.1 Modelling Data Structures To determine the effectiveness of our approach, we utilised two different sources of data: a news web-site and e-mail on a mail server. The news database data-set was collected from the BBC News web site4 . This site maintains a database of articles which are categorised according to subject, for example, UK News, Business News, Motorsport etc .. We had six months of data from February to July 2000 for 24 categories of the database. We modelled the data by decomposing it into the different categories and modelling each separately. This allowed us to explore the periodicity of each category independently. This is a sensible approach given that the nature of the data varies considerably across the categories. . Two extreme examples are Business news and FA Carling Premiership news5, Figure 1. Business news predominantly arrives during the week whereas FA Carling Premiership news arrives typically just after soccer games finish on a Saturday. Business news was best modelled on a Weekday/Weekend basis, and FA Carling Premiership news was best modelled on a Weekday /Saturday /Sunday basis. To evaluate the feasibility of our approach, we selected three consecutive months of data. The inference step consisted of constructing our models on data from the first two months. To restrict our investigations to the nature of the data evolution only, user access frequency was taken to be uniform and cost of connection was considered to be constant. For the decision step we considered 1 to 24 synchronisations a day. The synchronisation times were optimised for each category separately, they were initialised with a uniformly-spaced strategy, optimisation of the timings then proceeded as described in Section 4. The staleness associated with these timings was then computed for the third month. This value was compared with the staleness resulting from the uniformly-spaced strategy containing the same number of synchronisations6 . The percentage decrease in staleness is shown in figures 2 and 3 in the form of box-plots. 60 x x x 40 x x 20 12 -20 X X Xx 00 x x ~XX ~ g}40 ~,(¢<: ~ § ~x x '"@ ++ \xx t260 + )( .,« .S + x x -120 -140 -160 -180 -200 x + x + + + + + + x + + + + + + + + Figure 3: May/June based models tested on July. + signifies the FA Carling Premiership Generally an improvement in performance is observed, Stream. however, we note that in Figure 3 the performance for several categories is extremely 4http://news.bbc.co.uk. 5The FA Carling Premiership is England's premier division soccer. 6The uniformly-spaced strategy's staleness varies with the timing of the first of the K synchronisations. This figure was therefore an average of the staleness from all possible starting points taken at five minute intervals. poor. In particular the FA Carling Premiership stream in Figure 3. The poor performance is caused by the soccer season ending in May. As a result relatively few articles are written in July, most of them concerning player transfer speculation, and the timing of those articles is very different from those in May. In other words the data evolves in a non-stationary manner which we have not modelled. The other poor performers are also sports related categories exhibiting non-stationarities. The e-mail data-set was collected by examining the logs of e-mail arrival times for 9 researchers from Microsoft's Cambridge research lab. This data was collected for January and February 2001. We utilised the January data to build the probabilistic models and the February data to evaluate the average reduction in staleness. Figure 1 shows the results obtained. In practice, a user is more likely to be interested in a combination of different categories of data. Perhaps several different streams of news and his e-mail. Therefore, to recreate a more realistic situation where a user has a combination of interests, we also collected e-mail arrivals for three users from February, March and April 2000. We randomly generated user profiles by sampling, without replacement, five categories from the available twenty-seven, rejecting samples where more than one e-mail stream was selected. We then modelled the users' interests by constructing an unweighted mixture of the five categories and proceeded to optimise the synchronisation times based on this model. This was performed one hundred times. The average staleness for the different numbers of synchronisations per day is shown in Figure 4. Note that the performance for the combined categories is worse than it is for each individually. This is to be expected as the entropy of the combined model will always be greater than that of its constituents, we therefore have less information about arrival times, and as a result there are less gains to be made over the uniformlyspaced strategy7. 5.2 Affect of Cost and User Model In the previous sections we focussed on modelling the evolution of the databases. Here we now briefly turn our attention to the other portions of the system, user behaviour and connection cost. For this preliminary study, it proved difficult to obtain high quality data representing user access times. We therefore artificially generated a model which represents a user who accesses there device frequently at breakfast, lunchtime and during the evening, and rarely at night. Figure 4 simply shows the user model, Pa(t), along with the result of optimising the cost function for uniform data arrivals and fixed cost under this user model. Note how synchronisation times are suggested just before high periods of user activity are about to occur. Also in Figure 4 is the effect of a varying cost, c(t), under uniform Pa(t) and Pa(t). Currently most mobile internet access providers appear to be charging a flat fee for call costs (typically in the U.K. about 15 cents per minute). However, when demand on their systems rise they may wish to incorporate a varying cost to flatten peak demands. This cost could be an actual cost for the user, or alternatively a 'shadow price' specified by service provider for controlling demand (Kelly, 2000). We give a simple example of such a call cost in Figure 4. For this we considered user access and data update rates to be constant. Note how the times move away from periods of high cost. 7The uniformly-spaced strategy can be shown to be optimal when the entropy of the underlying distribution is maximised (a uniform distribution across the interval). "-' gj 60 >=1 Cl) 0.3 7Lo "-' X 0.25 1200 .S ....., 0.2 ~ 900 ~20 0.15 U '" Cl) :::::::: 600 .... 0.1 '" U U Cl) 0.05 300 "'0 ~ X 0 00:00 08:00 00:00 08:00 16:00 00:00 -20 Figure 4: Left: change in synchronisation times for variable user access rates. x shows the initialisation points, + the end points. Middle: change in synchronisation times for a variable cost. Right: performance improvements for the combination of news and e-mail. 6 Discussion The optimisation strategy we suggest could be sensitive to local minima, we did not try a range of different initialisations to explore this phenomena. However, by initialising with the uniformly-spaced strategy we ensured that we increased the objective function relative to the standard strategy. The month of July showed how a non-stationarity in the data structure can dramatically affect our performance. We are currently exploring on-line Bayesian models which we hope will track such non-stationarities. The system we have explored in this work assumed that the data replicated on the mobile device was only modified on the server. A more general problem is that of mutable replicas where the data may be modified on the server or the client. Typical applications of such technology include mobile databases, where sales personnel modify portions of the database whilst on the road, and a calendar application on a PDA, where the user adds appointments on the PDA. Finally there are many other applications of this type of technology beyond mobile devices. Web crawlers need to estimate when pages are modified to maintain a representative cache (eho and Garcia-Molina, 2000). Proxy servers could also be made to intelligent maintain their caches of web-pages up-to-date (Willis and Mikhailov, 1999; Wolman et al., 1999) . References Cho, J. and H. Garcia-Molina (2000). Synchronizing a database to improve freshness. In Proceedings 2000 ACM International Conference on Management of Data (SIGMOD). Kelly, F. P. (2000). Models for a self-managed internet. Philosophical Transactions of the Royal Society A358, 2335-2348. Mardia, K. V. (1972). Statistics of Directional Data. London: Academic Press. Rowstron, A. 1. T., N. D. Lawrence, and C. M. Bishop (2001). Probabilistic modelling of replica divergence. In Proceedings of the 8th Workshop on Hot Topics in Operating Systems HOTOS (VIII). Willis, C. E. and M. Mikhailov (1999). Towards a better understanding of web resources and server responses for improved caching. In Proceedings of the 8th International World Wide Web Conference, pp. 153-165. Wolman, A., G. M. Voelker, N. Sharma, N. Cardwell, A. Karlin, and H. M. Levy (1999). On the scale and performance of co-operative web proxy caching. In 17th ACM Symposium Operating System Principles (SOSP'99), pp. 16-3l. Yu, H. and A. Vahdat (2000). Design and evaluation of a continuous consistency model for replicated services. In 4th Symposium on Operating System Design and Implementation (OSDI).
|
2001
|
32
|
2,028
|
Spike timing and the coding of naturalistic sounds in a central auditory area of songbirds Brian D. Wright, Kamal Sen, William Bialek and Allison J. Doupe Sloan–Swartz Center for Theoretical Neurobiology Departments of Physiology and Psychiatry University of California at San Francisco, San Francisco, California 94143–0444 NEC Research Institute, 4 Independence Way, Princeton, New Jersey 08540 Department of Physics, Princeton University, Princeton, New Jersey 08544 bdwright/kamal/ajd @phy.ucsf.edu, wbialek@princeton.edu Abstract In nature, animals encounter high dimensional sensory stimuli that have complex statistical and dynamical structure. Attempts to study the neural coding of these natural signals face challenges both in the selection of the signal ensemble and in the analysis of the resulting neural responses. For zebra finches, naturalistic stimuli can be defined as sounds that they encounter in a colony of conspecific birds. We assembled an ensemble of these sounds by recording groups of 10-40 zebra finches, and then analyzed the response of single neurons in the songbird central auditory area (field L) to continuous playback of long segments from this ensemble. Following methods developed in the fly visual system, we measured the information that spike trains provide about the acoustic stimulus without any assumptions about which features of the stimulus are relevant. Preliminary results indicate that large amounts of information are carried by spike timing, with roughly half of the information accessible only at time resolutions better than 10 ms; additional information is still being revealed as time resolution is improved to 2 ms. Information can be decomposed into that carried by the locking of individual spikes to the stimulus (or modulations of spike rate) vs. that carried by timing in spike patterns. Initial results show that in field L, temporal patterns give at least
% extra information. Thus, single central auditory neurons can provide an informative representation of naturalistic sounds, in which spike timing may play a significant role. 1 Introduction Nearly fifty years ago, Barlow [1] and Attneave [2] suggested that the brain may construct a neural code that provides an efficient representation for the sensory stimuli that occur in the natural world. Slightly earlier, MacKay and McCulloch [3] emphasized that neurons that could make use of spike timing—rather than a coarser “rate code”—would have available a vastly larger capacity to convey information, although they left open the question of whether this capacity is used efficiently. Theories for timing codes and efficient representation have been discussed extensively, but the evidence for these attractive ideas remains tenuous. A real attack on these issues requires (at least) that we actually measure the information content and efficiency of the neural code under stimulus conditions that approximate the natural ones. In practice, constructing an ensemble of “natural” stimuli inevitably involves compromises, and the responses to such complex dynamic signals can be very difficult to analyze. At present the clearest evidence on efficiency and timing in the coding of naturalistic stimuli comes from central invertebrate neurons [4, 5] and from the sensory periphery [6, 7] and thalamus [8, 9] of vertebrates. The situation for central vertebrate brain areas is much less clear. Here we use the songbird auditory system as an accessible test case for these ideas. The set of songbird telencephalic auditory areas known as the field L complex is analogous to mammalian auditory cortex and contains neurons that are strongly driven by natural sounds, including the songs of birds of the same species (conspecifics) [10, 11, 12, 13]. We record from the zebra finch field L, using naturalistic stimuli that consist of recordings from groups of 10-40 conspecific birds. We find that single neurons in field L show robust and well modulated responses to playback of long segments from this song ensemble, and that we are able to maintain recordings of sufficient stability to collect the large data sets that are required for a model independent information theoretic analysis. Here we give a preliminary account of our experiments. 2 A naturalistic ensemble Auditory processing of complex sounds is critical for perception and communication in many species, including humans, but surprisingly little is known about how high level brain areas accomplish this task. Songbirds provide a useful model for tackling this issue, because each bird within a species produces a complex individualized acoustic signal known as a song, which reflects some innate information about the species’ song as well as information learned from a “tutor” in early life. In addition to learning their own song, birds use the acoustic information in songs of others to identify mates and group members, to discriminate neighbors from intruders, and to control their living space [14]. Consistent with how ethologically critical these functions are, songbirds have a large number of forebrain auditory areas with strong and increasingly specialized responses to songs [11, 15, 16]. The combination of a rich set of behaviorally relevant stimuli and a series of high-level auditory areas responsive to those sounds provides an opportunity to reveal general principles of central neural encoding of complex sensory stimuli. Many prior studies have chosen to study neural responses to individual songs or altered versions thereof. In order to make the sounds studied increasingly complex and natural, we have made recordings of the sounds encountered by birds in our colony of zebra finches. To generate the sound ensemble that was used in this study we first created long records of the vocalizations of groups of 10-40 zebra finches in a soundproof acoustic chamber with a directional microphone above the bird cages. The group of birds generated a wide variety of vocalizations including songs and a variety of different types of calls. Segments of these sounds were then joined to create the sounds presented in the experiment. One of the segments that was presented (
sec) was repeated in alternation with different segments. We recorded the neural responses in field L of one of the birds from the group to the ensemble of natural sounds played back through a speaker, at an intensity approximately equal to that in the colony recording. This bird was lightly anesthetized with urethane. We used a single electrode to record the neural response waveforms and sorted single units offline. Further details concerning experimental techniques can be found in Ref. [13]. 50 Hz 500 ms A B C D Figure 1: A. Spike raster of 4 seconds of the responses of a single neuron in field L to a 30 second segment of a natural sound ensemble of zebra finch sounds. The stimulus was repeated 80 times. B. Peri-stimulus time histogram (PSTH) with 1 ms bins. C. Sound pressure waveform for the natural sound ensemble. D. Blowup of segment shown in the box in A. The scale bar is 50 ms. 3 Information in spike sequences The auditory telencephalon of birds consists of a set of areas known as the field L complex, which receive input from the auditory thalamus and project to increasingly selective auditory areas such as NCM, cHV and NIf [12, 17] and ultimately to the brain areas specialized for the bird’s own song. Field L neurons respond to simple stimuli such as tone bursts, and are organized in a roughly tonotopic fashion [18], but also respond robustly to many complex sounds, including songs. Figure 1 shows 4 seconds of the responses of a cell in field L to repeated presentations of a 30 sec segment from the natural ensemble described above. Averaging over presentations, we see that spike rates are well modulated. Looking at the responses on a finer time resolution we see that aspects of the spike train are reproducible on at least a
ms time scale. This encourages us to measure the information content of these responses over a range of time scales, down to millisecond resolution. Our approach to estimating the information content of spike trains follows Ref. [4]. At some time (defined relative to the repeating stimulus) we open a window of size to look at the response. Within this window we discretize the spike arrival times with resolution so that the response becomes a “word” with letters. If the time resolution is very small, the allowed letters are only 1 and 0, but as becomes larger one must keep track of multiple spikes within each bin. Examining the whole experiment, we sample 0 0.01 0.02 0.03 0.04 0.05 0.06 0 5 10 15 20 25 30 35 40 1/Nrepeats Information Rate (bits/sec) Total Entropy Noise Entropy Mutual Info Figure 2: Mutual information rate for the spike train is shown as a function of data size for ms and ms. the probability distribution of words, , and the entropy of this distribution sets the capacity of the code to convey information about the stimulus:
!" #%$'&)(+* (1) where the notation reminds us that the entropy depends both on the size of the words that we consider and on the time resolution with which we classify the responses. We can think of this entropy as measuring the size of the neuron’s vocabulary. Because the whole experiment contributes to defining the vocabulary size, estimating the distribution , and hence the total entropy is not significantly limited by the problems of finite sample size. This can be seen in Fig. 2 in the stability of the total entropy with changing the number of repeats used in the analysis. Here we show the total entropy as a rate in bits per second by dividing the entropy by the time window . While the capacity of the code is limited by the total entropy, to convey information particular words in the vocabulary must be associated, more or less reliably, with particular stimulus features. If we look at one time relative to the (long) stimulus, and examine the words generated on repeated presentations, we sample the conditional distribution -/. . This distribution has an entropy that quantifies the noise in the response at time , and averaging over all times we obtain the average noise entropy, 10 )2 34 5 7689 "/. :'; "/. =<,>?#%$'&)(+* (2) where > indicates a time average (in general, denotes an average over the variable ). Technically, the above average should be an average over stimuli , however, for a sufficiently long and rich stimulus, the ensemble average over can be replaced by a time average. For the noise entropy, the problem of sampling is much more severe, since each distribution /. is estimated from a number of examples given by the number of repeats. Still, as shown in Fig. 2, we find that the dependence of our estimate on sample size is simple and regular; specifically, we find 5
4 4 3 5 4 4 3 (3) This is what we expect for any entropy estimate if the distribution is well sampled, and if we make stronger assumptions about the sampling process (independence of trials etc.) we can even estimate the correction coefficient [19]. In systems where much larger data sets are available this extrapolation procedure has been checked, and the observation of a good fit to Eq. (3) is a strong indication that larger sample sizes will be consistent with 5 ; further, this extrapolation can be tested against bounds on the entropy that are derived from more robust quantities [4]. Most importantly, failure to observe Eq. (3) means that we are in a regime where sampling is not sufficient to draw reliable conclusions without more sophisticated arguments, and we exclude these regions of and from our discussion. Ideally, to measure the spike train total and noise entropy rates, we want to go to the limit of infinite word duration. A true entropy is extensive, which here means that it grows linearly with spike train word duration , so that the entropy rate is constant. For finite word duration however, words sampled at neighboring times will have correlations between them due, in part, to correlations in the stimulus (for birdsong these stimulus autocorrelation time scales can extend up to
ms). Since the word samples are not completely independent, the raw entropy rate is an overestimate of the true entropy rate. The effect is larger for smaller word duration and the leading dependence of the raw estimate is 5 * (4) where and we have already taken the infinite data size limit. We cannot directly take the large limit, since for large word lengths we eventually reach a data sampling limit beyond which we are unable to reliably compute the word distributions. On the other hand, if there is a range of for which the distributions are sufficiently well sampled, the behavior in Eq. (4) should be observed and can be used to extrapolate to infinite word size [4]. We have checked that our data shows this behavior and that it sets in for word sizes below the limit where the data sampling problem occurs. For example, in the case of the noise entropy, for ms, it applies for below the limit of ms (above this we run into sampling problems). The total entropy estimate is nearly perfectly extensive. Finally, we combine estimates of total and noise entropies to obtain the information that words carry about the sensory stimulus,
5 0 2 3 4 5 #%$'&)( (5) Figure 2 shows the total and noise entropy rates as well as the mutual information rate for a time window ms and time resolution ms. The error bars on the raw entropy and information rates were estimated to be approximately ! bits/sec using a simple bootstrap procedure over the repeated trials. The extrapolation to infinite data size is shown for the mutual information rate estimate (error bars in the extrapolated values will be "#! bits/sec) and is consistent with the prediction of Eq. (3). Since the total entropy is nearly extensive and the noise entropy rate decreases with word duration due to subextensive corrections as described above, the mutual information rate shown in Fig. 2 grows with word duration. We find that there is an upward change in the mutual information 0 5 10 15 20 25 30 35 1.5 2 2.5 3 3.5 4 4.5 5 ∆τ (ms) Information Rate (bits/sec) Independent Events Spike Train Figure 3: Information rates for the spike train ( ms) and single spike events as a function of time resolution of the spike rasters, corrected for finite data size effects. rate (computed at ms and ms) of
%, in the large limit. For simplicity in the following, we shall look at a fixed word duration ms that is in the well-sampled region for all time resolutions considered. The mutual information rate measures the rate at which the spike train removes uncertainty about the stimulus. However, the mutual information estimate does not depend on identifying either the relevant features of the stimulus or the relevant features of the response, which is crucial in analyzing the response to such complex stimuli. In this sense, our estimates of information transmission and efficiency are independent of any model for the code, and provide a benchmark against which such models could be tested. One way to look at the information results is to fix our time window and ask what happens as we change our time resolution . When , the “word” describing the response is nothing but the number of spikes in the window, so we have a rate or counting code. As we decrease , we gradually distinguish more and more detail in the arrangement of spikes in the window. We chose a range of values from ms in our analyses to cover previously observed response windows for field L neurons and to probe the behaviorally relevant time scale (
ms) of individual song syllables or notes. For ms, we show the results (extrapolated to infinite data size) in the upper curve of Fig. 3. The spike train mutual information shows a clear increase as the timing resolution is improved. In addition, Fig. 3 shows that roughly half of the information is accessible at time resolutions better than ms and additional information is still being revealed as time resolution is improved to 2 ms. 4 Information in rate modulation Knowing the mutual information between the stimulus and the spike train (defined in the window ), we would like to ask whether this can be accounted for by the information in single spike events or whether there is some additional information conveyed by the patterns of spikes. In the latter case, we have precisely what we mean by a temporal or timing code: there is information beyond that attributable to the probability of single spike events occurring at time relative to the onset of the stimulus. By event at time , we mean that the event occurs between time and time , where is the resolution at which we are looking at the spike train. This probability is simply proportional to the firing rate (or peri-stimulus time histogram (PSTH)) at time normalized by the mean firing rate . Specifically if the duration of each repeated trial is 4 4 we have spk @ . 4 4 * (6) where denotes the stimulus history ( " ). The probability of a spike event at , a priori of knowing the stimulus history, is flat: spk @ 4 4 . Thus, the mutual information between the stimulus and the single spike events is [20]:
(%$ spk @ spk @ .
6 <-> #%$'&)(+* (7) where is the PSTH binned to resolution and the stimulus average in the first expression is replaced by a time average in the second (as discussed in the calculation of the noise entropy in spike train words in the previous section). We find that this information is approximately bit for ms. Supposing that the individual spike events are independent (i.e. no intrinsic spike train correlations), the information rate in single spike events is obtained by multiplying the mutual information per spike (Eq. 7) by the mean firing rate of the neuron (
Hz). This gives an upper bound to the single spike event contribution to the information rate and is shown in the lower curve of Fig. 3 (error bars are again " ! bits/sec). Comparing with the spike train information (upper curve), we see that at a resolution of 8 ms, there is at least
% of the total information in the spike train that cannot be attributable to single spike events. Thus there is some pattern of spikes that is contributing synergistically to the mutual information. The fact discussed, in the previous section, that the spike train information rate grows subextensively with the the word duration out to the point where data sampling becomes problematic is further confirmation of the synergy from spike patterns. Thus we have shown model-independent evidence for a temporal code in the neural responses. 5 Conclusion Until now, few experiments on neural responses in high level, central vertebrate brain areas have measured the information that these responses provide about dynamic, naturalistic sensory signals. As emphasized in earlier work on invertebrate systems, information theoretic approaches have the advantage that they require no assumptions about the features of the stimulus to which neurons respond. Using this method in the songbird auditory forebrain, we found that patterns of spikes seem to be special events in the neural code of these neurons, since they carry more information than expected by adding up the contributions of individual spikes. It remains to be determined what these spike patterns are, what stimulus features they may encode, and what mechanisms may be responsible for reading such codes at even higher levels of processing. Acknowledgments Work at UCSF was supported by grants from the NIH (NS34835) and the Sloan-Swartz Center for Theoretical Neurobiology. BDW and KS supported by NRSA grants from the NIDCD. We thank Katrin Schenk and Robert Liu for useful discussions. References 1. Barlow, H.B. (1961). Possible principles underlying the transformation of sensory messages. In Sensory Communication, W.A. Rosenblith, ed., pp. 217–234 (MIT Press, Cambridge, MA). 2. Attneave, F. (1954). Some informational aspects of visual perception. Psychol. Rev. 61, 183–193. 3. MacKay, D. and McCulloch, W.S. (1952). The limiting information capacity of a neuronal link. Bull. Math. Biophys. 14, 127–135. 4. Strong, S.P., Koberle, R., de Ruyter van Steveninck, R. and Bialek, W. (1998). Entropy and information in neural spike trains, Phys. Rev. Lett. 80, 197–200. 5. Lewen, G.D., Bialek, W. and de Ruyter van Steveninck, R.R. (2001). Neural coding of naturalistic motion stimuli. Network 12, 317–329. 6. Rieke, F., Bodnar, D.A. and Bialek, W. (1995). Naturalistic stimuli increase the rate and efficiency of information transmission by primary auditory afferents. Proc. R. Soc. Lond. B 262, 259–265. 7. Berry II, M.J., Warland, D.K. and Meister, M. (1997). The structure and precision of retinal spike trains. Proc. Nat. Acad. Sci. (USA) 94, 5411–5416. 8. Reinagel, P. and Reid, R.C. (2000). Temporal coding of visual information in the thalamus. J. Neurosci. 20, 5392–5400. 9. Liu, R.C., Tzonev, S., Rebrik, S. and Miller, K.D. (2001). Variability and information in a neural code of the cat lateral geniculate nucleus. J. Neurophysiol. 86, 2789–2806. 10. Scheich, H., Langner, G. and Bonke, D. (1979). Responsiveness of units in the auditory neostriatum of the guinea fowl (Numida meleagris) to species-specific calls and synthetic stimuli II. Discrimination of Iambus-Like Calls. J. Comp. Physiol. A 132, 257–276. 11. Lewicki, M.S. and Arthur, B.J. (1996). Hierarchical organization of auditory temporal context sensitivity. J. Neurosci. 16(21), 6987–6998. 12. Janata, P. and Margoliash, D. (1999). Gradual emergence of song selectivity in sensorimotor structures of the male zebra finch song system. J. Neurosci. 19(12), 5108–5118. 13. Theunissen, F.E., Sen, K. and Doupe, A.J. (2000). Spectral temporal receptive fields of nonlinear auditory neurons obtained using natural sounds. J. Neurosci. 20(6), 2315–2331. 14. Searcy, W.A. and Nowicki, S. (1999). In The Design of Animal Communication, M.D. Hauser and M. Konishi, eds., pp. 577–595 (MIT Press, Cambridge, MA). 15. Margoliash, D. (1983). Acoustic parameters underlying the responses of song-specific neurons in the white-crowned sparrow. J. Neurosci. 3(5), 1039–1057. 16. Sen, K., Theunissen, F.E. and Doupe, A.J. (2001). Feature analysis of natural sounds in the songbird auditory forebrain. J. Neurophysiol. 86, 1445–1458. 17. Stripling, R., Kruse, A.A. and Clayton, D.F. (2001). Development of song responses in the zebra finch caudomedial neostriatum: role of genomic and electrophysiological activities. J. Neurobiol. 48, 163–180. 18. Zaretsky, M.D. and Konishi, M. (1976). Tonotopic organization in the avian telencephalon. Brain Res. 111, 167–171. 19. Treves, A. and Panzeri, S. (1995). The upward bias in measures of information derived from limited data samples. Neural Comput., 7, 399–407. 20. Brenner, N., Strong, S., Koberle, R. and Bialek, W. (2000). Synergy in a neural code, Neural Comput. 12, 1531–1552.
|
2001
|
33
|
2,029
|
A Variational Approach to Learning Curves D¨orthe Malzahn Manfred Opper Neural Computing Research Group School of Engineering and Applied Science Aston University, Birmingham B4 7ET, United Kingdom. [malzahnd,opperm]@aston.ac.uk Abstract We combine the replica approach from statistical physics with a variational approach to analyze learning curves analytically. We apply the method to Gaussian process regression. As a main result we derive approximative relations between empirical error measures, the generalization error and the posterior variance. 1 Introduction Approximate expressions for generalization errors for finite dimensional statistical data models can be often obtained in the large data limit using asymptotic expansions. Such methods can yield approximate relations for empirical and true errors which can be used to assess the quality of the trained model see e.g. [1]. Unfortunately, such an approximation scheme does not seem to be easily applicable to popular non-parametric models like Gaussian process (GP) models and Support Vector Machines (SVMs). We apply the replica approach of statistical physics to asses the average case learning performance of these kernel machines. So far, the tools of statistical physics have been successfully applied to a variety of learning problems [2]. However, this elegant method suffers from the drawback that data averages can be performed exactly only under very idealistic assumptions on the data distribution in the ”thermodynamic” limit of infinite data space dimension. We try to overcome these limitations by combining the replica method with a variational approximation. For Bayesian models, our method allows us to express useful data averaged a-posteriori expectations by means of an approximate measure. The derivation of this measure requires no assumptions about the data density and no assumptions about the input dimension. The main focus of this article is Gaussian process regression where we demonstrate the various strengths of the presented method. It solves some of the problems stated at the end of our previous NIPS paper [3] which was based on a simpler somewhat unmotivated truncation of a cumulant expansion. For Gaussian process models we show that our method does not only give explicit approximations for generalization errors but also of their sample fluctuations. Furthermore, we show how to compute corrections to our theory and demonstrate the possibility of deriving approximate universal relations between average empirical and true errors which might be of practical interest. An earlier version of our approach, which was still restricted to the assumption of idealized data distributions appeared in [4]. 2 Setup and Notation We assume that a class of elementary predictors (neural networks, regressors etc.) is given by functions . In a Bayesian formulation, we have a prior distribution over this class of functions . Assuming that a set of observations
is conditionally independent given inputs
, we assign a likelihood term of the form
to each observation. Posterior expectations (denoted by angular brackets) of any functional !#" $ are expressed in the form % !#" $&(' ) * +-,.!#" $
, / 10 23 5464 (1) where the partition function * normalizes the posterior and + denotes the expectation with respect to the prior. We are interested in computing averages 7 % !#" $&589 of posterior expectations over different drawings of training data sets : ' " ;
# <$ were all data examples are independently generated from the same distribution. In the next section we will show how to derive a measure which enables us to compute analytically approximate combined data and posterior averages. 3 A Grand-Canonical Approach We utilize the statistical mechanics approach to the analysis of learning. Our aim is to compute the so-called averaged ”free energy” 7 >=1? * 8@9 which serves as a generating function for suitable data averages of posterior expectations. The partition function * is * ' + , / 10 . 3 4 (2) To perform the average 7 =A? * 8B9 we use the replica trick 7 =1? * 8B9C'C=AD1EGFHJILK.M NPO Q.R SUTWV K F , where 7 * F 8B9 is computed for integer X and the continuation is performed at the end [5]. We obtain * FYZ[ ' 7 * F 8B9\' + F^]_G` , F / a 0 a . 54Jbc1de fghi (3) where + F denotes the expectation over the replicated prior measure. Eq.(3) can be transformed into a simpler form by introducing the ”grand canonical” partition function j F kl j F kl 'nm / 0YIpoq Zsr * F Z>t' + F vu F (4) with the Hamiltonian uwFx'y o q ` , F / a 0 l a z4Jbc1de fg (5) The density o|{} R evaluates all X replicas
~F of the predictor at the same data point and the expectation 7W
8 c1de fg is taken with respect to the true data density . The ”grand canonical” partition function j Fkl represents a ”poissonized” version of the original model with fluctuating number of examples. The ”chemical potential” k determines the expected value of Z' K.M N R c q g K q which yields simply k'
=A?Z for X . For sufficiently large Z , we can replace the sum in Eq. (4) by its dominating term =A? * FYZ>t=A? j FYkllZs=1?Z ) UZk (6) thereby neglecting relative fluctuations. We recover the original (canonical) free energy as K M NQ R c g K F K6M N R c M N g K F . 4 Variational Approximation For most interesting cases, the partition function j FYkl can not be computed in closed form for given X . Hence, we use a variational approach to approximate u F by a different tractable Hamiltonian u I F . It is easy to write down the first terms in an expansion of the ”grand canonical” free energy with respect to the difference u F u I F >=A? j FYkl ' >=1? + F o {} R % uwFYu I F &5I ) % u#Fu I F &5I; % uwFYu I F & I
(7) The brackets %
&zI denote averages with respect to the effective measure which is induced by the prior and o;{.} R and acts in the space of replicated variables. As is well known, the first two leading terms in Eq.(7) present an upper bound [6] to >=A? j FYkl . Although differentiating the bound with respect to X will usually not preserve the inequality, we still expect 1 that an optimization with respect to u I F is a sensible thing to do [7]. 4.1 Variational Equations The grand-canonical ensemble was chosen such that Eq.(5) can be rewritten as an integral over a local quantity in the input variable , i.e. in the form u F ' Z
| " a $ with s " a .<$ '
, F / a 0 l a z4 (8) We will now specialize to Gaussian priors over , for which a local quadratic expression u I F '
| / a ) a . a .l / a a a . (9) is a suitable trial Hamiltonian, leading to Gaussian averages %
& I . The functions a . and a . are variational parameters to be optimized. It is important to have an explicit dependence on the input variable in order to take a non uniform input density into account. To perform the variation of the first two terms in Eq.(7) we note that the locality of Eq.(8) makes the ”variational free energy” >=A? + F zvu I F % uwF u I F &zI an explicit function of the first two local moments a . ' % a . .&5I a . ' % a &5I (10) Hence, a straightforward variation yields Z
% s &zI
a . ' a Z
% . &zI
a . ' a (11) To extend the variational solutions to non-integer values of X , we assume that for all the optimal parameters are replica symmetric, ie. a ' . as well as a .' . for ' and aa ' I| . We also use a corresponding notation for a . and a . 1Guided by the success of the method in physical applications, for instance in polymer physics. 4.2 Interpretation of u I F Note, that our approach is not equivalent to a variational approximation of the original posterior. In contrast, uwI contains the full information of the statistics of the training data. We can use the distribution induced by the prior and o {.} R in order to compute approximate combined data and posterior averages. As an example, we first consider the expected local posterior variance . ' 7 % .& % .& 8 9 . Following the algebra of the replica method (see [5]) this is approximated within the variational replica approach as ' =AD1E F|H I % a & I % a . & I ' I .t . (12) Second, we consider the noisy local mean square prediction error of the posterior mean predictor ' % & which is given by ' 7 .t 8@9 . In this case ' =AD1E FH I % a . &5I % a .&5I ' U
( (13) We can also calculate fluctuations with respect to the data average, for example 7 .t | B @8 9 ' =AD1E FH I a 0 e a ( 0 e B @
I (14) 5 Regression with Gaussian Processes This statistical model assumes that data are generated as ' 23
, where is Gaussian white noise with variance { . The prior over functions has zero mean and covariance t' + 7 . 58 . Hence, we have l t' . . Using the definitions Eqs.(12,13), we get % " a .<$&5I X 'y . =1? ) |.
. { (15) which yields the set of variational equations (11). They become particularly easy when the regression model uses a translationally invariant kernel and the input distribution is homogeneous in a finite interval. The variational equations (11) can then be solved in terms of the eigenvalues of the Gaussian process kernel. [8, 9] studied learning curves for Gaussian process regression which are not only averaged over the data but also over the data generating process using a Gaussian process prior on . Applying these averages to our theory and adapting the notation of [9] simply replaces in Eq.(15) the term
. by . while . . . 5.1 Learning Curves and Fluctuations Practical situations differ from this ”typical case” analysis. The data generating process is unknown but assumed to be fixed. The resulting learning curve is then conditioned on this particular ”teacher” . The left panel of Fig.1 shows an example. Displayed are the mean square prediction error (circle and solid line) and its sample fluctuations (error bars) with respect to the data average (cross and broken line). The target was a random but fixed realization from a Gaussian process prior with a periodic Radial Basis Function kernel p' "! z $# &% ' , ' ' ) . We keep the example simple, e.g the Gaussian process regression model used the same kernel and noise { ' ' ) . The inputs are one dimensional, independent and uniformly distributed )( 7 ) 8 . Symbols represent simulation data. A typical property of our theory (lines) is that it becomes very accurate for sufficiently large number of example data. 0 50 100 150 200 Number m of Example Data 10 −4 10 −3 10 −2 10 −1 10 0 Generalization Error ε, Fluctuation ∆ε ε ∆ε Theory: Lines Simulation: Symbols 0 20 40 60 80 100 Number m of Example Data −4 −3 −2 −1 0 Correction of Free Energy β −1=0.25 β −1=0.01 β −1=0.0001 Figure 1: Gaussian process regression using a periodic Radial Basis Function kernel, input dimension d=1, ( 7 ) 8 , and homogeneous input density. Left: Generalization error and fluctuations for data noise ' { ' ) . Right: Correction of the free energy. Symbols: We subtracted the first two contributions to Eq.(7) from the true value of the free energy. The latter was obtained by simulations. Lines show the third contribution of Eq.(7). The value of the noise variance { decreases from top to bottom. All y-data was set equal to zero. 5.2 Corrections to the Variational Approximation It is a strength of our method that the quality of the variational approximation Eq.(7) can be characterized and systematically improved. In this paper, we restrict ourself to a characterization and consider the case where all -data is set equal to zero. Since the posterior variance is independent of the data this is still an interesting model from which the posterior variance can be estimated. We consider the third term in the expansion to the free energy Eq.(7). It is a correction to the variational free energy and evaluates to =ADAE FH I X ) % u F u I F & I % u F u I F & I 'y[) 7 I . I B ; B58 d e d Z `=A? , ) l^ { 4Jb d e d Z ` I { b d e d (16) with ' =1DAE FH I % a . a p a . & I . Eq.(16) is shown by lines in the right panel of Fig.1 for different values of the model noise { . We considered a homogeneous input density, the input dimension is one and the regression model uses a periodic RBF kernel. The symbols in Fig.1 show the difference between the true value of the free energy which is obtained by simulations and the first two terms of Eq.(7). The correction term is found to be qualitatively accurate and emphasizes a discrepancy between free energy and the first two terms of the expansion Eq.(7) for a medium amount of example data. The calculated learning curves inherit this behaviour. 5.3 Universal Relations We can relate the training error and the empirical posterior variance ' ) Z ` / A0 b 9 ' ) Z ` / 10 ; b 9 (17) 0 0.2 0.4 0.6 0.8 1 βσT 2 0 0.2 0.4 0.6 0.8 1 [βσ 2(x)/(βσ 2(x)+1)]x Theory d=1, periodic d=2, periodic d=3, periodic d=2+2, non-periodic 0 0.2 0.4 0.6 0.8 1 βεT 0 0.2 0.4 0.6 0.8 1 [βε(x,y)/(βσ 2(x)+1) 2](x,y) Theory 1d, periodic 2d, periodic 3d, periodic Figure 2: Illustration of relation Eq.(19) (left) and Eq.(20) (right). All error measures are scaled with . Symbols show simulation results for Radial Basis Function (RBF) regression and a homogeneous input distribution in
x' ) dimensions (square, circle, diamond). The RBF kernel was periodic. Additionally, the left figure shows an example were the inputs lie on a quasi two-dimensional manifold which is embedded in
' dimensions (cross). In this case the RBF kernel was non-periodic. to the free energy 7 =A? * 8 9 ' . Using Eqs.(6,7) and the stationarity of the grand-canonical free energy with respect to the variational parameters we obtain the following relation
| 7 >=1? * 8B9 yZ
| % " a $&5I X (18) We use the fact that the posterior variance is independent of the -data and simply estimate it from the model where all -data is set equal to zero. In this case, Eq.(18) yields '
) ^ . (19) which relates the empirical posterior variance to the local posterior variance . at test inputs . Similarly, we can derive an expression for the training error by using Eqs.(15,18) in combination with Eq.(19) '
|
) . (20) It is interesting to note, that the relations (19,20) contain no assumptions about the data generating process. They hold in general for Gaussian process models with a Gaussian likelihood. An illustration of Eqs.(19,20) is given by Fig.2 for the example of Gaussian process regression with a Radial Basis Function kernel. In the left panel of Fig.2, learning starts in the upper right corner as the rescaled empirical posterior variance is initially one and decreases with increasing number of example data. For the right panel of Fig.2, learning starts in the lower left corner. The rescaled training error on the noisy data set is initially zero and increases to one with increasing number of example data. The theory (line) holds for a sufficiently large number of example data and its accuracy increases with the input dimension. Eqs.(19,20) can also be tested on real data. For common benchmark sets such as Abalone and Boston Housing data we find that Eqs.(19,20) hold well even for small and medium sizes of the training data set. 6 Outlook One may question if our approximate universal relations are of any practical use as, for example, the relation between training error and generalization error involves also the unknown posterior variance . . Nevertheless, this relation could be useful for cases, where a large number of data inputs without output labels are available. Since for regression, the posterior variance is independent of the output labels, we could use these extra input points to estimate . The application of our technique to more complicated models is possible and technically more involved. For example, replacing o { by # .. ) in Eq.(1) and further rescaling the kernel t' % of the Gaussian process prior gives a model for hard margin Support Vector Machine Classification with SVM kernel . The condition of maximum margin classification will be ensured by the limes . Of particular interest is the computation of empirical estimators that can be used in practice for model selection as well as the calculation of fluctuations (error bars) for such estimators. A prominent example is an efficient approximate leave-one-out estimator for SVMs. Work on these issues is in progress. Acknowledgement We would like to thank Peter Sollich for may inspiring discussions. The work was supported by EPSRC grant GR/M81601. References [1] N. Murata, S. Yoshizawa, S. Amari, IEEE Transactions on Neural Networks 5, p. 865-872, (1994). [2] A. Engel, C. Van den Broeck, Statistical Mechanics of Learning, Cambridge University Press (2001). [3] D. Malzahn, M. Opper, Neural Information Processing Systems 13, p. 273, T. K. Leen, T. G. Dietterich and V. Tresp, eds., MIT Press, Cambridge MA (2001). [4] D. Malzahn, M. Opper, Lecture Notes in Computer Science 2130, p. 271, G. Dorffner, H. Bischof and K. Hornik, eds., Springer, Berlin (2001). [5] M. M´ezard, G. Parisi, M. Virasoro, Spin Glass Theory and Beyond, World Scientific, Singapore, (1987). [6] R. P. Feynman and A. R. Hibbs, Quantum mechanics and path integrals, Mc GrawHill Inc., (1965). [7] T. Garel, H. Orland, Europhys. Lett. 6, p. 307 (1988). [8] P. Sollich, Neural Information Processing Systems 11, p. 344, M. S. Kearns, S. A. Solla and D. A. Cohn, eds., MIT Press, Cambridge MA (1999). [9] P. Sollich, Neural Information Processing Systems 14, T. G. Dietterich, S. Becker, Z. Ghahramani, eds., MIT Press (2002).
|
2001
|
34
|
2,030
|
Why neuronal dynamics should control synaptic learning rules Jesper Tegner Stockholm Bioinformatics Center Dept. of Numerical Analysis & Computing Science Royal Institute for Technology S-10044 Stockholm, Sweden jespert@nada.kth.se Adam Kepecs Volen Center for Complex Systems Brandeis University Waltham, MA 02454 kepecs@brandeis.edu Abstract Hebbian learning rules are generally formulated as static rules. Under changing condition (e.g. neuromodulation, input statistics) most rules are sensitive to parameters. In particular, recent work has focused on two different formulations of spike-timing-dependent plasticity rules. Additive STDP [1] is remarkably versatile but also very fragile, whereas multiplicative STDP [2, 3] is more robust but lacks attractive features such as synaptic competition and rate stabilization. Here we address the problem of robustness in the additive STDP rule. We derive an adaptive control scheme, where the learning function is under fast dynamic control by postsynaptic activity to stabilize learning under a variety of conditions. Such a control scheme can be implemented using known biophysical mechanisms of synapses. We show that this adaptive rule makes the additive STDP more robust. Finally, we give an example how meta plasticity of the adaptive rule can be used to guide STDP into different type of learning regimes. 1 Introduction Hebbian learning rules are widely used to model synaptic modification shaping the functional connectivity of neural networks [4, 5]. To ensure competition between synapses and stability of learning, constraints have to be added to correlational Hebbian learning rules [6]. Recent experiments revealed a mode of synaptic plasticity that provides new possibilities and constraints for synaptic learning rules [7, 8, 9]. It has been found that synapses are strengthened if a presynaptic spike precedes a postsynaptic spike within a short (::::: 20 ms) time window, while the reverse spike order leads to synaptic weakening. This rule has been termed spike-timing dependent plasticity (STDP) [1]. Computational models highlighted how STDP combines synaptic strengthening and weakening so that learning gives rise to synaptic competition in a way that neuronal firing rates are stabilized. Recent modeling studies have, however, demonstrated that whether an STDP type rule results in competition or rate stabilization depends on exact formulation of the weight update scheme [3, 2]. Sompolinsky and colleagues [2] introduced a distinction between additive and multiplicative weight updating in STDP. In the additive version of an STDP update rule studied by Abbott and coworkers [1, 10], the magnitude of synaptic change is independent on synaptic strength. Here, it is necessary to add hard weight bounds to stabilize learning. For this version of the rule (aSTDP), the steady-state synaptic weight distribution is bimodal. In sharp contrast to this, using a multiplicative STDP rule where the amount of weight increase scales inversely with present weight size produces neither synaptic competition nor rate normalization [3, 2]. In this multiplicative scenario the synaptic weight distribution is unimodal. Activity-dependent synaptic scaling has recently been proposed as a separate mechanism to ensure synaptic competition operating on a slow (days) time scale [3]. Experimental data as of today is not yet sufficient to determine the circumstances under which the STDP rule is additive or multiplicative. In this study we examine the stabilization properties of the additive STDP rule. In the first section we show that the aSTDP rule normalizes postsynaptic firing rates only in a limited parameter range. The critical parameter of aSTDP becomes the ratio (0;) between the amount of synaptic depression and potentiation. We show that different input statistics necessitate different 0; ratios for aSTDP to remain stable. This lead us to consider an adaptive version of aSTDP in order to create a rule that is both competitive as well as rate stabilizing under different circumstances. Next, we use a Fokker-Planck formalism to clarify what determines when an additive STDP rule fails to stabilize the postsynaptic firing rate. Here we derive the requirement for how the potentiation to depression ratio should change with neuronal activity. In the last section we provide a biologically realistic implementation of the adaptive rule and perform numerical simulations to show the how different parameterizations of the adaptive rule can guide STDP into differentially rate-sensitive regimes. 2 Additive STDP does not always stabilize learning First, we numerically simulated an integrate-and-fire model receiving 1000 excitatory and 250 inhibitory afferents. The weights of the excitatory synapses were updated according to the additive STDP rule. We used the model developed by Song et al, 2000 [1]. The learning kernel L(T) is A+exp(T/T+) if T < 0 or -A_ exp( -T/L) if T > 0 where A_ / A+ denotes the amplitude of depression/potentiation respectively. Following [1] we use T + = T _ = 20 ms for the time window of learning. The integral over the temporal window of the synaptic learning function (L) is always negative. Synaptic weights change according to dWi J ill = L(T)Spre(t + T)Spost(T)dT , Wi E[O,Wmax ] (1) where s(t) denotes a delta function representing a spike at time t. Correlations between input rates were generated by adding a common bias rate in a graded manner across synapses so that the first afferent is has zero while the last afferent has the maximal correlation, Cmax . We first examine how the depression/potentiation ratio (0; = LTD / LT P) [2] controls the dependence of the output firing rate on the synaptic input rate, here referred to as the effective neuronal gain. Provided that 0; is sufficiently large, the STDP rule controls the postsynaptic firing rate (Fig. 1A). The stabilizing effect of the STDP rule is therefore equivalent to having weak a neuronal gain. 600 500 100 10 ~ ~ ~ W W M 00 00 Inpul Rate (liz) B 250 ;; mD , -; Increasing j:: I'lP"'C'~~ 50~ % m w w w ~ ~ Input Rale(Hz) c ,---~--------~ 250 200 150 Increasing t.05 Reference Ratio LTDlt.TPratios I ': I~ 0 -o ~ W 00 00 100 Input Rate (Hz) Figure I: A STDP controls neuronal gain. The slope of the dependence of the postsynaptic output rate on the presynaptic input rate is referred to as the effective neuronal gain. The initial firing rate is shown by the upper curve while the lower line displays the final postsynaptic firing rate. The gain is reduced provided that the depression/potentiation ratio (0: = 1.05 here) is large enough. The input is uncorrelated. B Increasing input correlations increases neuronal gain. When the synaptic input is strongly correlated the postsynaptic neuron operates in a high gain mode characterized by a larger slope and larger baseline rate. Input correlations were uniformly distributed between 0 and a maximal value, Cm a x . The maximal correlation increases in the direction of the arrow: 0.0; 0.2; 0.3; 0.4; 0.5; 0.6; 0.7. The 0: ratio is 1.05. Note that for further increases in the presynaptic rates, postsynaptic firing can increase to over 1000 Hz. C The depression/potentiation ratio sets the neuronal gain. The 0: ratios increase in the direction of arrow:1.025;1.05;1.075;1.1025;1.155;1.2075. Cm a x is 0.5. We find that the neuronal gain is extremely sensitive to the value of 0: as well as to the amount of afferent input correlations. Figure IB shows that increasing the amount of input correlations for a given 0: value increases the overall firing rate and the slope of the input-output curve, thus leading to larger effective gain. Increasing the amount of correlations between the synaptic afferents could therefore be interpreted as increasing the effective neuronal gain. Note that the baseline firing at a presynaptic drive of 20Hz is also increased. Next, we examined how neuronal gain depends on the value of 0: in the STDP rule (Figure IC). The high gain and high rate mode induced by strong input correlations was reduced to a lower gain and lower rate mode by increasing 0: (see arrow in Figure IC). Note, however, that there is no correct 0: value as it depends on both the input statistics as well as the desired input/output relationship. 3 Conditions for an adaptive additive STDP rule Here we address how the learning ratio, 0:, should depend on the input rate in order to produce a given neuronal input-output relationship. Using this functional form we will be able to formulate constraints for an adaptive additive STDP rule. This will guide us in the derivation of a biophysical implementation of the adaptive control scheme. The problem in its generality is to find (i) how the learning ratio should depend on the postsynaptic rate and (ii) how the postsynaptic rate depends on the input rate and the synaptic weights. By performing self-consistent calculations using a Fokker-Planck formulation, the problem is reduced to finding conditions for how the learning ratio should depend on the input rates only. Let 0: denote depression/potentiation ratio 0: = LTD/LTP as before. Now we meanw A 30 ouput rate B ,-------------~ 0.6,-------------~ 25 20 15 D o. 1 ,-----,----==:=::=--,---~ 0.05 T···· • • ••• • •••• •••••••••• • • •••• • ••• . . . . °0L--2~0---4~0--6~0---8~0 input rate 0.5 0.4 0.3 0.2 0.1 . ~ . .•.... . . . . . . •.... . . . . . .. . ... . . ... •....... . . .. . . -- ... - . . . . : .. . . . . 0.5 w C WTOT 40,-------------~ 35 30 25 0.5 w Figure 2: Self consistent Fokker-Planck calculations. Conditions for zero neuronal gain. U A The output rate does not depend on the input rate. Zero neuronal gain. B Dependence of the mean synaptic weight on input rates. C W tot ex: Tpre < W >, see text. D The dependence of j3 = a - 1 on input rate. E,F A( w) and P( w) are functions of the synaptic strength and depend on the input rate .. Note that eight different input rates are used but only traces 1, 3, 5, 7 are shown for A(w) and pew) in which the dashed line correspond to the case with the lowest presynaptic rate. determine how the parameter fJ = 0: - 1 should scale with presynaptic rates in order to control the neuronal gain. The Fokker-Planck formulation permits an analytic calculation of the steady state distribution of synaptic weights [3]. The competition parameter for N excitatory afferents is given by Wtot = twrpreN < w > where the time window tw is defined as the probability for depression (Pd = tw/tisi) that a synaptic event occurs within the time window (tw < tisi ). The amount of potentiation and depression for the additive STDP yields in the steady-state, neglecting the exponential timing dependence, the following expression for the drift term A(w) A(w) = PdA-[W/Wtot - (1 - 1/0:)] (2) A( w) represents the net weight "force field" experienced by an individual synapse. Thus, A( w) determines whether a given synapse (w) will increase or decrease as a function of its synaptic weight. The steepness of the A( w) function determines the degree of synaptic competition. The w /Wtot is a competition term whereas the (1 - 1/0:) provides a destabilizing force. When Wmax > (1 - l/o:)Wtot the synaptic weight distribution is bimodal. The steady state distribution reads P(w) = Ke[(-w(1-1 /a) +w 2 /(2 Wt ot ))/(A _ )] (3) where K normalizes the P(w) distribution [3]. Now, equations (2-3), with appropriate definitions of the terms, constitute a selfconsistent system. Using these equations one can calculate how the parameter fJ should scale with the presynaptic input rate in order to produce a given postsynaptic firing rate. For a given presynaptic rate, equations (2-3) can be iterated in until a self-consistent solution is found. At that point, the postsynaptic firing rate can be calculated. Here, instead we impose a fixed postsynaptic output rate for a given input rate and search for a self-consistent solution using (3 as a free parameter. Performing this calculation for a range of input rates provides us with the desired dependency of (3 on the presynaptic firing rate. Once a solution is reached we also examine the resulting steady state synaptic weight distribution (P(w)) and the corresponding drift term A( w) as a function of the presynaptic input rate. The results of such a calculation are illustrated in Figure 2. The neuronal gain, the ratio between the postsynaptic firing rate and the input rate is set to be zero (Fig 2A). To normalize postsynaptic firing rates the average synaptic weight has to decrease in order to compensate for the increasing presynaptic firing rate. This can be seen in (Fig 2B). The condition for a zero neuronal gain is that the average synaptic weight should decrease as 1 j r pre. This makes Wtot constant as shown in Fig 2C. For these values, (3 has to increase with input rate as shown in Fig 2D. Note that this curve is approximately linear. The dependence of A( w) and the synaptic weight distribution P( w) on different presynaptic rates is illustrated in Fig 2E and F. As the presynaptic rates increase, the A(w) function is lowered (dashed line indicates the smallest presynaptic rate), thus pushing more synapses to smaller values since they experience a net negative "force field". This is also reflected in the synaptic weight distribution which is pushed to the lower boundary as the input rates increase. When enforcing a different neuronal gain, the dependence of the (3 term on the presynaptic rates remains approximately linear but with a different slope (not shown). 4 Derivation of an adaptive learning rule with biophysical components The key insight from the above calculations is the observed linear dependence of (3 on presynaptic rates. However, when implementing an adaptive rule with biophysical elements it is very likely that individual components will have a non-linear dependence on each other. The Fokker-Planck analysis suggests that the non-linearities should effectively cancel. Why should the system be linear? Another way to see from where the linearity requirement comes is that the (w jWtot - (3) term in expression for A(w) (valid for small (3) has to be appropriately balanced when the input rates increases. The linearity of (3(rpre ) follows from Wtot being linear in r pre . Now, how could (3 depend on presynaptic rates? A natural solution would be to use postsynaptic calcium to measure the postsynaptic firing and therefore indirectly the presynaptic firing rate. Moreover, the asymmetry ((3) of the learning ratio could depend on the level of postsynaptic calcium. It is known that increased resting calcium levels inhibit NMDA channels and thus calcium influx due to synaptic input. Additionally, the calcium levels required for depression are easier to reach. Both of these effects in turn increase the probability of LTD induction. Incorporating these intermediate steps gives the following scheme: (3 q c p h +-'-+ a t-=--+ r po st +-'---+ r pr e This scheme introduces parameters (p and q) and a function Ut} to control for the linearity jnon-linearity between the variables. The global constraint from the Fokker-Planck is that the effective relation between (3 and r pre should be linear. A biophysical formulation of the above scheme is the following 200 150 ~l oo 5 o 50 No Adaptive Tracking Adaptive Tracking 20 40 60 80 input rat. i-2:WlUlliWWU] ~-40 > -60 0~----------~5~00~--------~10~0~0----------~1~500 '·'r ~A_ .~ 1 ~'ll V'~ "1 100 '0 500 1000 1500 Time (ms) Figure 3: Left Steady-state response with (squares) or without (circles) the adaptive tracking scheme. When the STDP rule is extended with an adaptive control loop, the output rates are normalized in the presence of correlated input. Right Fast adaptive tracking. Since (3 tracks changes in intracellular calcium on a rapid time-scale, every spike experiences a different learning ratio, 0:. Note that the adaptive scheme approximates the learning ratio (0: = 1.05) used in [1]. d(3 T(3 = - (3 + [Ca]q dt (4) (5) The parameter p determines how the calcium concentration scales with the postsynaptic firing rate (delta spikes r5 above) and q controls the learning sensitivity. "( controls the rise of steady-state calcium with increasing postsynaptic rates (rpost). The time constants TCa and T(3 determine the calcium dynamics and the time course of the adaptive rule respectively. Note that we have not specified the neuronal transfer function, it. To ensure a linear relation between (3 and r pre it follows from the Fokker-Planck analysis that [it (rpre)]pq is approximately linear in r pre . The neuronal gain can now be independently be controlled by the parameter T Moreover, the drift term A( w) becomes (6) for (3 < < 1. A( w) can be written in this form since we use that Wd - A_ = -A+CI: = -A+(l + [TCa"(r~ost]q). The w/Wtot is a competition term whereas the [TCa"(r~ost ]q provides a destabilizing force. Note also, that when W max > [TCa"(r~ost ]qWtot there is a bimodal synaptic weight distribution and synaptic competition is preserved. A complete stability analysis is beyond the scope of the present study. A B C .-.. 75 75 75 ...• ~. N :EO) 50 50 50 1\1 ... - 25 25 .... ~ . ; 25 :::J Co :::J 0 0 0 0 0 25 50 75 100 0 25 50 75 100 0 25 50 75 100 Input rate (Hz) Input rate (Hz) Input rate (Hz) Figure 4: Full numerical simulation of the adaptive additive STDP rule. Parameters: p = q = 1. Tea = 10ms, T f! = lOOms. A I = 1.25. B I = 0.25. C Input correlations are Cmax = 0, 0.3, 0.6 5 Numerical simulations Next, we examine whether the theory of adaptive normalization carryover to a full scale simulation of the integrate-and-fire model with the STDP rule and the biophysical adaptive scheme as described above. First, we studied the neuronal gain (cf. Figure 1) when the inputs were strongly correlated. Driving a neuron with increasing input rates increases the output rate significantly when there is no adaptive scheme (squares, Figure 3 Left) as observed previously (cf. Figure IB). Adding the adaptive loop normalizes the output rates (circles, Figure 3 Left). This simulation shows that the average postsynaptic firing rate is regulated by the adaptive tracking scheme. This is expected since the Fokker-Planck analysis is based on the steady-state synaptic weight distribution. To further gain insight into the operation of the adaptive loop we examined the spike-to-spike dependence of the tracking scheme. Figure 3 (Right) displays the evolution of the membrane potential (top) and the learning ratio 0: = 1 + (3 (bottom) . The adaptive rule tracks fast changes in firing by adjusting the learning ratio for each spike. Thus, the strength plasticity is different for every spike. Interestingly, the learning ratio (0:) fluctuates around the value 1.05 which was used in previous studies [1] . Our fast, spike-to-spike tracking scheme is in contrast to other homeostatic mechanisms operating on the time-scale of hours to days [11, 12, 13, 14]. In our formulation, the learning ratio, via (3, tracks changes in intra-cellular calcium, which in turn reflects the instantaneous firing rate. Slower homeostatic mechanisms are unable to detect these rapid changes in firing statistics. Because this fast adaptive scheme depends on recent neuronal firing, pairing several spikes on the time-scale comparable to the calcium dynamics introduces non-linear summation effects. Neurons with this adaptive STDP control loop can detect changes in the input correlation while being only weakly dependent on the presynaptic firing rate. Figure 4a and 4b show two different regimes corresponding to two different values of the parameter , . In the high , regime (Fig. 4a) the neuronal gain is zero. The neuronal gain increased when , decreased (Fig. 4b) as expected from the theory. In a different regime where we introduce increasing correlations between the synaptic inputs [1] we find that the neuronal gain is changed little with increasing input rates but increases substantially with increasing input correlations (Fig 4c) . Thus, the adaptive aSTDP rule can normalize the mean postsynaptic rate even when the input statistics change. With other adaptive parameters we also found learning regimes where the responses to input correlations were affected differentially (not shown). 6 Discussion Synaptic learning rules have to operate under widely changes conditions such as different input statistics or neuromodulation. How can a learning rule dynamically guide a network into functionally similar operating regime under different conditions? We have addressed this issue in the context of spike-timing-dependent plasticity (STDP) [1, 10J. We found that STDP is very sensitive to the ratio of synaptic strengthening to weakening, (t, and requires different values for different input statistics. To correct for this, we proposed an adaptive control scheme to adjust the plasticity rule. This adaptive mechanisms makes the learning rule more robust to changing input conditions while preserving its interesting properties, such as synaptic competition. We suggested a biophysically plausible mechanism that can implement the adaptive changes consistent with the requirements derived using the Fokker-Planck analysis. Our adaptive STDP rule adjusts the learning ratio on a millisecond time-scale. This in contrast to other, slow homeostatic controllers considered previously [11, 12, 13, 14, 3J. Because the learning rule changes rapidly, it is very sensitive the input statistics. Furthermore, the synaptic weight changes add non-linearly due to the rapid self-regulation. In recent experiments similar non-linearities have been detected (Y. Dan, personal communication) which might have roles in making synaptic plasticity adaptive. Finally, the new set of adaptive parameters could be independently controlled by meta-plasticity to bring the neuron into different operating regimes. Acknowledgments We thank Larry Abbott, Mark van Rossum, and Sen Song for helpful discussions. J.T. was supported by the Wennergren Foundation, and grants from Swedish Medical Research Foundation, and The Royal Academy for Science. A.K. was supported by the NIH Grant 2 ROI NS27337-12 and 5 ROI NS27337-13. Both A.K. and J.T. thank the Sloan Foundation for support. References [1] Song, S., Miller, K , & Abbott, L. Nature Neuroscience, 3:919-926, 2000. [2] Rubin, J., Lee, D., & Sompolinsky, H. Physical Review Letter, 86:364-367, 200l. [3] van Rossum, M., G-Q, B. , & Thrrigiano, G. J Neurosci, 20:8812- 8821, 2000. [4] Sejnowski, T. J Theoretical Biology, 69:385- 389, 1997. [5] Abbott, L. & Nelson, S. Nature Neuroscience, 3:1178- 1183, 2000. [6] Miller, K & MacKay, D. Neural Computation, 6:100- 126, 1994. [7] Markram, H., Lubke, J., Frotscher, M., & Sakmann, B. Science, 275:213- 215, 1997. [8] Bell, C., Han, V., Sugawara, Y., & Grant, K Nature, 387:278- 81, 1997. [9] Bi, G.-Q. & Poo, M. J Neuroscience, 18:10464- 10472, 1998. [10] Kempter, R., Gerstner, W., & van Hemmen, J. Neural Computation, 13:2709- 2742, 200l. [11] Bell, A. In Moody, J., Hanson, S., & Lippmann, R., editors, Advances in Neural Information Processing Systems, volume 4. Morgan-Kaufmann, 1992. [12] LeMasson, G., Marder, E., & Abbott, L. Science, 259:1915- 7, 1993. [13] Thrrigiano, G. , Leslie, K , Desai, N., Rutherford, L., & Nelson, S. Nature, 391:892- 6, 1998. [14] Thrrigiano, G. & Nelson, S. Curr Opin Neurobiol, 10:358- 64, 2000.
|
2001
|
35
|
2,031
|
Latent Dirichlet Allocation David M. Blei, Andrew Y. Ng and Michael I. Jordan University of California, Berkeley Berkeley, CA 94720 Abstract We propose a generative model for text and other collections of discrete data that generalizes or improves on several previous models including naive Bayes/unigram, mixture of unigrams [6], and Hofmann's aspect model, also known as probabilistic latent semantic indexing (pLSI) [3]. In the context of text modeling, our model posits that each document is generated as a mixture of topics, where the continuous-valued mixture proportions are distributed as a latent Dirichlet random variable. Inference and learning are carried out efficiently via variational algorithms. We present empirical results on applications of this model to problems in text modeling, collaborative filtering, and text classification. 1 Introduction Recent years have seen the development and successful application of several latent factor models for discrete data. One notable example, Hofmann's pLSI/aspect model [3], has received the attention of many researchers, and applications have emerged in text modeling [3], collaborative filtering [7], and link analysis [1]. In the context of text modeling, pLSI is a "bag-of-words" model in that it ignores the ordering of the words in a document. It performs dimensionality reduction, relating each document to a position in low-dimensional "topic" space. In this sense, it is analogous to PCA, except that it is explicitly designed for and works on discrete data. A sometimes poorly-understood subtlety of pLSI is that, even though it is typically described as a generative model, its documents have no generative probabilistic semantics and are treated simply as a set of labels for the specific documents seen in the training set. Thus there is no natural way to pose questions such as "what is the probability of this previously unseen document?". Moreover, since each training document is treated as a separate entity, the pLSI model has a large number of parameters and heuristic "tempering" methods are needed to prevent overfitting. In this paper we describe a new model for collections of discrete data that provides full generative probabilistic semantics for documents. Documents are modeled via a hidden Dirichlet random variable that specifies a probability distribution on a latent, low-dimensional topic space. The distribution over words of an unseen document is a continuous mixture over document space and a discrete mixture over all possible topics. 2 Generative models for text 2.1 Latent Dirichlet Allocation (LDA) model To simplify our discussion, we will use text modeling as a running example throughout this section, though it should be clear that the model is broadly applicable to general collections of discrete data. In LDA, we assume that there are k underlying latent topics according to which documents are generated, and that each topic is represented as a multinomial distribution over the IVI words in the vocabulary. A document is generated by sampling a mixture of these topics and then sampling words from that mixture. More precisely, a document of N words w = (W1,'" ,W N) is generated by the following process. First, B is sampled from a Dirichlet(a1,'" ,ak) distribution. This means that B lies in the (k I)-dimensional simplex: Bi 2': 0, 2:i Bi = 1. Then, for each of the N words, a topic Zn E {I, ... ,k} is sampled from a Mult(B) distribution p(zn = ilB) = Bi. Finally, each word Wn is sampled, conditioned on the znth topic, from the multinomial distribution p(wlzn). Intuitively, Bi can be thought of as the degree to which topic i is referred to in the document. Written out in full, the probability of a document is therefore the following mixture: p(w) = Ie (11 z~/(wnl zn; ,8)P(Zn IB») p(B; a)dB, (1) where p(B; a) is Dirichlet, p(znIB) is a multinomial parameterized by B, and p( Wn IZn;,8) is a multinomial over the words. This model is parameterized by the kdimensional Dirichlet parameters a = (a1,' .. ,ak) and a k x IVI matrix,8, which are parameters controlling the k multinomial distributions over words. The graphical model representation of LDA is shown in Figure 1. As Figure 1 makes clear, this model is not a simple Dirichlet-multinomial clustering model. In such a model the innermost plate would contain only W n ; the topic node would be sampled only once for each document; and the Dirichlet would be sampled only once for the whole collection. In LDA, the Dirichlet is sampled for each document, and the multinomial topic node is sampled repeatedly within the document. The Dirichlet is thus a component in the probability model rather than a prior distribution over the model parameters. We see from Eq. (1) that there is a second interpretation of LDA. Having sampled B, words are drawn iid from the multinomial/unigram model given by p(wIB) = 2::=1 p(wlz)p(zIB). Thus, LDA is a mixture model where the unigram models p(wIB) are the mixture components, and p(B; a) gives the mixture weights. Note that unlike a traditional mixture of unigrams model, this distribution has an infinite o 1'0 '. I Zn Wn Nd D Figure 1: Graphical model representation of LDA. The boxes are plates representing replicates. The outer plate represents documents, while the inner plate represents the repeated choice of topics and words within a document. Figure 2: An example distribution on unigram models p(wIB) under LDA for three words and four topics. The triangle embedded in the x-y plane is the 2-D simplex over all possible multinomial distributions over three words. (E.g., each of the vertices of the triangle corresponds to a deterministic distribution that assigns one of the words probability 1; the midpoint of an edge gives two of the words 0.5 probability each; and the centroid of the triangle is the uniform distribution over all 3 words). The four points marked with an x are the locations of the multinomial distributions p(wlz) for each of the four topics, and the surface shown on top of the simplex is an example of a resulting density over multinomial distributions given by LDA. number of continuously-varying mixture components indexed by B. The example in Figure 2 illustrates this interpretation of LDA as defining a random distribution over unigram models p(wIB). 2.2 Related models The mixture of unigrams model [6] posits that every document is generated by a single randomly chosen topic: (2) This model allows for different documents to come from different topics, but fails to capture the possibility that a document may express multiple topics. LDA captures this possibility, and does so with an increase in the parameter count of only one parameter: rather than having k - 1 free parameters for the multinomial p(z) over the k topics, we have k free parameters for the Dirichlet. A second related model is Hofmann's probabilistic latent semantic indexing (pLSI) [3], which posits that a document label d and a word ware conditionally independent given the hidden topic z : p(d, w) = L~= l p(wlz)p(zld)p(d). (3) This model does capture the possibility that a document may contain multiple topics since p(zld) serve as the mixture weights of the topics. However, a subtlety of pLSIand the crucial difference between it and LDA-is that d is a dummy index into the list of documents in the training set. Thus, d is a multinomial random variable with as many possible values as there are training documents, and the model learns the topic mixtures p(zld) only for those documents on which it is trained. For this reason, pLSI is not a fully generative model and there is no clean way to use it to assign probability to a previously unseen document. Furthermore, the number of parameters in pLSI is on the order of klVl + klDI, where IDI is the number of documents in the training set. Linear growth in the number of parameters with the size of the training set suggests that overfitting is likely to be a problem and indeed, in practice, a "tempering" heuristic is used to smooth the parameters of the model. 3 Inference and learning Let us begin our description of inference and learning problems for LDA by examining the contribution to the likelihood made by a single document. To simplify our notation, let w~ = 1 iff Wn is the jth word in the vocabulary and z~ = 1 iff Zn is the ith topic. Let j3ij denote p(wj = Ilzi = 1), and W = (WI, ... ,WN), Z = (ZI, ... ,ZN). Expanding Eq. (1), we have: (4) This is a hypergeometric function that is infeasible to compute exactly [4]. Large text collections require fast inference and learning algorithms and thus we have utilized a variational approach [5] to approximate the likelihood in Eq. (4). We use the following variational approximation to the log likelihood: logp(w; a, 13) log r :Ep(wlz; j3)p(zIB)p(B; a) q~:, z:" ~~ dB le z q ,Z", > Eq[logp(wlz;j3) +logp(zIB) +logp(B;a) -logq(B,z; , ,¢)], where we choose a fully factorized variational distribution q(B, z;" ¢) q(B; ,) fIn q(Zn; ¢n) parameterized by , and ¢n, so that q(B; ,) is Dirichlet({), and q(zn; ¢n) is MUlt(¢n). Under this distribution, the terms in the variational lower bound are computable and differentiable, and we can maximize the bound with respect to, and ¢ to obtain the best approximation to p(w;a,j3). Note that the third and fourth terms in the variational bound are not straightforward to compute since they involve the entropy of a Dirichlet distribution, a (k - I)-dimensional integral over B which is expensive to compute numerically. In the full version of this paper, we present a sequence of reductions on these terms which use the log r function and its derivatives. This allows us to compute the integral using well-known numerical routines. Variational inference is coordinate ascent in the bound on the probability of a single document. In particular, we alternate between the following two equations until the objective converges: (5) ,i ai + 2:~=1 ¢ni (6) where \]i is the first derivative of the log r function. Note that the resulting variational parameters can also be used and interpreted as an approximation of the parameters of the true posterior. In the current paper we focus on maximum likelihood methods for parameter estimation. Given a collection of documents V = {WI' ... ' WM}, we utilize the EM algorithm with a variational E step, maximizing a lower bound on the log likelihood: M logp(V) 2:: l:= Eqm [logp(B, z, w)]- Eqm [logqm(B, z)]. (7) m=l The E step refits qm for each document by running the inference step described above. The M step optimizes Eq. (7) with respect to the model parameters a and (3. For the multinomial parameters (3ij we have the following M step update equation: M Iwml (3ij ex: l:= l:= ¢>mniwtnn· (8) m=l n=l The Dirichlet parameters ai are not independent of each other and we apply N ewton-Raphson to optimize them: The variational EM algorithm alternates between maximizing Eq. (7) with respect to qm and with respect to (a, (3) until convergence. 4 Experiments and Examples We first tested LDA on two text corpora.1 The first was drawn from the TREC AP corpus, and consisted of 2500 news articles, with a vocabulary size of IVI = 37,871 words. The second was the CRAN corpus, consisting of 1400 technical abstracts, with IVI = 7747 words. We begin with an example showing how LDA can capture multiple-topic phenomena in documents. By examining the (variational) posterior distribution on the topic mixture q(B; ')'), we can identify the topics which were most likely to have contributed to many words in a given document; specifically, these are the topics i with the largest ')'i. Examining the most likely words in the corresponding multinomials can then further tell us what these topics might be about. The following is an article from the TREC collection. The William Randolph Hearst Foundation will give $1.25 million to Lincoln Center, Metropolitan Opera Co., New York Philharmonic and Juilliard School. "Our board felt that we had a real opportunity to make a mark on the future of the performing arts with these grants an act every bit as important as our traditional areas of support in health, medical research, education and the social services," Hearst Foundation President Randolph A. Hearst said Monday in announcing the grants. Lincoln Center's share will be $200,000 for its new building, which will house young artists and provide new public facilities. The Metropolitan Opera Co. and New York Philharmonic will receive $400,000 each. The Juilliard School, where music and the performing arts are taught, will get $250,000. The Hearst Foundation, a leading supporter of the Lincoln Center Consolidated Corporate Fund, will make its usual annual $100,000 donation, too. Figure 3 shows the Dirichlet parameters of the corresponding variational distribution for those topics where ')'i > 1 (k = 100), and also lists the top 15 words (in iTo enable repeated large scale comparison of various models on large corpora, we implemented our variational inference algorithm on a parallel computing cluster. The (bottleneck) E step is distributed across nodes so that the qm for different documents are calculated in parallel. Topic 1 Topic 2 Topic 3 Topic 4 Topic 5 SCHOOL MILLION SAID SAID SAID SAID YEAR AIDS NEW NEW STUDENTS SAID HEALTH PRESIDENT MUSIC BOARD SALES DISEASE CHIEF YEAR SCHOOLS BILLION VIRUS CHAIRMAN THEATER STUDENT TOTAL CHILDREN EXECUTIVE MUSICAL TEACHER SHARE BLOOD VICE BAND POLICE EARNINGS PATIENTS YEARS PLAY PROGRAM PROFIT TREATMENT COMPANY WON TEACHERS QUARTER STUDY YORK TWO MEMBERS ORDERS IMMUNE SCHOOL AVAILABLE YEAROLD LAST CANCER TWO AWARD GANG DEC PEOPLE TODAY OPERA I" DEPARTMENT REVENUE PERCENT COLUMBIA BEST Figure 3: The Dirichlet parameters where Ii > 1 (k = 100), and the top 15 words from the corresponding topics, for the document discussed in the text. __ LDA -x- pLSI .. pLSI(00 lemper) ~ MIx1Un;grams v · ram -><--------------k (number of topics) woo ' .~ 4500 ',.. _ .l\ ! k (number of topiCS) Figure 4: Perplexity results on the CRAN and AP corpora for LDA, pLSI, mixture of unigrams, and the unigram model. order) from these topics. This document is mostly a combination of words about school policy (topic 4) and music (topic 5). The less prominent topics reflect other words about education (topic 1), finance (topic 2), and health (topic 3). 4.1 Formal evaluation: Perplexity To compare the generalization performance of LDA with other models, we computed the perplexity of a test set for the AP and CRAN corpora. The perplexity, used by convention in language modeling, is monotonically decreasing in the likelihood of the test data, and can be thought of as the inverse of the per-word likelihood. More formally, for a test set of M documents, perplexity(Vtest ) = exp (-l:m logp(wm)/ l:m Iwml}. We compared LDA to both the mixture of unigrams and pLSI described in Section 2.2. We trained the pLSI model with and without tempering to reduce overfitting. When tempering, we used part of the test set as the hold-out data, thereby giving it a slight unfair advantage. As mentioned previously, pLSI does not readily generate or assign probabilities to previously unseen documents; in our experiments, we assigned probability to a new document d by marginalizing out the dummy training set indices2 : pew ) = l:d(rr: =1l:z p(wn lz)p(zld))p(d) . 2 A second natural method, marginalizing out d and z to form a unigram model using the resulting p(w)'s, did not perform well (its performance was similar to the standard unigram model). 1-:- ~Dc.~UrUg,ams I ~ W' • M" x NaiveBaes k (number of topics) k (number of topics) Figure 5: Results for classification (left) and collaborative filtering (right) Figure 4 shows the perplexity for each model and both corpora for different values of k. The latent variable models generally do better than the simple unigram model. The pLSI model severely overfits when not tempered (the values beyond k = 10 are off the graph) but manages to outperform mixture of unigrams when tempered. LDA consistently does better than the other models. To our knowledge, these are by far the best text perplexity results obtained by a bag-of-words model. 4.2 Classification We also tested LDA on a text classification task. For each class c, we learn a separate model p(wlc) of the documents in that class. An unseen document is classified by picking argmaxcp(Clw) = argmaxcp(wlc)p(c). Note that using a simple unigram distribution for p(wlc) recovers the traditional naive Bayes classification model. Using the same (standard) subset of the WebKB dataset as used in [6], we obtained classification error rates illustrated in Figure 5 (left). In all cases, the difference between LDA and the other algorithms' performance is statistically significant (p < 0.05). 4.3 Collaborative filtering Our final experiment utilized the EachMovie collaborative filtering dataset. In this dataset a collection of users indicates their preferred movie choices. A user and the movies he chose are analogous to a document and the words in the document (respectively) . The collaborative filtering task is as follows. We train the model on a fully observed set of users. Then, for each test user, we are shown all but one of the movies that she liked and are asked to predict what the held-out movie is. The different algorithms are evaluated according to the likelihood they assign to the held-out movie. More precisely define the predictive perplexity on M test users to be exp(- ~~=llogP(WmNd lwml' ... ,Wm(Nd-l))/M). With 5000 training users, 3500 testing users, and a vocabulary of 1600 movies, we find predictive perplexities illustrated in Figure 5 (right). 5 Conclusions We have presented a generative probabilistic framework for modeling the topical structure of documents and other collections of discrete data. Topics are represented explicitly via a multinomial variable Zn that is repeatedly selected, once for each word, in a given document. In this sense, the model generates an allocation of the words in a document to topics. When computing the probability of a new document, this unknown allocation induces a mixture distribution across the words in the vocabulary. There is a many-to-many relationship between topics and words as well as a many-to-many relationship between documents and topics. While Dirichlet distributions are often used as conjugate priors for multinomials in Bayesian modeling, it is preferable to instead think of the Dirichlet in our model as a component of the likelihood. The Dirichlet random variable e is a latent variable that gives generative probabilistic semantics to the notion of a "document" in the sense that it allows us to put a distribution on the space of possible documents. The words that are actually obtained are viewed as a continuous mixture over this space, as well as being a discrete mixture over topics.3 The generative nature of LDA makes it easy to use as a module in more complex architectures and to extend it in various directions. We have already seen that collections of LDA can be used in a classification setting. If the classification variable is treated as a latent variable we obtain a mixture of LDA models, a useful model for situations in which documents cluster not only according to their topic overlap, but along other dimensions as well. Another extension arises from generalizing LDA to consider Dirichlet/multinomial mixtures of bigram or trigram models, rather than the simple unigram models that we have considered here. Finally, we can readily fuse LDA models which have different vocabularies (e.g., words and images); these models interact via a common abstract topic variable and can elegantly use both vocabularies in determining the topic mixture of a given document. Acknowledgments A. Ng is supported by a Microsoft Research fellowship. This work was also supported by a grant from Intel Corporation, NSF grant IIS-9988642, and ONR MURI N 00014-00-1-0637. References [1] D. Cohn and T. Hofmann. The missing link- A probabilistic model of document content and hypertext connectivity. In Advances in Neural Information Processing Systems 13, 2001. [2] P.J. Green and S. Richardson. Modelling heterogeneity with and without the Dirichlet process. Technical Report, University of Bristol, 1998. [3] T. Hofmann. Probabilistic latent semantic indexing. Proceedings of the Twenty-Second Annual International SIGIR Conference, 1999. [4] T. J. Jiang, J. B. Kadane, and J. M. Dickey. Computation of Carlson's multiple hypergeometric functions r for Bayesian applications. Journal of Computational and Graphical Statistics, 1:231- 251, 1992. [5] M. I. Jordan, Z. Ghahramani, T. S. Jaakkola, and L. K. Saul. Introduction to variational methods for graphical models. Machine Learning, 37:183- 233, 1999. [6] K. Nigam, A. Mccallum, S. Thrun, and T. Mitchell. Text classification from labeled and unlabeled documents using EM. Machine Learning, 39(2/3):103- 134, 2000. [7] A. Popescul, L. H. Ungar, D. M. Pennock, and S. Lawrence. Probabilistic models for unified collaborative and content-based recommendation in sparse-data environments. In Uncertainty in Artificial Intelligence, Proceedings of the Seventeenth Conference, 2001. 3These remarks also distinguish our model from the Bayesian Dirichlet/Multinomial allocation model (DMA)of [2], which is a finite alternative to the Dirichlet process. The DMA places a mixture of Dirichlet priors on p(wlz) and sets O i = 00 for all i.
|
2001
|
36
|
2,032
|
A Bayesian Network for Real-Time Musical Accompaniment Christopher Raphael Department of Mathematics and Statistics, University of Massachusetts at Amherst, Amherst, MA 01003-4515, raphael~math.umass.edu Abstract We describe a computer system that provides a real-time musical accompaniment for a live soloist in a piece of non-improvised music for soloist and accompaniment. A Bayesian network is developed that represents the joint distribution on the times at which the solo and accompaniment notes are played, relating the two parts through a layer of hidden variables. The network is first constructed using the rhythmic information contained in the musical score. The network is then trained to capture the musical interpretations of the soloist and accompanist in an off-line rehearsal phase. During live accompaniment the learned distribution of the network is combined with a real-time analysis of the soloist's acoustic signal, performed with a hidden Markov model, to generate a musically principled accompaniment that respects all available sources of knowledge. A live demonstration will be provided. 1 Introduction We discuss our continuing work in developing a computer system that plays the role of a musical accompanist in a piece of non-improvisatory music for soloist and accompaniment. The system begins with the musical score to a given piece of music. Then, using training for the accompaniment part as well as a series of rehearsals, we learn a performer-specific model for the rhythmic interpretation of the composition. In performance, the system takes the acoustic signal of the live player and generates the accompaniment around this signal, in real-time, while respecting the learned model and the constraints imposed by the score. The accompaniment played by our system responds both flexibly and expressively to the soloist's musical interpretation. Our system is composed of two high level tasks we call "Listen" and "Play." Listen takes as input the acoustic signal of the soloist and, using a hidden Markov model, performs a real-time analysis of the signal. The output of Listen is essentially a running commentary on the acoustic input which identifies note boundaries in the solo part and communicates these events with variable latency. The HMM framework is well-suited to the listening task and has several attributes we regard as indispensable to any workable solution: 1. The HMM allows unsupervised training using the Baum-Welch algorithm. Thus we can automatically adapt to changes in solo instrument, microphone placement, ambient noise, room acoustics, and the sound of the accompaniment instrument. 2. Musical accompaniment is inherently a real-time problem. Fast dynamic programming algorithms provide the computational efficiency necessary to process the soloist's acoustic signal at a rate consistent with the real-time demands of our application. 3. Musical signals are occasionally ambiguous locally in time, but become easier to parse when more context is considered. Our system owes much of its accuracy to the probabilistic formulation of the HMM. This formulation allows one to compute the probability that an event is in the past. We delay the estimation of the precise location of an event until we are reasonably confident that it is, in fact, past. In this way our system achieves accuracy while retaining the lowest latency possible in the identification of musical events. Our work on the Listen component is documented thoroughly in [1] and we omit a more detailed discussion here. The heart of our system, the Play component, develops a Bayesian network consisting of hundreds of Gaussian random variables including both observable quantities, such as note onset times, and unobservable quantities, such as local tempo. The network can be trained during a rehearsal phase to model both the soloist's and accompanist's interpretations of a specific piece of music. This model then forms the backbone of a principled real-time decision-making engine used in performance. We focus here on the Play component which is the most challenging part of our system. A more detailed treatment of various aspects of this work is given in [2- 4]. 2 Knowledge Sources A musical accompaniment requires the synthesis of a number of different knowledge sources. From a modeling perspective, the fundamental challenge of musical accompaniment is to express these disparate knowledge sources in terms of a common denominator. We describe here the three knowledge sources we use. 1. We work with non-improvisatory music so naturally the musical score, which gives the pitches and relative durations of the various notes, as well as points of synchronization between the soloist and accompaniment, must figure prominently in our model. The score should not be viewed as a rigid grid prescribing the precise times at which musical events will occur; rather, the score gives the basic elastic material which will be stretched in various ways to to produce the actual performance. The score simply does not address most interpretive aspects of performance. 2. Since our accompanist must follow the soloist, the output of the Listen component, which identifies note boundaries in the solo part, constitutes our second knowledge source. While most musical events, such as changes between neighboring diatonic pitches, can be detected very shortly after the change of note, some events, such as rearticulations and octave slurs, are much less obvious and can only be precisely located with the benefit of longer term hindsight. With this in mind, we feel that any successful accompaniment system cannot synchronize in a purely responsive manner. Rather it must be able to predict the future using the past and base its synchronization on these predictions, as human musicians do. 3. While the same player's performance of a particular piece will vary from rendition to rendition, many aspects of musical interpretation are clearly established with only a few repeated examples. These examples, both of solo performances and human (MIDI) performances of the accompaniment part constitute the third knowledge source for our system. The solo data is used primarily to teach the system how to predict the future evolution of the solo part. The accompaniment data is used to learn the musicality necessary to bring the accompaniment to life. We have developed a probabilistic model, a Bayesian network, that represents all of these knowledge sources through a jointly Gaussian distribution containing hundreds of random variables. The observable variables in this model are the estimated soloist note onset times produced by Listen and the directly observable times for the accompaniment notes. Between these observable variables lies a layer of hidden variables that describe unobservable quantities such as local tempo, change in tempo, and rhythmic stress. 3 A Model for Rhythmic Interpretation We begin by describing a model for the sequence of note onset times generated by a monophonic (single voice) musical instrument playing a known piece of music. For each of the notes, indexed by n = 0, . . . ,N, we define a random vector representing the time, tn, (in seconds) at which the note begins, and the local "tempo," Sn, (in secs. per measure) for the note. We model this sequence ofrandom vectors through a random difference equation: (1) n = 0, ... , N - 1, where in is the musical length of the nth note, in measures, and the {(Tn' CTnY} and (to, so)t are mutually independent Gaussian random vectors. The distributions of the {CTn} will tend concentrate around ° expressing the notion that tempo changes are gradual. The means and variances of the {CT n} show where the soloist is speeding-up (negative mean), slowing-down (positive mean), and tell us if these tempo changes are nearly deterministic (low variance), or quite variable (high variance). The { Tn} variables describe stretches (positive mean) or compressions (negative mean) in the music that occur without any actual change in tempo, as in a tenuto or agogic accent. The addition of the {Tn} variables leads to a more musically plausible model, since not all variation in note lengths can be explained through tempo variation. Equally important, however, the {Tn} variables stabilize the model by not forcing the model to explain, and hence respond to, all note length variation as tempo variation. Collectively, the distributions of the (Tn' CTn)t vectors characterize the solo player's rhythmic interpretation. Both overall tendencies (means) and the repeatability of these tendencies (covariances) are captured by these distributions. 3.1 Joint Model of Solo and Accompaniment In modeling the situation of musical accompaniment we begin with the our basic rhythm model of Eqn. 1, now applied to the composite rhythm. More precisely, Listen Update Composite Accomp Figure 1: A graphical description of the dependency structure of our model. The top layer of the graph corresponds to the solo note onset times detected by Listen. The 2nd layer of the graph describes the ( Tn, 0" n) variables that characterize the rhythmic interpretation. The 3rd layer of the graph is the time-tempo process {(Sn, tn)}. The bottom layer is the observed accompaniment event times. let mo, ... , mivs and mg, ... , m'Na denote the positions, in measures, of the various solo and accompaniment events. For example, a sequence of quarter notes in 3/ 4 time would lie at measure positions 0, 1/ 3, 2/ 3, etc. We then let mo, ... , mN be the sorted union of these two sets of positions with duplicate times removed; thus mo < ml < ... < mN· We then use the model of Eqn. 1 with In = mn+1 - m n, n = 0, . . . , N - 1. A graphical description of this model is given in the middle two layers of Figure 1. In this figure, the layer labeled "Composite" corresponds to the time-tempo variables, (tn, sn)t, for the composite rhythm, while the layer labeled "Update" corresponds to the interpretation variables ( Tn, 0" n) t. The directed arrows of this graph indicate the conditional dependency structure of our model. Thus, given all variables "upstream" of a variable, x, in the graph, the conditional distribution of x depends only on the parent variables. Recall that the Listen component estimates the times at which solo notes begin. How do these estimates figure into our model? We model the note onset times estimated by Listen as noisy observations of the true positions {tn}. Thus if m n is a measure position at which a solo note occurs, then the corresponding estimate from Listen is modeled as an = tn + an where an rv N(O, 1I2). Similarly, if m n is the measure position of an accompaniment event, then we model the observed time at which the event occurs as bn = tn + f3n where f3n rv N(O, ",2). These two collections of observable variables constitute the top layer of our figure, labeled "Listen," and the bottom layer, labeled "Accomp." There are, of course, measure positions at which both solo and accompaniment events should occur. If n indexes such a time then an and bn will both be noisy observations of the true time tn. The vectors/ variables {(to, so)t, (Tn ' O"n)t, an, f3n} are assumed to be mutually independent. 4 Training the Model Our system learns its rhythmic interpretation by estimating the parameters of the (Tn,O"n) variables. We begin with a collection of J performances of the accompaniment part played in isolation. We refer to the model learned from this accompaniment data as the "practice room" distribution since it reflects the way the accompanist plays when the constraint of following the soloist is absent. For each Listen Update Composite Accomp Figure 2: Conditioning on the observed accompaniment performance (darkened circles), we use the message passing algorithm to compute the conditional distributions on the unobservable {Tn' O"n} variables. such performance, we treat the sequence of times at which accompaniment events occur as observed variables in our model. These variables are shown with darkened circles in Figure 2. Given an initial assignment of of means and covariances to the (Tn ,O"n) variables, we use the "message passing" algorithm of Bayesian Networks [8,9] to compute the conditional distributions (given the observed performance) of the (Tn,O"n) variables. Several such performances lead to several such estimates, enabling us to improve our initial estimates by reestimating the (Tn' O"n) parameters from these conditional distributions. More specifically, we estimate the (Tn,O"n) parameters using the EM algorithm, as follows, as in [7]. We let J-L~, ~~ be our initial mean and covariance matrix for the vector ( Tn, 0" n). The conditional distribution of ( Tn, 0" n) given the jth accompaniment performance, and using {J-L~ , ~~} , has a N(m;,n, S~ ) distribution where the m;,n and S~ parameters are computed using the message passing algorithm. We then update our parameter estimates by 1 J . } Lmj,n j = l ~ i+ l n The conventional wisdom of musicians is that the accompaniment should follow the soloist. In past versions of our system we have explicitly modeled the asymmetric roles of soloist and accompaniment through a rather complicated graph structure [2- 4]. At present we deal with this asymmetry in a more ad hoc, however, perhaps more effective, manner, as follows. Training using the accompaniment performances allows our model to learn some of the musicality these performances demonstrate. Since the soloist's interpretation must take precedence, we want to use this accompaniment interpretation only to the extent that it does not conflict with that of the soloist. We accomplish this by first beginning with the result of the accompaniment training described above. We use the practice room distributions, (the distributions on the {(Tn, O"n)} learned from the accompaniment data) , as the initial distributions, {J-L~ , ~~} . We then run the EM algorithm as described above now treating the currently available collection of solo performances as the observed data. During this phase, only those parameters relevant to the soloist's rhythmic interpretation will be modified significantly. Parameters describing the interpretation of a musical segment in which the soloist is mostly absent will be largely unaffected by the second training pass. Listen Update Composite Accomp Figure 3: At any given point in the performance we will have observed a collection of solo note times estimated estimated by Listen, and the accompaniment event times (the darkened circles). We compute the conditional distribution on the next unplayed accompaniment event, given these observations. This solo training actually happens over the course of a series of rehearsals. We first initialize our model to the practice room distribution by training with the accompaniment data. Then we iterate the process of creating a performance with our system, (described in the next section), extracting the sequence of solo note onset times in an off-line estimation process, and then retraining the model using all currently available solo performances. In our experience, only a few such rehearsals are necessary to train a system that responds gracefully and anticipates the soloist's rhythmic nuance where appropriate generally less than 10. 5 Real Time Accompaniment The methodological key to our real-time accompaniment algorithm is the computation of (conditional) marginal distributions facilitated by the message-passing machinery of Bayesian networks. At any point during the performance some collection of solo notes and accompaniment notes will have been observed, as in Fig. 3. Conditioned on this information we can compute the distribution on the next unplayed accompaniment. The real-time computational requirement is limited by passing only the messages necessary to compute the marginal distribution on the pending accompaniment note. Once the conditional marginal distribution of the pending accompaniment note is calculated we schedule the note accordingly. Currently we schedule the note to be played at the conditional mean time, given all observed information, however other reasonable choices are possible. Note that this conditional distribution depends on all of the sources of information included in our model: The score information, all currently observed solo and accompaniment note times, and the rhythmic interpretations demonstrated by both the soloist and accompanist captured during the training phase. The initial scheduling of each accompaniment note takes place immediately after the previous accompaniment note is played. It is possible that a solo note will be detected before the pending accompaniment is played; in this event the pending accompaniment event is rescheduled by recomputing the its conditional distribution using the newly available information. The pending accompaniment note is rescheduled each time an additional solo note is detected until its currently scheduled time arrives, at which time it is finally played. In this way our accompaniment makes use of all currently available information. Does our system pass the musical equivalent of the Turing Test? We presume no more objectivity in answering this question than we would have in judging the merits of our other children. However, we believe that the level of musicality attained by our system is truly surprising, while the reliability is sufficient for live demonstration. We hope that the interested reader will form an independent opinion, even if different from ours, and to this end we have made musical examples demonstrating our progress available on the web page: http://fafner.math.umass.edu/musicplus_one. Acknowledgments This work supported by NSF grants IIS-998789 and IIS-0113496. References [1] Raphael C. (1999), "Automatic Segmentation of Acoustic Musical Signals Using Hidden Markov Models," IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 21, No.4, pp. 360-370. [2] Raphael C. (2001), "A Probabilistic Expert System for Automatic Musical Accompaniment," Journal of Computational and Graphical Statistics, vol. 10 no. 3, 487-512. [3] Raphael C. (2001), "Can the Computer Learn to Play Expressively?" Proceedings of Eighth International Workshop on Artificial Intelligence and Statistics, 113-120, Morgan Kauffman. [4] Raphael C. (2001), "Synthesizing Musical Accompaniments with Bayesian Belief Networks," Journal of New Music Research, vol. 30, no. 1, 59-67. [5] Spiegelhalter D., Dawid A. P., Lauritzen S., Cowell R. (1993), "Bayesian Analysis in Expert Systems," Statistical Science, Vol. 8, No.3, pp. 219-283. [6] Cowell R., Dawid A. P., Lauritzen S., Spiegelhalter D. (1999), "Probabilistic Networks and Expert Systems," Springer, New York. [7] Lauritzen S. L. (1995), "The EM Algorithm for Graphical Association Models with Missing Data," Computational Statistics and Data Analysis, Vol. 19, pp. 191-20l. [8] Lauritzen S. L. (1992), "Propagation of Probabilities, Means, and Variances in Mixed Graphical Association Models," Journal of the American Statistical Association, Vol. 87, No. 420, (Theory and Methods), pp. 1098-1108. [9] Lauritzen S. L. and F. Jensen (1999), "Stable Local Computation with Conditional Gaussian Distributions," Technical Report R-99-2014, Department of Mathematic Sciences, Aalborg University.
|
2001
|
37
|
2,033
|
Efficiency versus Convergence of Boolean Kernels for On-Line Learning Algorithms Roni Khardon Tufts University Medford, MA 02155 roni@eecs.tufts.edu Dan Roth University of Illinois Urbana, IL 61801 danr@cs.uiuc.edu Rocco Servedio Harvard University Cambridge, MA 02138 rocco@deas.harvard.edu Abstract We study online learning in Boolean domains using kernels which capture feature expansions equivalent to using conjunctions over basic features. We demonstrate a tradeoff between the computational efficiency with which these kernels can be computed and the generalization ability of the resulting classifier. We first describe several kernel functions which capture either limited forms of conjunctions or all conjunctions. We show that these kernels can be used to efficiently run the Perceptron algorithm over an exponential number of conjunctions; however we also prove that using such kernels the Perceptron algorithm can make an exponential number of mistakes even when learning simple functions. We also consider an analogous use of kernel functions to run the multiplicative-update Winnow algorithm over an expanded feature space of exponentially many conjunctions. While known upper bounds imply that Winnow can learn DNF formulae with a polynomial mistake bound in this setting, we prove that it is computationally hard to simulate Winnow’s behavior for learning DNF over such a feature set, and thus that such kernel functions for Winnow are not efficiently computable. 1 Introduction The Perceptron and Winnow algorithms are well known learning algorithms that make predictions using a linear function in their feature space. Despite their limited expressiveness, they have been applied successfully in recent years to several large scale real world classification problems. The SNoW system [7, 2], for example, has successfully applied variations of Perceptron [6] and Winnow [4] to problems in natural language processing. The system first extracts Boolean features from examples (given as text) and then runs learning algorithms over restricted conjunctions of these basic features. There are several ways to enhance the set of features after the initial extraction. One idea is to expand the set of basic features using conjunctions such as and use these expanded higher-dimensional examples, in which each conjunction plays the role of a basic feature, for learning. This approach clearly leads to an increase in expressiveness and thus may improve performance. However, it also dramatically increases the number of features (from
to if all conjunctions are used) and thus may adversely affect both the computation time and convergence rate of learning. This paper studies the computational efficiency and convergence of the Perceptron and Winnow algorithms over such expanded feature spaces of conjunctions. Specifically, we study the use of kernel functions to expand the feature space and thus enhance the learning abilities of Perceptron and Winnow; we refer to these enhanced algorithms as kernel Perceptron and kernel Winnow. 1.1 Background: Perceptron and Winnow Throughout its execution Perceptron maintains a weight vector which is initially Upon receiving an example
the algorithm predicts according to the linear threshold function If the prediction is and the label is (false positive prediction) then the vector is set to , while if the prediction is and the label is (false negative) then is set to No change is made if the prediction is correct. The famous Perceptron Convergence Theorem [6] bounds the number of mistakes which the Perceptron algorithm can make: Theorem 1 Let be a sequence of labeled examples with ! "# $ $%& and (') +* for all , . Let -. /10 be such that -1 "/ for all , Then Perceptron makes at most 243658795:3 ; 3 mistakes on this example sequence. The Winnow algorithm [4] has a very similar structure. Winnow maintains a hypothesis vector <# which is initially >= Winnow is parameterized by a promotion factor ? and a threshold @ 0 BA upon receiving an example C' +*D Winnow predicts according to the threshold function ( @ If the prediction is and the label is then for all , such that =E the value of is set to GF ? ; this is a demotion step. If the prediction is and the label is then for all , such that =H the value of is set to ?I ; this is a promotion step. No change is made if the prediction is correct. For our purposes the following mistake bound, implicit in [4], is of interest: Theorem 2 Let the target function be a J -literal monotone disjunction K = 8LM NN M PO For any sequence of examples in ' +*D labeled according to K the number of prediction mistakes made by Winnow ? @ is at most Q QBR S J ?TU VXWZY\[ Q @ 1.2 Our Results Our first result in Section 2 shows that it is possible to efficiently run the kernel Perceptron algorithm over an exponential number of conjunctive features: Theorem 3 There is an algorithm that simulates Perceptron over the -dimensional feature space of all conjunctions of
basic features. Given a sequence of ] labeled examples in ' ^* the prediction and update for each example take poly
] time steps. This result is closely related to one of the main open problems in learning theory: efficient learnability of disjunctions of conjunctions, or DNF (Disjunctive Normal Form) expressions.1 Since linear threshold elements can represent disjunctions (e.g. M !_ M is true iff !_ ), Theorems 1 and 3 imply that kernel Perceptron can be used to learn DNF. However, in this framework the values of ` and & in Theorem 1 can be exponentially large, and hence the mistake bound given by Theorem 1 is exponential rather than polynomial in
The question thus arises whether, for kernel Perceptron, the exponential 1Angluin [1] proved that DNF expressions cannot be learned efficiently using hypotheses which are themselves DNF expressions from equivalence queries and thus also in the mistake bound model which we are considering here. However this result does not preclude the efficient learnability of DNF using a different class of hypotheses such as those generated by the kernel Perceptron algorithm. upper bound implied by Theorem 1 is essentially tight. We give an affirmative answer, thus showing that kernel Perceptron cannot efficiently learn DNF: Theorem 4 There is a monotone DNF K over and a sequence of examples labeled according to K which causes the kernel Perceptron algorithm to make mistakes. Turning to Winnow, an attractive feature of Theorem 2 is that for suitable ? @ the bound is logarithmic in the total number of features ` (e.g. ?= and @ =H` ). Therefore, as noted by several researchers [5], if a Winnow analogue of Theorem 3 could be obtained this would imply efficient learnability of DNF. We show that no such analogue can exist: Theorem 5 There is no polynomial time algorithm which simulates Winnow over exponentially many monotone conjunctive features for learning monotone DNF, unless every problem in #P can be solved in polynomial time. We observe that, in contrast to Theorem 5, Maass and Warmuth have shown that the Winnow algorithm can be simulated efficiently over exponentially many conjunctive features for learning some simple geometric concept classes [5]. While several of our results are negative, in practice one can achieve good performance by using kernel Perceptron (if
is small) or the limited-conjunction kernel described in Section 2 (if
is large). This is similar to common practice with polynomial kernels2 where typically a small degree is used to aid convergence. These observations are supported by our preliminary experiments in an NLP domain which are not reported here. 2 Theorem 3: Kernel Perceptron with Exponentially Many Features It is easily observed, and well known, that the hypothesis of the Perceptron algorithm is a sum of the previous examples on which prediction mistakes were made. If we let "' ^* denote the label of example , then =
where is the set of examples on which the algorithm made a mistake. Thus the prediction of Perceptron on is 1 iff U =
B =
. For an example ' +* let denote its transformation into an enhanced feature space such as the space of all conjunctions. To run the Perceptron algorithm over the enhanced space we must predict iff where is the weight vector in the enhanced space; from the above discussion this holds iff
B B . Denoting = B this holds iff
. Thus we never need to construct the enhanced feature space explicitly; we need only be able to compute the kernel function efficiently. This is the idea behind all so-called kernel methods, which can be applied to any algorithm (such as support vector machines) whose prediction is a function of inner products of examples; see e.g. [3] for a discussion. The result in Theorem 3 is simply obtained by presenting a kernel function capturing all conjunctions. We also describe kernels for all monotone conjunctions which allow no negative literals, and kernels capturing all (monotone) conjunctions of up to J literals. The general case: When includes all conjunctions (with positive and negative literals) must compute the number of conjunctions which are true in both and . Clearly, any literal in such a conjunction must satisfy both and and thus the corresponding bit in must have the same value. Counting all such conjunctions gives = "!$# &%' ( where )+*-,/. is the number of original features that have the same value in and . This kernel has been obtained independently by [8]. 2Our Boolean kernels are different than standard polynomial kernels in that all the conjunctions are weighted equally. While expressive power does not change, convergence and behavior, do. Monotone Monomials: In some applications the total number
of basic features may be very large but in any one example only a small number of features take value 1. In other applications the number of features
may not be known in advance (e.g. due to unseen words in text domains). In these cases it may be useful to consider only monotone monomials. To express all monotone monomials we take = +! # &%' ( where )+*,/.!Y) is the number of active features common to both and . A parameterized kernel: In general, one may want to trade off expressivity against number of examples and convergence time. Thus we consider a parameterized kernel which captures all conjunctions of size at most J for some J
The number of such conjunctions that satisfy both and is =
"!$# &%' ( . This kernel is reported also in [10]. For monotone conjunctions of size at most J we have = +! # &%' ( . 3 Theorem 4: Kernel Perceptron with Exponentially Many Mistakes We describe a monotone DNF target function and a sequence of labeled examples which cause the monotone kernel Perceptron algorithm to make exponentially many mistakes. For C' +* we write to denote the number of 1’s in and . to denote )+*,/.!Y) We use the following lemma (constants have not been optimized): Lemma 6 There is a set of
-bit strings =E' *' +* with ] = such that \=
F for % , % ] and ! #" %
F%$ for % ,&(' % ] Proof: The proof uses the probabilistic method. For each ,V= ] let ! X' +* be chosen by independently setting each bit to with probability 1/10. For any , it is clear that )+*, =
F A a Chernoff bound implies that .0/%*1 !
F % R 2 and thus the probability that any satisfies 3
F is at most ]4 R 2 Similarly, for any ,5 =6' we have )+*1 7 #" -#=
F DBA a Chernoff bound implies that .8/%*1 9 #" 0
F:$ % R 2 and thus the probability that any " with ,;5 =6' satisfies " 0
F%$ is at most _ R <2 For ] == the value of _ R 3 <2 ]4 R 2 is less than 1. Thus for some choice of we have each
F and # #" %
F:$ For any which has 0
F we can set ^
F of the 1s to 0s, and the lemma is proved. The target DNF is very simple: it is the single conjunction !_ While the original Perceptron algorithm over the
features makes at most poly
mistakes for this target function, we now show that the monotone kernel Perceptron algorithm which runs over all monotone monomials can make > mistakes. Recall that at the beginning of the Perceptron algorithm’s execution all coordinates of are 0. The first example is the negative example A since = Perceptron incorrectly predicts 1 on this example. The resulting update causes the coefficient ? corresponding to the empty monomial (satisfied by any example ) to become but all other coordinates of remain 0. The next example is the positive example For this example we have = so Perceptron incorrectly predicts Since all monotone conjunctions are satisfied by this example the resulting update causes ? to become 0 and all other coordinates of to become 1. The next examples are the vectors described in Lemma 6. Since each such example has )=
F each example is negative; however as we now show the Perceptron algorithm will predict on each of these examples. Fix any value % , % and consider the hypothesis vector just before example is received. Since !=
F the value of is a sum of the _ different coordinates which correspond to the monomials satisfied by ! More precisely we have =
) where contains the monomials which are satisfied by and #" for some '65 =E, and contains the monomials which are satisfied by but no #" with '5 =, We lower bound the two sums separately. Let be any monomial in By Lemma 6 any H contains at most
F%$ variables and thus there can be at most 2 _ monomials in Using the well known bound Q
" ; " % Q where ? % F and ? is the binary entropy function there can be at most terms in Moreover the value of each must be at least 34 since decreases by at most 1 for each example, and hence
; 34 0 _ On the other hand, for any we clearly have = By Lemma 6 for any 0
F:$ every -variable monomial satisfied by must belong to and hence
3_ 2 _ 0 Combining these inequalities we have _ < 0 and hence the Perceptron prediction on is 1. 4 Theorem 5: Learning DNF with Kernel Winnow is Hard In this section, for .' ^* denotes the -element vector whose coordinates are all nonempty monomials (monotone conjunctions) over A sequence of labeled examples is monotone consistent if it is consistent with some monotone function, i.e. % " for all J =
implies % " If is monotone consistent and has ] labeled examples then clearly there is a monotone DNF formula consistent with which contains at most ] conjunctions. We consider the following problem: KERNEL WINNOW PREDICTION ? @ (KWP) Instance: Monotone consistent sequence C= of labeled examples with each .' ^* and each .') +* A unlabeled example ' +* Question: Is @ where is the ` = . -dimensional hypothesis vector generated by running Winnow ? @ on the example sequence ? In order to run Winnow over all
. nonempty monomials to learn monotone DNF, one must be able to solve KWP efficiently. The main result of this section is proved by showing that KWP is computationally hard for any parameter settings which yield a polynomial mistake bound for Winnow via Theorem 2. Theorem 7 Let ` =
U and ? 0 @ be such that , * Q QBR S ? > WZY\[ Q @ = poly ! Then KWP ? @ is #P-hard. Proof of Theorem 7: For ` ? and @ as described above it can easily be verified that poly ? poly ! and _#" poly @ poly The proof of the theorem is a reduction from the following #P-hard problem [9]: (See [9] also for details on #P.) MONOTONE 2-SAT (M2SAT) Instance: Monotone 2-CNF Boolean formula $=&% (' % _)' *' % with % = 8L4M 3 and each ,+ ' * A integer such that % % Question: Is $ R i.e. does $ have at least satisfying assignments in ' +* ? 4.1 High-Level Idea of the Proof The high level idea of the proof is simple: let $ be an instance of M2SAT where $ is defined over variables The Winnow algorithm maintains a weight for each monomial over variables We define a 1-1 correspondence between these monomials and truth assignments ' +* for $ and we give a sequence of examples for Winnow which causes . if $ = and = if $ = The value of is thus related to $ R A some additional work ensures that @ if and only if $ R In more detail, let =
WZYD[ Q W YD[ ? = Q =
_ Q and =
C
_ We describe a polynomial time transformation which maps an
-variable instance $ of M2SAT to an -variable instance of KWP ? @ where X=E is monotone consistent, each and belong to ' ^* and @ if and only if $ R The Winnow variables are divided into three sets and where = ' * = '
* and = '
* The unlabeled example is
R R
i.e. all variables in and are set to 1 and all variables in are set to 0. We thus have = ( where = ? = ? and ( = ' ? ' ? We refer to monomials 5 = as type monomials, monomials 5 = as type monomials, and monomials ! 5 = 5 = as type monomials. The example sequence is divided into four stages. Stage 1 results in $ R A as described below the
variables in correspond to the
variables in the CNF formula $ Stage 2 results in ?" $ R for some positive integer # Stages 3 and 4 together result in ( @ ? " Thus the final value of is approximately @# ? " $ R ^ so we have @ if and only if $ R Since all variables in are 0 in if includes a variable in then the value of does not affect The variables in are “slack variables” which (i) make Winnow perform the correct promotions/demotions and (ii) ensure that is monotone consistent. 4.2 Details of the Proof Stage 1: Setting $&%('*),+.-0/214365) . We define the following correspondence between truth assignments ' +* and monomials 7 &8 = if and only if is not present in For each clause L M 3 in $ Stage 1 contains negative examples such that 8L = 3 = and = for all other Assuming that (1) Winnow makes a false positive prediction on each of these examples and (2) in Stage 1 Winnow never does a promotion on any example which has any variable in set to 1, then after Stage 1 we will have that = if $ = and < |