index
int64 0
20.3k
| text
stringlengths 0
1.3M
| year
stringdate 1987-01-01 00:00:00
2024-01-01 00:00:00
| No
stringlengths 1
4
|
|---|---|---|---|
1,900
|
Active Support Vector Machine Classification o. L. Mangasarian Computer Sciences Dept. University of Wisconsin 1210 West Dayton Street Madison, WI 53706 olvi@cs.wisc.edu David R. Musicant Dept. of Mathematics and Computer Science Carleton College One North College Street Northfield, MN 55057 dmusican@carleton.edu Abstract An active set strategy is applied to the dual of a simple reformulation of the standard quadratic program of a linear support vector machine. This application generates a fast new dual algorithm that consists of solving a finite number of linear equations, with a typically large dimensionality equal to the number of points to be classified. However, by making novel use of the Sherman-MorrisonWoodbury formula, a much smaller matrix of the order of the original input space is inverted at each step. Thus, a problem with a 32-dimensional input space and 7 million points required inverting positive definite symmetric matrices of size 33 x 33 with a total running time of 96 minutes on a 400 MHz Pentium II. The algorithm requires no specialized quadratic or linear programming code, but merely a linear equation solver which is publicly available. 1 Introduction Support vector machines (SVMs) [23, 5, 14, 12] are powerful tools for data classification. Classification is achieved by a linear or nonlinear separating surface in the input space of the dataset. In this work we propose a very fast simple algorithm, based on an active set strategy for solving quadratic programs with bounds [18]. The algorithm is capable of accurately solving problems with millions of points and requires nothing more complicated than a commonly available linear equation solver [17, 1, 6] for a typically small (100) dimensional input space of the problem. Key to our approach are the following two changes to the standard linear SVM: 1. Maximize the margin (distance) between the parallel separating planes with respect to both orientation (w) as well as location relative to the origin b). See equation (7) below. Such an approach was also successfully utilized in the successive overrelaxation (SOR) approach of [15] as well as the smooth support vector machine (SSVM) approach of [12]. 2. The error in the soft margin (y) is minimized using the 2-norm squared instead of the conventional 1-norm. See equation (7). Such an approach has also been used successfully in generating virtual support vectors [4]. These simple, but fundamental changes, lead to a considerably simpler positive definite dual problem with nonnegativity constraints only. See equation (8). In Section 2 of the paper we begin with the standard SVM formulation and its dual and then give our formulation and its simpler dual. We corroborate with solid computational evidence that our simpler formulation does not compromise on generalization ability as evidenced by numerical tests in Section 4 on 6 public datasets. See Table 1. Section 3 gives our active support vector machine (ASVM) Algorithm 3.1 which consists of solving a system of linear equations in m dual variables with a positive definite matrix. By invoking the Sherman-Morrison-Woodbury (SMW) formula (1) we need only invert an (n + 1) x (n + 1) matrix where n is the dimensionality of the input space. This is a key feature of our approach that allows us to solve problems with millions of points by merely inverting much smaller matrices of the order of n. In concurrent work [8] Ferris and Munson also use the SMW formula but in conjunction with an interior point approach to solve massive problems based on our formulation (8) as well as the conventional formulation (6). Burges [3] has also used an active set method, but applied to the standard SVM formulation (2) instead of (7) as we do here. Both this work and Burges' appeal, in different ways, to the active set computational strategy of More and Toraldo [18]. We note that an active set computational strategy bears no relation to active learning. Section 4 describes our numerical results which indicate that the ASVM formulation has a tenfold testing correctness that is as good as the ordinary SVM, and has the capability of accurately solving massive problems with millions of points that cannot be attacked by standard methods for ordinary SVMs. We now describe our notation and give some background material. All vectors will be column vectors unless transposed to a row vector by a prime I. For a vector x E Rn, x+ denotes the vector in Rn with all of its negative components set to zero. The notation A E Rmxn will signify a real m x n matrix. For such a matrix A' will denote the transpose of A and Ai will denote the i-th row of A. A vector of ones or zeroes in a real space of arbitrary dimension will be denoted by e or 0, respectively. The identity matrix of arbitrary dimension will be denoted by I. For two vectors x and y in Rn, x ..1 y denotes orthogonality, that is x' y = O. For U E R m, Q E Rmxm and B C {I, 2, ... , m}, UB denotes UiEB, QB denotes QiEB and QBB denotes a principal submatrix of Q with rows i E B and columns j E B. The notation argminxEs f(x) denotes the set of minimizers in the set S of the real-valued function f defined on S. We use := to denote definition. The 2-norm of a matrix Q will be denoted by IIQI12. A separating plane, with respect to two given point sets A and B in Rn , is a plane that attempts to separate Rn into two halfspaces such that each open halfspace contains points mostly of A or B. A special case of the Sherman-Morrison-Woodbury (SMW) formula [9] will be utilized: (Ilv + HH')-l = v(I - H(Ilv + H'H)-l H'), (1) where v is a positive number and H is an arbitrary m x k matrix. This formula enables us to invert a large m x m matrix by merely inverting a smaller k x k matrix. 2 The Linear Support Vector Machine We consider the problem of classifying m points in the n-dimensional real space R n , represented by the m x n matrix A, according to membership of each point Ai in the class A+ or A- as specified by a given m x m diagonal matrix D with +l's or -1 's along its diagonal. For this problem the standard SVM with a linear kernel [23, 5] is given by the following quadratic program with parameter v > 0: . 1 mm ve'y + -w'w s.t. D(Aw - e-y) + y 2:: e, y 2:: O. (2) (w,'Y,y)ERn+l+= 2 x'w = 1 + 1 x o 0 000 o 0 0 Ox A0 0000 X'W = 1-1 0 o ( M · 2 argln= IIwl12 x x x A+ x x x x x x x'w =1 Figure 1: The bounding planes (3) with a soft (i.e. with some errors) margin 2/llwI12, and the plane (4) approximately separating A+ from A-. Here w is the normal to the bounding planes: x'w = 'Y ± 1 (3) and'Y determines their location relative to the origin (Figure 1.) The plane x'w = 'Y + 1 bounds the A+ points, possibly with error, and the plane x'w = 'Y -1 bounds the A - points, also possibly with some error. The separating surface is the plane: x'w = 'Y, (4) midway between the bounding planes (3). The quadratic term in (2), is twice the reciprocal of the square of the 2-norm distance 2/llw112 between the two bounding planes of (3) (see Figure 1). This term maximizes this distance which is often called the "margin". If the classes are linearly inseparable, as depicted in Figure 1, then the two planes bound the two classes with a "soft margin". That is, they bound each set approximately with some error determined by the nonnegative error variable y: ~ 'Y + 1, for Dii = 1, ::; 'Y - 1, for Dii = - 1. (5) Traditionally the I-norm of the error variable y is minimized parametrically with weight v in (2) resulting in an approximate separation as depicted in Figure 1. The dual to the standard quadratic linear SVM (2) [13, 22, 14, 7] is the following: . 1 mill - u'DAA'Du - e'u s.t. e'Du = 0, 0 < u < ve. uER=2 (6) The variables (w, 'Y) of the primal problem which determine the separating surface (4) can be obtained from the solution of the dual problem above [15, Eqns. 5 and 7]. We note immediately that the matrix DAA'D appearing in the dual objective function (6) is not positive definite in general because typically m > > n. Also, there is an equality constraint present, in addition to bound constraints, which for large problems necessitates special computational procedures such as SMO [21]. Furthermore, a one-dimensional optimization problem [15] must be solved in order to determine the locator 'Y of the separating surface (4). In order to overcome all these difficulties as well as that of dealing with the necessity of having to essentially invert a very large matrix of the order of m x m , we propose the following simple but critical modification of the standard SVM formulation (2). We change Ily lll to Ilyll§ which makes the constraint y ~ 0 redundant. We also append the term 'Y2 to w'w. This in effect maximizes the margin between the parallel separating planes (3) with respect to both wand 'Y [15], that is with respect to both orientation and location of the planes, rather that just with respect to w which merely determines the orientation of the plane. This leads to the following reformulation of the SVM: y'y 1 min v- + -(w'w + ,2) s.t. D(Aw - er) + y ~ e. (7) (w ,'Y,y)ERn+l+", 2 2 the dual of this problem is [13]: 1 I min -u'( - + D(AA' + ee')D)u - e'u. O~uER'" 2 v (8) The variables (w,,) of the primal problem which determine the separating surface (4) are recovered directly from the solution of the dual (8) above by the relations: w=A'Du, y=u/v, ,=-e'Du. (9) We immediately note that the matrix appearing in the dual objective function is positive definite and that there is no equality constraint and no upper bound on the dual variable u. The only constraint present is a simple nonnegativity one. These facts lead us to our simple finite active set algorithm which requires nothing more sophisticated than inverting an (n + 1) x (n + 1) matrix at each iteration in order to solve the dual problem (8). 3 ASVM (Active Support Vector Machine) Algorithm The algorithm consists of determining a partition of the dual variable u into nonbasic and basic variables. The nonbasic variables are those which are set to zero. The values of the basic variables are determined by finding the gradient of the objective function of (8) with respect to these variables, setting this gradient equal to zero, and solving the resulting linear equations for the basic variables. If any basic variable takes on a negative value after solving the linear equations, it is set to zero and becomes nonbasic. This is the essence of the algorithm. In order to make the algorithm converge and terminate, a few additional safeguards need to be put in place in order to allow us to invoke the More-Toraldo finite termination result [18]. The other key feature of the algorithm is a computational one and makes use of the SMW formula. This feature allows us to invert an (n + 1) x (n + 1) matrix at each step instead of a much bigger matrix of order m x m. Before stating our algorithm we define two matrices to simplifY notation as follows: H = D[A - e], Q = I /v + HH'. (10) With these definitions the dual problem (8) becomes . 1 mm f(u):= -u'Qu - eu. (11) O~uER'" 2 It will be understood that within the ASVM Algorithm, Q- 1 will always be evaluated using the SMW formula and hence only an (n+l) x (n+l) matrix is inverted. We state our algorithm now. Note that commented (%) parts of the algorithm are not needed in general and were rarely used in our numerical results presented in Section 4. The essence of the algorithm is displayed in the two boxes below. Algorithm 3.1 Active SVM (ASVM) Algorithm for (8). (0) Start with UO := (Q- 1e)+. For i = 1,2, .. . , having u i compute Ui+1 as Ifollows. (1) Define Bi := {j I u; > a}, N i := {.i I u~ = a}. (2) Determine Ui+l .- (Q-1 e·) u i+1.- a Bi ' BiBi B ' +, Ni .. Stop if Ui+1 is the global solution, that is if a ~ Ui+1 -.l QUi+1 - e ~ a. (2a) % If f(uiH ) ~ f(ui ), then go to (4a). (2b) % If 0 :s; Ut.~l .1 QBi+1Bi+l nt.~ 1 -eBi+1 ~ 0, then UH1 is a global solution on the face of active constraints: UNi = O. Set ui := uiH and go to (4b). (3) ISet i := i + 1 and go to (1). I (4a) % Move in the direction of the global minimum on the face of act · t· t 0 S t -HI Q- l d H I .we cons razn s, UNi = . e UBi := BiBi eBi an UBi argmino95. df(uki + ).(ut.1 - nki)) I nki + ).(ut.1 - Uki ) ~ O}. If U~+1 = 0 for some j E B i , set i := i + 1 and go to (1). Otherwise UH1 is a global minimum on the face UNi = 0, and go to (4b). (4b) % Iterate a gradient projection step. Set k := 0 and uk := u i . Iterate Uk+l:= argminO<A<l f(uk - ).(uk -(Quk -e))+), k:= k + l untilf(uk) < f(11/). Set uiH ::: ilk. Set i:= i + 1 and go to (1). Remark 3.2 All commented (%) parts of the algorithm are optional and are not usually implemented unless the algorithm gets stuck, which it rarely did on our examples. Hence our algorithm is particularly simple and consists of steps (0), (1),(2) and (3). The commented parts were inserted in order to comply with the active set strategy of Morr!-Toraldo result [18] for which they give finite termination. Remark 3.3 The iteration in step (4b) is a gradient projection step which is guaranteed to converge to the global solution of (8) [2, pp 223-225] and is placed here to ensure that the strict inequality f(uk) < f(u') eventually holds as required in [18]. Similarly, the step in (4a) ensures that the function value does not increase when it remains on the same face, in compliance with [18, Algortihm BCQP(b)j. 4 Numerical Implementation and Comparisons We implemented ASVM in Visual C++ 6.0 under Windows NT 4.0. The experiments were run on the UW-Madison Data Mining Institute Locop2 machine, which utilizes a 400 MHz Pentium II Xeon Processor and a maximum of 2 Gigabytes of memory available per process. We wrote all the code ourselves except for the linear equation solver, for which we used CLAPACK [1, 6]. Our stopping criterion for ASVM is triggered when the error bound residual [16] Ilu - (u - Qu + e)+ II, which is zero at the solution of (11) , goes below O.l. The first set of experiments are designed to show that our reformulation (8) of the SVM (7) and its associated algorithm ASVM yield similar performance to the standard SVM (2), referred to here as SVM-QP. For six datasets available from the UCI Machine Learning Repository [19], we performed tenfold cross validation in order to compare test set accuracies between ASVM and SVM-QP. We implemented SVM-QP using the high-performing CPLEX barrier quadratic programming solver [10], and utilized a tuning set for both algorithms to find the optimal value of the parameter v , using the default stopping criterion of CPLEX. Altering the CPLEX default stopping criterion to match that of ASVM did not result in significant change in timing relative to ASVM, but did reduce test set correctness. In order to obtain additional timing comparison information, we also ran the wellknown SVM optimized algorithm SVM1ight [11]. Joachims, the author of SVM1ight , provided us with the newest version of the software (Version 3.lOb) and advice on setting the parameters. All features for these experiments were normalized to the range [-1, + 1] as recommended in the SVM1ight documentation. We chose to use Dataset Training Testing Time Dataset Training Testing Time m x n 1\lqorithm Correctness Correctness CPU sec) m x n ~Iqorithm Correctness Correctness (CPU sec) Liver Disorders CPLEX 70.76% 68.41% 7.87 Ionosphere CPLEX 92.81% 88.60% 9.84 ~VMf~ht 70.37% 68.12% 0.26 ~VMf~ht 92.81% 88.60% 0.23 345 x 6 ",SVM 70.40% 67.25% 0.03 351 x 34 ",SVM 93.29% 87.75% 0.26 Cleveland Heart CPLEX 87.50% 64.20% 4.17 ic Tae Toe CPLEX 65.34% 65.34% 206.52 ~VMf~ht 87.50% 64.20% 0.17 ~VMf~ht 65.34% 65.34% 0.23 297 x 13 SVM 87.24% 85.56% 0.05 958 x 9 SVM 70.27% 69.72% 0.05 Pima Diabetes CPLEX 77.36% 76.95% 128.90 Votes CPLEX 96.02% 95.85% 27.26 ~VMf~ht 77.36% 76.95% 0.19 ~VMf~ht 96.02% 95.85% 0.06 768 x 8 SVM 78.04% 78.12% 0.08 435 x 16 SVM 96.73% 96.07% 0.09 Table 1: ASVM compared with conventional SVM-QP (CPLEX and SVM1ight ) on VCI datasets. ASVM test correctness is comparable to SVM-QP, with timing much faster than CPLEX and faster than or comparable to SVM1ight • #01 Training Testing Time Points Iterations Correctness Correctness (CPU min) 4 million 5 86.09% 86.06% 38.04 7 million 5 86.10% 86.28% 95.57 Table 2: Performance of ASVM on NDC generated datasets in R 32 . (1/ = 0.01) the default termination error criterion in SVM1ight of 0.001, which is actually a less stringent criterion than the one we used for ASVM. This is because the criterion we used for ASVM (see above) is an aggregate over the errors for all points, whereas the SVM1ight criterion reflects a minimum error threshold for each point. The second set of experiments show that ASVM performs well on massive datasets. We created synthetic data of Gaussian distribution by using our own NDC Data Generator [20] as suggested by Usama Fayyad. The results of our experiments are shown in Table 2. We did try to run SVM1ight on these datasets as well, but we ran into memory difficulties. Note that for these experiments, all the data was brought into memory. As such, the running time reported consists of the time used to actually solve the problem to termination excluding I/O time. This is consistent with the measurement techniques used by other popular approaches [11, 21]. Putting all the data in memory is simpler to code and results in faster running times. However, it is not a fundamental requirement of our algorithm block matrix multiplications, incremental evaluations of Q-1 using another application of the SMW formula, and indices on the dataset can be used to create an efficient disk based version of ASVM. 5 Conclusion A very fast, finite and simple algorithm, ASVM, capable of classifying massive datasets has been proposed and implemented. ASVM requires nothing more complex than a commonly available linear equation solver for solving small systems with few variables even for massive datasets. Future work includes extensions to parallel processing of the data, handling very large datasets directly from disk as well as extending our approach to nonlinear kernels. Acknow ledgements We are indebted to our colleagues Thorsten Joachims for helping us to get SVM1ight running significantly faster on the UCI datasets, and to Glenn Fung for his efforts in running the experiments for revisions of this work. Research described in this Data Mining Institute Report 00-04, April 2000, was supported by National Science Foundation Grants CCR-9729842 and CDA-9623632, by Air Force Office of Scientific Research Grant F49620-00-1-0085 and by Microsoft. References [1] E. Anderson, Z. Bai, C. Bischof, J. Demmel, J. Dongarra, J. Du Cros, A. Greenbaum, S. Hammarling, A. McKenney, S. Ostrouchov, and D. Sorensen. LAPACK User's Guide. SIAM, Philadelphia, Pennsylvania, second edition, 1995. [2] D. P. Bertsekas. Nonlinear Programming. Athena Scientific, Belmont, MA, second edition, 1999. [3] C. J. C. Burges. A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery, 2(2):121-167, 1998. [4] C. J. C. Burges and B. Sch6lkopf. Improving the accuracy and speed of support vector machines. In M. C. Mozer, M. I. Jordan, and T. Petsche, editors, Advances in Neural Information Processing Systems -9-, pages 375-381, Cambridge, MA, 1997. MIT Press. [5] V. Cherkassky and F. Mulier. Learning from Data - Concepts, Theory and Methods. John Wiley & Sons, New York, 1998. [6] CLAPACK. f2c'ed version of LAPACK. http://www.netlib.org/clapack. [7] N. Cristianini and J. Shawe-Taylor. An Introduction to Support Vector Machines. Cambridge University Press, Cambridge, 2000. [8] M. C. Ferris and T. S. Munson. Interior point methods for massive support vector machines. Technical Report 00-05, Computer Sciences Department, University of Wisconsin, Madison, Wisconsin, May 2000. [9] G. H. Golub and C. F. Van Loan. Matrix Computations. The John Hopkins University Press, Baltimore, Maryland, 3rd edition, 1996. [10] ILOG, Incline Village, Nevada. CPLEX 6.5 Reference Manual, 1999. [11] T. Joachims. SVMlight, 1998. http://www-ai . informatik . uni -dortmund. de/FORSCHUNG/VERFAHREN/SVM_LIGHT/sVID_light.eng.html. [12] Yuh-Jye Lee and O. L. Mangasarian. SSVM: A smooth support vector machine. Computational Optimization and Applications, 2000. [13] O. L. Mangasarian. Nonlinear Programming. SIAM, Philadelphia, PA, 1994. [14] O. L. Mangasarian. Generalized support vector machines. In A. Smola, P. Bartlett, B. Sch6lkopf, and D. Schuurmans, editors, Advances in Large Margin Classifiers, pages 135- 146, Cambridge, MA, 2000. MIT Press. [15] O. L. Mangasarian and D. R. Musicant. Successive overrelaxation for support vector machines. IEEE Transactions on Neural Networks, 10:1032- 1037, 1999. [16] O. L. Mangasarian and J. Ren. New improved error bounds for the linear complementarity problem. Mathematical Programming, 66:241- 255, 1994. [17] MATLAB. User's Guide. The MathWorks, Inc., Natick, MA 01760,1992. [18] J. J. More and G. Toraldo. Algorithms for bound constrained quadratic programs. Numerische Mathematik, 55:377-400, 1989. [19] P. M. Murphy and D. W. Aha. UCI repository of machine learning databases, 1992. www.ics.uci.edu/ rvmlearn/MLRepository.html. [20] D. R. Musicant. NDC: normally distributed clustered datasets, 1998. www.cs.wisc.edu/rvmusicant/data/ndc/. [21] J. Platt. Sequential minimal optimization: A fast algorithm for training support vector machines. In Sch6lkopf et al. [22], pages 185- 208. [22] B. Sch6lkopf, C. Burges, and A. Smola (editors). Advances in Kernel Methods: Support Vector Machines. MIT Press, Cambridge, MA, 1998. [23] V. N. Vapnik. The Nature of Statistical Learning Theory. Springer, NY, 1995.
|
2000
|
96
|
1,901
|
Decomposition of Reinforcement Learning for Admission Control of Self-Similar Call Arrival Processes Jakob Carlstrom Department of Electrical Engineering, Technion, Haifa 32000, Israel jakob@ee . technion . ac . il Abstract This paper presents predictive gain scheduling, a technique for simplifying reinforcement learning problems by decomposition. Link admission control of self-similar call traffic is used to demonstrate the technique. The control problem is decomposed into on-line prediction of near-future call arrival rates, and precomputation of policies for Poisson call arrival processes. At decision time, the predictions are used to select among the policies. Simulations show that this technique results in significantly faster learning without any performance loss, compared to a reinforcement learning controller that does not decompose the problem. 1 Introduction In multi-service communications networks, such as Asynchronous Transfer Mode (ATM) networks, resource control is of crucial importance for the network operator as well as for the users. The objective is to maintain the service quality while maximizing the operator's revenue. At the call level, service quality (Grade of Service) is measured in terms of call blocking probabilities, and the key resource to be controlled is bandwidth. Network routing and call admission control (CAC) are two such resource control problems. Markov decision processes offer a framework for optimal CAC and routing [1]. By modelling the dynamics of the network with traffic and computing control policies using dynamic programming [2], resource control is optimized. A standard assumption in such models is that calls arrive according to Poisson processes. This makes the models of the dynamics relatively simple. Although the Poisson assumption is valid for most user-initiated requests in communications networks, a number of studies [3, 4, 5] indicate that many types of arrival processes in wide-area networks as well as in local area networks are statistically selfsimilar. This makes it difficult to find models of the dynamics, and the models become large and complex. If the number of system states is large, straightforward application of dynamic programming is unfeasible. Nevertheless, the "fractal" burst structure of self-similar traffic should be possible to exploit in the design of efficient resource control methods. We have previously presented a method based on temporal-difference (TD) learning for CAC of self-similar call traffic, which yields higher revenue than a TD-based controller assuming Poisson call arrival processes [7]. However, a drawback of this method is the slow convergence of the control policy. This paper presents an alternative solution to the above problem, called predictive gain scheduling. It decomposes the control problem into two parts: time-series prediction of near-future call arrival rates and precomputation of a set of control policies for Poisson call arrival processes. At decision time, a policy is selected based on these predictions. Thus, the self-similar arrival process is approximated by a quasi-stationary Poisson process. The rate predictions are made by (artificial) neural networks (NNs), trained on-line. The policies can be computed using dynamic programming or other reinforcement learning techniques [6]. This paper concentrates on the link admission control problem. However, the controllers we describe can be used as building block in optimal routing, as shown in [8] and [9]. Other recent work on reinforcement learning for CAC and routing includes [10], where Marbach et al. show how to extend the use of TD learning to network routing, and [11] where Tong et al. apply reinforcement learning to routing subject to Quality of Service constraints. 2 Self-Similar Call Arrival Processes The limitations of the traditional Poisson model for network arrival processes have been demonstrated in a number of studies, e.g. [3, 4, 5], which indicate the existence of heavytailed inter-arrival time distributions and long-term correlations in the arrival processes. Self-similar (fractal-like) models have been shown to correspond better with this traffic. A self-similar arrival process has no "natural" burst length. On the contrary, its arrival intensity varies considerably over many time scales. This makes the variance of its sample mean decay slowly with the sample size, and its auto-correlation function decay slowly with time, compared to Poisson traffic [4]. The complexity of control and prediction of Poisson traffic is reduced by the memory-less property of the Poisson process: its expected future depends on the arrival intensity, but not on the process history. On the other hand, the long-range dependence of self-similar traffic makes it possible to improve predictions of the process future by observing the history. A compact statistical measure of the degree of self-similarity of a stochastic process is the Hurst parameter [4]. For self-similar traffic this parameter takes values in the interval (0.5, 1], whereas Poisson processes have a Hurst parameter of 0.5. 3 The Link Admission Control Problem In the link admission control (LAC) problem, a link with capacity C [units/s] is offered calls from K different service classes. Calls belonging to such a class j E J = {I, ... , K} have the same bandwidth requirements hj [units/s]. The per-class call holding times are assumed to be exponentially distributed with mean 1/ftj [s]. Access to the link is controlled by a policy:rc that maps states x E X to actions a EA,:rc: X -+ A. The set X contains all feasible link states, and the action set is A = ((ai, ... ,aK) : aj E {O, Il,j E J), where aj is ° for rejecting a presumptive class-j call and 1 for accepting it. The set of link states is given by X = N x H, where N is the set of feasible call number tuples, and His the Cartesian product of some representations, '1, of the history of the per-class call arrival processes (needed because of the memory of self-similar arrival processes). N is given by N = {n : nj ;:: 0, j E J; Injhj ::; C}' jEJ where nj is the number of type-j calls accepted on the link. We assume uniform call charging, which means that the reward rate p(t) at time t is equal to the carried bandwidth: pet) = p(x(t» = I n/t)bj (1) jEl Time evolves continuously, with discrete call arrival and departure events, enumerated by k = 0,1,2, ... Denote by rk+l the immediate reward obtained from entering a state Xk at time tk until entering the next state Xk+l at time tk+1• The expectation of this reward is E,,{rk+l} = E,,{P(Xk)[tk+1 t,)} = P(Xk)1:(X",:rr(Xk» (2) where t'(xk,:rr) is the expected sojourn time in state Xk under policy:rr. By taking optimal actions, the policy controls the probabilities of state transitions so as to increase the probability of reaching states that yield high long-term rewards. The objective of link: admission control is to find a policy :rr that maximizes the average reward per stage: R(,,) ~ )~"! E.{~ ~ 'He I X, ~ x}. x E X (3) Note that the average reward does not depend on the initial state x, as the contribution from this state to the average reward tends to zero as N -+ 00 (assuming, for example, that the probability of reaching any other state y E X from every state x E X is positive). Certain states are of special interest for the optimal policy. These are the states that are candidates for intelligent blocking. The set of such states X ib C X is given by X ib = Nib X H, where Nib is the set of call number tuples for which the available bandwidth is a multiple of the bandwidth of a wideband call. In the states of X ib, the long-term reward may be increased by rejecting narrowband calls to reserve bandwidth for future, expected wideband calls. 4 Solution by Predictive Gain Scheduling Gain scheduling is a control theory technique, where the parameters of a controller are changed as a function of operating conditions [12]. The approach taken here is to look up policies in a table from predictions of the near-future per-class call arrival rates. For Poisson call arrival processes, the optimal policy for the link: admission control problem does not depend on the history, H, of the arrival processes. Due to the memory-less property, only the (constant) per-class arrival rates Aj , j E J, matter. In our gain scheduled control of self-similar call arrival processes, near-future Aj are predicted from hj- The selfsimilar call arrival processes are approximated by quasi-stationary Poisson processes, by selecting precomputed polices (for Poisson arrival processes) based on predicted A/s. One radial-basis function (REF) NN per class is trained to predict its near-future arrival rate. 4.1 Solving the Link Admission Control problem for Poisson Traffic For Poisson call arrival processes, dynamic programming offers well-established techniques for solving the LAC problem [1]. In this paper, policy iteration is used. It involves two steps: value determination and policy improvement. The value determination step makes use of the objective function (3), and the concept of relative values [1]. The difference v(x,:rr) v(y,:rr) between two relative values under a policy :rr is the expected difference in accumulated reward over an infinite time interval, starting in state X instead of state y. In this paper, the relative values are computed by solving a system of linear equations, a method chosen for its fast convergence. The dynamics of the system are characterized by state transition probabilities, given by the policy, the perclass call arrival intensities, (,q, and mean holding times, (1/,ll J The policy improvement step consists of finding the action that maximizes the relative value at each state. After improving the policy, the value determination and policy improvement steps are iterated until the policy does not change [9]. 4.2 Determining The Prediction Horizon Over what future time horizon should we predict the rates used to select policies? In this work, the prediction horizon is set to an average of estimated mean first passage times from states back to themselves, in the following referred to as the mean return time. The arrival process is approximated by a quasi-stationary Poisson process within this time interval. The motivation for this choice of prediction horizon is that the effects of a decision (action) in a state Xd influence the future probabilities of reaching other states and receiving the associated rewards, until the state Xd is reached the next time. When this happens, a new decision can be made, where the previous decision does no longer influence the future expected reward. In accordance with the assumption of quasi-stationarity, the mean return time can be estimated for call tuples n instead of the full state descriptor, x. In case of Poisson call arrival processes, the mean first passage times E,.{ Tin} from other states to a state n are the unique solution of the linear system of equations E,,{TmJ = T(m, a) + I E,,{Tln }, m E N\{n}, a = n(m) (4) IEN\!n} The limiting probability qn of occupying state n is determined for all states that are candidates for intelligent blocking, by solving a linear system of equations qB = 0. B is a matrix containing the state transition intensities, given by (Aj} and (1/,llj}. The mean return time for the link, TI, is defmed as the average of the individual mean return times of the states of Nib, weighted by their limiting probabilities and normalized: (5) For ease of implementation, this time window is expressed as a number of call arrivals. The window length Lj for class j is computed by multiplying the mean return time by the arrival rate, Lj = Aj T[, and rounding off to an integer. Although the window size varies with Aj, this variation is partly compensated by T[ decreasing with increasing Aj • 4.3 Prediction of Future Call Arrival Rates The prediction of future arrival call rates is naturally based on measures of recent arrival rates. In this work, the following representation of the history of the arrival process is used: for all classes j E J, exponentially weighted running averages hj = (hj), ... , hjM) of the inter-arrival times are computed on different time scales. These history vectors are computed using forgetting factors {a), ... ,aM } taking values in the interval (0, 1): hik) = a i[t/k) - t/k 1) 1 + (1 - a;)hik 1) , where fj(k) is the arrival time of the k-th call from class j. (6) In studies of time-series prediction, non-linear feed-forward NN s outperform linear predictors on time series with long memory [13]. We employ RBF NNs with symmetric Gaussian basis functions. The activations of the RBF units are normalized by division by the sum of activations, to produce a smooth output function. The locations and widths of the RBF units can be determined by inspection of the data sets, to cover the region of history vectors. The NN is trained with the average inter-arrival time as target. After every new call arrival, the prediction error €j(k) is computed: Lj Elk) = L I [ t(k + i) t(k + i-I)] - y/k). J i~ ' (7) Learning is performed on-line using the least mean squares rule, which means that the upd)lting must be delayed by Lj call arrivals. The predicted per-class arrival rates A/k) = y(k)-' are used to select a control policy on the arrival of a call request. Given the prediction horizon and the arrival rate predictor, ai' ... ,aM can be tuned by linear search to minimize the prediction error on sample traffic traces. 5 Numerical study The performance of the gain scheduled admission controller was evaluated on a simulated link with capacity C = 24 [units/s], that was offered calls from self-similar call arrival processes. For comparison, the simulations were repeated with three other link admission controllers: two TD-based controllers, one table-based and one NN based, and a controller using complete sharing, i.e. to accept a call if the free capacity on the link is sufficient. The NN based TD controller [7] uses RBF NNs (one per n EN), receiving (h" h2) as input. Each NN has 65 hidden units, factorized to 8 units per call class, plus a default activation unit. Its weights were initialized to favor acceptance of all feasible calls in all states. The table-based TD controller assumes Poisson call arrival processes. From this, it follows that the call number tuples n E N constitute Markovian states. Consequently, the value function table stores only one value per n. This controller was used for evaluation of the performance loss from incorrectly modelling self-similar call traffic by Poisson traffic. 5.1 Synthesis of Call Traffic Synthetic traffic traces were generated from a Gaussian fractional auto-regressive integrated moving average model, FARIMA (0, d, 0). This results in a statistically self-similar arrival process, where the Hurst parameter is easily tuned [7]. We generated traces containing arrival/departure pairs from two call classes, characterized by bandwidth requirements bi = 1 (narrow-band) and ~ = 6 (wide-band) [units/s] and call holding times with mean 1/,u1 = 1/,u2= 1 [s]. A Hurst parameter of 0.85 was used, and the call arrival rates were scaled to make the expected long-term arrival rates A, and A2 for the two classes fulfill b,A,/,u, + b).2/,u2 = 1.25 C. The ratio A,/A2 was varied from 0.4 to 2.0. 5.2 Gain Scheduling For simplicity, a constant prediction horizon was used throughout the simulations. This was computed according to section 4.2. By averaging the resulting prediction windows for A,/A2 = 0.4, 1.0 and 2.0, a window size L, = L2 = 6 was obtained. A A The table of policies to be used for gain scheduling was computed for predicted A, and A2 ranging from 0.5 to 15 with step size 0.5; in total 900 policies. The two rate-prediction NNs both had 9 hidden units. The NNs' weights were initialized to O. 5.3 Numerical results Both the TD learning controllers and the gain scheduling controller were allowed to adapt to the first 400 000 simulated call arrivals of the traffic traces. The throughput obtained by all four methods was measured on the subsequent 400000 call arrivals. o 1000 2000 3000 4000 0.5 1 1.5 2 2.5 3 3.5 4.0 call arrivals x 105 call arrivals (a) Initial weight evolution in neural predictor (b) Long-term weight evolution in neural predictor 11 9 1.5 2 2.5 3 3.5 4.0 x 105 call arrivals Throughput [units/s] 17.4 17.2 17.0 16.8 16.6 16.4 16.2 16.0 15.8 GSIRBF TDIRBF TDITBL CS 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 AdA2 (c) Weight evolution in NN based TD controller (d) Throughput versus arrival rate ratio Figure 1: Weight evolution for NN predictor (a, b); NN based TD-controller (c). Performance (d). Figure 1 (a, b) shows the evolution of the weights of the call arrival rate predictor for class 2, and figure 1 (c) displays nine weights of the RBF NN corresponding to the call number tuple (n!, n2) = (6,2), which is a candidate for intelligent blocking. These weights correspond to eight different class-2 center vectors, plus the default activation. The majority of the weights of the gain scheduling RBF NN seems to converge in a few thousand call arrivals, whereas the TD learning controller needs about tOO 000 call arrivals to converge. This is not surprising, since the RBF NNs of the TD learning controllers split up the set of training data, so that a single NN is updated much less frequently than a ratepredicting NN in the gain scheduling controller. Secondly, the TD learning NNs are trained on moving targets, due to the temporal-difference learning rule, stochastic action selection and a changing policy. A few of the weights of the gain scheduling NN change considerably even after long training. These weights correspond to RBF units that are activated by rare, large inputs. Figure t (d) evaluates performance in terms of throughput versus arrival rate ratio. Each data point is the averaged throughput for 10 traffic traces. Gain scheduling (GS/RBF) achieves the same throughput as TD learning with RBF NNs (TD/RBF), up to 1.3% compared to tabular TD learning (TDITBL), and up to 5.7% better than complete sharing (CS). The difference in throughput between TD learning and complete sharing is greatest for low arrival rate ratios, since the throughput increase by reserving bandwidth for highrate wideband calls is considerably higher than the loss of throughput from the blocked lowrate narrowband traffic. 6 Conclusion We have presented predictive gain scheduling, a technique for decomposing reinforcement learning problems. Link admission control, a sub-problem of network routing, was used to demonstrate the technique. By predicting near-future call arrival rates from one part of the full state descriptor, precomputed policies for Poisson call arrival processes (computed from the rest of the state descriptor) were selected. This increased the on-line convergence rate approximately 50 times, compared to a TD-based admission controller getting the full state descriptor as input. The decomposition did not result in any performance loss. The computational complexity of the controller using predictive gain scheduling may reach a computational bottleneck if the size of the state space is increased: the determination of optimal policies for Poisson traffic by policy iteration. This can be overcome by state aggregation [2], or by parametrization the relative value function combined with temporaldifference learning [10]. It is also possible to significantly reduce the number of relative value functions. In [14], we showed that linear interpolation of relative value functions distributed by an error-driven algorithm enables the use of less than 30 relative value functions without performance loss. Further, we have successfully employed gain scheduled link admission control as a building block of network routing [9], where the performance improvement compared to conventional methods is larger than for the link admission control problem. The use of gain scheduling to reduce the complexity of reinforcement learning problems is not limited to link admission control. In general, the technique should be applicable to problems where parts of the state descriptor can be used, directly or after preprocessing, to select among policies for instances of a simplified version of the original problem. References [1] Z. Dziong, ATM Network Resource Management, McGraw-Hill, 1997. [2] D.P. Bertsekas, Dynamic Programming and Optimal Control, Athena Scientific, Belmont, Mass., 1995. [3] V. Paxson and S. Floyd, "Wide-Area Traffic: The Failure of Poisson Modeling", IEEF/ACM Transactions on Networking, vol. 3, pp. 226-244, 1995. [4] W.E. Leland, M.S. Taqqu, W. Willinger and D.V. Wilson, "On the Self-Similar Nature of Ethemet Traffic (Extended Version)", IEEF/ACM Transactions on Networking, vol. 2, no. 1, pp. 1- 15, Feb. 1994. [5] A Feldman, AC. Gilbert, W. Willinger and T.G. Kurtz, "The Changing Nature of Network Traffic: Scaling Phenomena", Computer Communication Review, vol. 28, no. 2, pp. 5- 29, April 1998. [6] R.S. Sutton and AG. Barto, Reinforcement Learning: An Introduction, MIT Press, Cambridge, Mass., 1998. [7] J. Carlstrom and E. Nordstrom, "Reinforcement Learning for Control of Self-Similar Call Traffic in Broadband Networks", Teletraffic Engineering in a Competitive World - Proceedings of The 16th International Teletraffic Congress (ITC 16), pp. 571- 580, Elsevier Science B.V., 1999. [8] Z. Dziong and L. Mason,"Call Admission Control and Routing in Multi-service Loss Networks", IEEE Transactions on Communications, vol. 42, no. 2. pp. 2011- 2022, Feb. 1994. [9] J. Carlstrom and E. Nordstrom, "Gain Scheduled Routing in Multi-Service Networks", Technical Report 2000-009, Dept. of Information Technology, Uppsala University, Uppsala, Sweden, April 2000. [10] P. Marbach, O. Mihatsch and J.N. Tsitsiklis, "Call Admission Control and Routing in Integrated Service Networks Using Neuro-Dynarnic Programming", IEEE J. Sel. Areas ofComm, Feb. 2000. [11] H. Tong and T. Brown, "Adaptive Call Admission Control Under Quality of Service Constraints: A Reinforcement Learning Solution", IEEE Journal on Selected Areas in Communications, Feb. 2000. [12] K.J. Astrom and B. Wittenmark, Adaptive Control, 2nd ed., Addison-Wesley, 1995. [13] S. Haykin, Neural Networks: A Comprehensive Foundation, 2nd ed., Macmillan College Publishing Co., Englewood Cliffs, NJ, 1999. [14] J. Carlstrom, "Efficient Approximation of Values in Gain Scheduled Routing", Technical Report 2000-010, Dept. of Information Technology, Uppsala University, Uppsala, Sweden, April 2000.
|
2000
|
97
|
1,902
|
Generalizable Singular Value Decomposition for Ill-posed Datasets Ulrik Kjerns Lars K. Hansen Department of Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby, Denmark uk, lkhansen@imm. dtu. dk Abstract Stephen C. Strother PET Imaging Service VA medical center Minneapolis steve@pet. med. va. gov We demonstrate that statistical analysis of ill-posed data sets is subject to a bias, which can be observed when projecting independent test set examples onto a basis defined by the training examples. Because the training examples in an ill-posed data set do not fully span the signal space the observed training set variances in each basis vector will be too high compared to the average variance of the test set projections onto the same basis vectors. On basis of this understanding we introduce the Generalizable Singular Value Decomposition (GenSVD) as a means to reduce this bias by re-estimation of the singular values obtained in a conventional Singular Value Decomposition, allowing for a generalization performance increase of a subsequent statistical model. We demonstrate that the algorithm succesfully corrects bias in a data set from a functional PET activation study of the human brain. 1 Ill-posed Data Sets An ill-posed data set has more dimensions in each example than there are examples. Such data sets occur in many fields of research typically in connection with image measurements. The associated statistical problem is that of extracting structure from the observed high-dimensional vectors in the presence of noise. The statistical analysis can be done either supervised (Le. modelling with target values: classification, regresssion) or unsupervised (modelling with no target values: clustering, PCA, ICA). In both types of analysis the ill-posedness may lead to immediate problems if one tries to apply conventional statistical methods of analysis, for example the empirical covariance matrix is prohibitively large and will be rank-deficient. A common approach is to use Singular Value Decomposition (SVD) or the analogue Principal Component Analysis (PCA) to reduce the dimensionality of the data. Let the N observed i-dimensional samples Xj, j = L .N, collected in the data matrix X = [Xl ... XN] of size I x N, I> N . The SVD-theorem states that such a matrix can be decomposed as (1) where U is a matrix of the same size as X with orthogonal basis vectors spanning the space of X, so that UTU = INxN. The square matrix A contains the singular values in the diagonal, A = diag( AI, ... , >w), which are ordered and positive Al ~ A2 ~ ... ~ AN ~ 0, and V is N x N and orthogonal V TV = IN. If there is a mean value significantly different from zero it may at times be advantageous to perform the above analysis on mean-subtracted data, i.e. X - X = U A V T where columns of X all contain the mean vector x = Lj xj/N. Each observation Xj can be expressed in coordinates in the basis defined by the vectors of U with no loss of information[Lautrup et al., 1995]. A change of basis is obtained by qj = U T Xj as the orthogonal basis rotation Q = [ql ... qN] = U T X = UTUAVT = AVT. (2) Since Q is only N x Nand N « I, Q is a compact representation of the data. Having now N examples of N dimension we have reduced the problem to a marginally illposed one. To further reduce the dimensionality, it is common to retain only a subset of the coordinates, e.g. the top P coordinates (P < N) and the supervised or unsupervised model can be formed in this smaller but now well-posed space. So far we have considered the procedure for modelling from a training set. Our hope is that the statistical description generalizes well to new examples proving that is is a good description of the generating process. The model should, in other words, be able to perform well on a new example, x*, and in the above framework this would mean the predictions based on q* = U T x* should generalize well. We will show in the following, that in general, the distribution of the test set projection q* is quite different from the statistics of the projections of the training examples qj. It has been noted in previous work [Hansen and Larsen, 1996, Roweis, 1998, Hansen et al., 1999] that PCA/SVD of ill-posed data does not by itself represent a probabilistic model where we can assign a likelihood to a new test data point, and procedures have been proposed which make this possible. In [Bishop, 1999] PCA has been considered in a Bayesian framework, but does not address the significant bias of the variance in training set projections in ill-posed data sets. In [Jackson, 1991] an asymptotic expression is given for the bias of eigen-values in a sample covariance matrix, but this expression is valid only in the well-posed case and is not applicable for ill-posed data. 1.1 Example Let the signal source be I-dimensional multivariate Gaussian distribution N(O,~) with a covariance matrix where the first K eigen-values equal u 2 and the last 1- K are zero, so that the covariance matrix has the decomposition ~=u2YDyT, D=diag(1, ... ,1,0, ... ,0), yTY=I (3) Our N samples of the distribution are collected in the matrix X = [Xij] with the SVD A = diag(Al, ... , AN) (4) and the representation ofthe N examples in the N basis vector coordinates defined by U is Q = [%] = U T X = A V T. The total variance per training example is ~ LX;j ~Tr(XTX) = ~Tr(VAUTUAVT) = ~Tr(VA2VT) i,j = ~ Tr(VVT A2) = ~ Tr(A2) = ~L A; i (5) Note that this variance is the same in the U-basis coordinates: 1 '" 2 N L...J qij = ~Tr(QTQ) = ~Tr(VA2VT) = ~ LA~ i,j i We can derive the expected value of this variance: (~ LX;) = (LxL) = (x? Xl) = Tr:E = a2K i ,j (6) (7) Now, consider a test example X* '" N(O,:E) with the projection q* = U T x* which will have the average total variance (Tr[(UT x*{ (UT x*)]) = Tr [(x*x* T)UUT] Tr[:EUUT] = Tr[DUUT] = a2min(N,K) (8) In summary, this means that the orthogonal basis U computed from the training set spans all the variance in the training set but fails to do so on the test examples when N < K, i.e. for ill-posed data. The training set variance is K / N a 2 on average per coordinate, compared to a 2 for the test examples. So which of the two variances is "correct" ? From a modelling point of view, the variance from the test example tells us the true story, so the training set variance should be regarded as biased. This suggests that the training set singular values should be corrected for this bias, in the above example by re-estimating the training set projections using Q = J N / K Q. In the more general case we do not know K, and the true covariance may have an arbitrary eigen-spectrum. The GenSVD algorithm below is a more general algorithm for correcting for the training set bias. 2 The GenSVD Algorithm The data matrix consists of N statistically independent samples X = [Xl ... XN ] so X is size I x N, and each column of X is assumed multivariate Gaussian, Xj '" N(O,:E) and is ill-posed with rank:E > N. With the SVD X = UoAoVaT, we now make the approximation that Uo contains an actual subset of the true eigen-vectors of :E (9) where we have collected the remaining (unspanned by X) eigen-vectors and values in UJ. and Ai , satisfying uluJ. = I and UJUJ. = 0. The unknown 'true' eigen-values corresponding to the observed eigen-vectors are collected in A = diag(Al, ... AN), which are the values we try to estimate in the following. It should be noted that a direct estimation of :E using f: = j;y X X T yields f: = j;yuoAoVaTVoAoUJ = j;yUoA~UJ, i.e., the nonzero eigen-vectors and values of f: is Uo and Ao. The distribution of test samples x* inside the space spanned by Uo is (10) The problem is that Uo and the examples Xj are not independent, so UJ Xj is biased, e.g. the SVD estimate -k A ~ of A 2 assigns all variance to lie within Uo. The GenSVD algorithm bypasses this problem by, for each example, computing a basis on all other examples, estimating the variances in A 2 in a leave-one-out manner. Consider (11) where we introduce the notation X_j for the matrix of all examples except the j'th, and this matrix is decomposed as X_j = B_jA_j lC;' The operation B_jB_; Xj projects the example onto the basis defined by the remaining examples, and back again, so it 'strips' off the part of signal space which is special for Xj which could be signal which does not generalize across examples. Since B_j and Xj are independent B-"J Xj has the same distribution as the projection of a test example x*, B_; x*. Thus, B_jB_; Xj and B_jB_; x* have the same distribution as well. Now, since span B_j=span X_j and span Uo=span [X_j Xj] we have that span B_j~span Uo so we see that Z j and U J B_jB-"J X* are identically distributed. This means that Zj has the covariance UJ B_jB-"J~B_jB_;Uo and using Eq. (9) and that ul B_j = 0 (since uluo = 0) we get (12) We note that this distribution is degenerate because the covariance is of rank N -l. For a sample Zj from the above distribution we have that UJ B_jB_;Uozj = UJ B_jB_;UoUJ B_jB_; Xj = UJ B_jB_; Xj = Zj (13) As a second approximation, assume that the observed Zj are independent so that we can write the likelihood of A ~ log [(27r)N/21(uJ B_J(B-"JUo)A2(UJ B_J(R;Uo)l l/2] +~ ~zJ (UJ B_J(B_;Uo)A -2(UJ B_j) (B-"JUo)zj j N ~ 2 1~ T 2 C + log A· + z· A - z· 2 J 2 J J j j (14) where we have used Eq. (13) and that the determinantl is approximated by IA21. This above expression is maximized when 5.~ = ~ ~ Z~j' (15) j A T A A A The GenSVD of X is then X = UoAV ,A = diag(Al' ... , AN). In practice, using Eq. (11) directly to compute an SVD of the matrix X_j for each example is computationally demanding. It is possible to compute Zj in a more efficient two-level procedure with the following algorithm: Compute UOAoVOT = svd(X) and Qo = [qj] = AoVOT lSince Zj is degenerate, we define the likelihood over the space where Zj occur, i.e. the determinant in Eq. 14 should be read as 'the product of non-zero eigenvalues'. foreach j = L.N Compute B_;A_; v..; = svd( Q.J Zj = B_;B-"J qj A2 1 2 '\ = Iii L:j Zij If the data has a mean value that we wish to remove prior to the SVD it is important that this is done within the GenSVD algorithm. Consider a centered matrix Xc = X - X where X contains the mean x replicated in all N columns. The signal space in Xc is now corrupted because each centered example will contain a component of all examples, which means the 'stripping' of signal components not spanned by other examples no longer works: B_; Xj is no longer distributed like B_; x*. This suggests the alternative algorithm for data with removal of mean component: Compute UOAoVOT = svd(X) and Qo = [qj] = AoVOT foreach j = L.N 1 '"" q-j = N-1 6j'¥j qj' T Compute B_;A_; v..; = svd(Q_; - Q.;) Zj = B_;B-"J (qj ii-j) A2 _ 1 2 Ai N -1 L:j Zij Finally, note that it is possible to leave out more than one example at a time if the data is independent only in block, i.e. let Q.k would be Qo with the k'th block left out. Example With PET Scans We compared the performance of GenSVD to conventional SVD on a functional [150] water PET activation study of the human brain. The study consisted of 18 subjects, who were scanned four times while tracing a star-shaped maze with a joy-stick with visual feedback, in total 72 scans of dimension '" 25000 spatial voxels. After the second scan, the visual feedback was mirrored, and the subject accomodated to and learned the new control environment during the last two scans. Scans were normalized by 1) dividing each scan by the average voxel value measured inside a brain mask and 2) for each scan subtracting the average scan for that subject thereby removing subject effects and 3) intra and inter-subject normalization and transformation using rigid body reorientation and affine linear transformations respectively. Voxels inside aforementioned brain mask were arranged in the data matrix with one scan per column. Figure 1 shows the results of an SVD decomposition compared to GenSVD. Each marker represents one scan and the glyphs indicate scan number out of the four (circle-square-star-triangle). The ellipses indicate the mean and covariances of the projections in each scan number. The 32 scans from eight subjects were used as a training set and 40 scans from the remaining 10 subjects for testing. The training set projections are filled markers, test-set projections onto the basis defined by the training set are open markers (i.e. we plot the first two columns of UoAo for SVD and of UoA for GenSVD). We see that there is a clear difference in variance in the train- and test-examples, which is corrected quite well by GenSVD. The lower plot in Figure 1 shows the singular values for the PET data set. We see that GenSVD estimates are much closer to the actual test projection standard deviations than the SVD singular values. 3 Conclusion We have demonstrated that projection of ill-posed data sets onto a basis defined by the same examples introduces a significant bias on the observed variance when comparing to projections of test examples onto the same basis. The GenSVD algorithm has been presented as a tool for correcting for this bias using a leave-one-out re-estimation scheme, and a computationally efficient implementation has been proposed. We have demonstrated that the method works well on an ill-posed real-world data set, were the distribution of the GenSVD-corrected training test set projections matched the distribution of the observed test set projections far better than the uncorrected training examples. This allows a generalization performance increase of a subsequent statistical model, in the case of both supervised and unsupervised models. Acknowledgments This work was supported partly by the Human Brain Project grant P20 MH57180, the Danish Research councils for the Natural and Technical Sciences through the Danish Computational Neural Network Center (CONNECT) and the Technology Center Through Highly Oriented Research (THOR). References [Bishop, 1999] Bishop, C. (1999). Bayesian pca. In Kearns, M. S., Soli a, S. A., and Cohn, D. A., editors, Advances in Neural Information Processing Systems, volume 11. The MIT Press. [Hansen et al., 1999] Hansen, L., Larsen, J. , Nielsen, F., Strother, S., Rostrup, E., Savoy, R., Lange, N., Sidtis, J., Svarer, C., and Paulson, O. (1999) . Generalizable patterns in neuroimaging: How many principal components? NeuroImage, 9:534- 544. [Hansen and Larsen, 1996] Hansen, L. K. and Larsen, J. (1996). Unsupervised learning and generalization. In Proceedings of IEEE International Conference on Neural Networks, pages 25- 30. [Jackson, 1991] Jackson, J . E. (1991). A User's Guide to Principal Components. Wiley Series on Probability and Statistics, John Wiley and Sons. [Lautrup et aI., 1995] Lautrup, B., Hansen, L. K., Law, I., M0rch, N., Svarer, C., and Strother, S. (1995). Massive weight sharing: A cure for extremely ill-posed problems. In Hermann, H. J ., Wolf, D. E., and Poppel, E. P., editors, Proceedings of Workshop on Supercomputing in Brain Research: Prom Tomography to Neural Networks: Prom tomography to neural networks, HLRZ, KFA Jillich, Germany, pages 137- 148. World Scientific. [Roweis, 1998] Roweis, S. (1998) . Em algorithms for pca and spca. In Jordan, M. I., Kearns, M. J., and Soli a, S. A., editors, Advances in Neural Information Processing Systems, volume 10. The MIT Press. 3.00 2.00 'E 11 1.00 0 "E 8 0 0.00 > (jJ -g 0 i;l - 1.00 (jJ - 2.00 - 3.00 - 4.00 1.50 1.00 " 11 ~ 0.50 8 o ~ 0.00 <D (!) "C 8 - 0.50 ell - 1.00 * * * .. - 3.00 - 2.00 * 1< Conventional SVD • * 1< • ~ . Oc .· . ~ " ~: j~.~ .Ii. .Ii. ..: .J>. .Ii. • - 1.00 0.00 1.00 First SVD component Generalizable SVD • '!' • .. • • 2.00 3.00 Solid: Train Open:Test o Trace scan 1 o Trace scan 2 * Mirror scan 1 £J. Mirror scan 2 4.00 - 1.50'--_~ __ ~ _______________ ---J c: o iii .~ - 2.00 - 1.50 - 1.00 - 0.50 0.00 0.50 1.00 1.50 2.00 First GenSVD component 2.00 r-~--------------------'---' SVD training set projection stdev -- GenSVD training set proj. stdev Test set projection stdev 1.50 \ \ ~ 1.00 {g ~ 0.50 0.00 '-----------------------'---' 5 10 Component 15 20 Figure 1: Projections of PET data in SVD and GenSVD. Each subject's four scans are indicated by: circle, square, star, triangle. Training set scans are marked with filled glyphs and test set with open glyphs. Solid and dotted Ellipses indicate test/train covariance per scan number. The third plot shows the standard deviations for the training and test set for SVD and GenSVD projections.
|
2000
|
98
|
1,903
|
Vicinal Risk Minimization Olivier Chapelle, Jason Weston* , Leon Bottou and Vladimir Vapnik AT&T Research Labs, 100 Schultz drive, Red Bank, NJ, USA * Barnhill BioInformatics.com, Savannah, GA, USA. {chapelle, weston,leonb, vlad}@research.att.com Abstract The Vicinal Risk Minimization principle establishes a bridge between generative models and methods derived from the Structural Risk Minimization Principle such as Support Vector Machines or Statistical Regularization. We explain how VRM provides a framework which integrates a number of existing algorithms, such as Parzen windows, Support Vector Machines, Ridge Regression, Constrained Logistic Classifiers and Tangent-Prop. We then show how the approach implies new algorithms for solving problems usually associated with generative models. New algorithms are described for dealing with pattern recognition problems with very different pattern distributions and dealing with unlabeled data. Preliminary empirical results are presented. 1 Introduction Structural Risk Minimisation (SRM) in a learning system can be achieved using constraints on the parameter vectors, using regularization terms in the cost function, or using Support Vector Machines (SVM). All these principles have lead to well established learning algorithms. It is often said, however, that some problems are best addressed by generative models. The first problem is of missing data. We may for instance have a few labeled patterns and a large number of unlabeled patterns. Intuition suggests that these unlabeled patterns carry useful information. The second problem is of discriminating classes with very different pattern distributions. This situation arises naturally in anomaly detection systems. This also occurs often in recognition systems that reject invalid patterns by defining a garbage class for grouping all ambiguous or unrecognizable cases. Although there are successful non-generative approaches (Schuurmans and Southey, 2000) (Drucker, Wu and Vapnik, 1999), the generative framework is undeniably appealing. Recent results (Jaakkola, Meila and Jebara, 2000) even define generative models that contain SVM as special cases. This paper discusses the Vicinal Risk Minimization (VRM) principle, summarily introduced in (Vapnik, 1999). This principle was independently hinted at by Tong and Koller (Tong and Koller, 2000) with a useful generative interpretation. In particular, they proved that SVM are a limiting case of their Restricted Bayesian Classifiers. We extend Tong's and Koller's result by showing that VRM subsumes several well known techniques such as Ridge Regression (Hoerl and Kennard, 1970), Constrained Logistic Classifier, or Tangent Prop (Simard et aI., 1992). We then go on to show how VRM naturally leads to simple algorithms that can deal with problems for which one would have formally considered purely generative models. We provide algorithms and preliminary empirical results for dealing with unlabeled data or recognizing classes with very different pattern distributions. 2 Vicinal Risk Minimization The learning problem can be formulated as the search of the function f E F that minimizes the expectation of a given loss £(f(x), y). R(f) = f £(f(x), y) dP(x, y) (1) In the classification framework, y takes values ±1 and £(f(x) , y) is a step function such as 1 - Sign(yf(x)), whereas in the regression framework, y is a real number and commonly £(f(x), y) is the mean squared error (f(x) _ y)2. The expectation (1) cannot be computed since the distribution P(x, y) is unknown. However, given a training set {(Xi, Yi) h <i<n, it is common to minimize instead the empirical ri~: -1 n Remp(f) = - L £(f(Xi)' Yi) n i=l Empirical Risk Minimization (ERM) is therefore equivalent to minimizing the expectation of the loss function with respect to an empirical distribution Pemp(x,y) formed by assembling delta functions located on each example: 1 n dPemp(x, y) = - LOx, (X)Oy, (y) n i=l (2) It is quite natural to consider improved density estimates by replacing the delta functions ox, (x) by some estimate of the density in the vicinity of the point Xi, PXi (X). 1 n dPest(x, y) = - L dPx, (x)Oy'(y) n i=l (3) We can define in this way the vicinal risk of a function as: Rvic(f) = f £(f(x),y) dPest(x,y) = ~ t f £(f(x), Yi)dPXi (x) (4) ~=1 The Vicinal Risk Minimization principle consists of estimating argmin!EFR(f) by the function which minimizes the vicinal risk (4). In general, one can construct the VRM functional using any estimate dPest (x, y) of the density dP(x, y), instead of restricting our choices to pointwise kernel estimates. Spherical gaussian kernel functions Nu(x - Xi) are otherwise an obvious choice for the local density estimate dPXi (x). The corresponding density estimate dPest is a Parzen windows estimate. The parameter u controls the scale of the density estimate. The extreme case u = 0 leads to the estimation of the density by delta functions and therefore leads to ERM. This must be distinguished from the case u -t 0 because the limit is taken after the minimization of the integral, leading to different results as shown in the next section. The theoretical analysis of ERM (Vapnik, 1999) shows that the crucial factor is the capacity of the class F offunctions. Large classes entail the risk of overfitting, whereas small classes entail the risk of underfitting. Two factors however are responsible for generalization of VRM, namely the quality of the estimate dPest and the size of the class F of functions. If dPest is a poor approximation to P then VRM can still perform well if F has suitably small capacity. ERM indeed uses a very naive estimate of dP and yet can provide good results. On the other hand, if F is not chosen with suitably small capacity then VRM can still perform well if the estimate dPest is a good approximation to dP. One can even take the set of all possible functions (whose capacity is obviously infinite) and still find a good solution if the estimate dPest is close enough to dP with an adequate metric. For example, if dPest is a Parzen window density estimate, then the Vicinal Risk minimizer is the Parzen window classifier. This latter property contrasts nicely with the ERM principle whose results strongly depend on the choice of the class of functions. Although we do not have a full theoretical understanding of VRM at this time, we expect considerable differences in the theoretical analysis of ERM and VRM. 3 Special Cases We now discuss the relationship of VRM to existing methods. There are obvious links between VRM and Parzen windows or Nearest Neighbour when the set of functions F is unconstrained. Furthermore, many existing algorithms can be viewed as special cases of VRM for different choices of F and dPest . a) VRM Regression and Ridge Regression Consider the case of VRM for regression with spherical Parzen windows (using gaussian kernel) with standard deviation u and with a family F of linear functions fw ,b(X) = W . x + b. We can write the vicinal risk as: Rvic(f) The resulting expression is the empirical risk augmented by a regularization term. The particular cost function above is known as the Ridge Regression cost function (Hoed and Kennard, 1970). This result can be extended to the case of non linear functions f by performing a Taylor expansion of f(Xi + £). The corresponding regularization term then combines successive derivatives offunction f. Useful mathematical arguments can be found in (Leen, 1995). b) VRM and Invariant Learning Generating synthetic examples is a simple way to incorporate selected invariances in a learning system. For instance, we can augment a optical character recognition database by applying applying translations or rotations to the initial examples. In the limit, this is equivalent to replacing each initial example by a distribution whose shape represents the desired invariances. This formulation naturally leads to a special case of VRM in which the local density estimates PXi (x) are elongated in the direction of invariance. Tangent-Prop (Simard et aI., 1992) is a more sophisticated way to incorporate invariances by adding an adequate regularization term to the cost function. Tangent-Prop has been formally proved to be equivalent to generating synthetic examples with infinitesimal deformations (Leen, 1995). This analysis makes Tangent-Prop a special case ofVRM. The local density estimate PXi is simply formed by Gaussian kernels with a covariance matrix whose eigenvectors describe the tangent direction to the invariant manifold. The eigenvalues then represent the respective strengths of the selected invariances. The tangent covariance matrix used in the SVM context by (Scholkopf et aI., 1998) specifies invariances globally. It can also been seen as a special case of VRM. c) VRM Classifier and Constrained Logistic Classifier Consider the case of VRM for classification with spherical Parzen windows with standard deviation 0' and with a family F of linear functions fw,b(X) = W . x + b. We can assume without loss of generality that JJwJJ = 1. We can write the vicinal risk as: RVic(w,b) 1 n f :;;: L -Yi Sign(b + w . x) dPXi (x) i=l = 1 n f :;;: L -Yi Sign(b + W· Xi + W· e:) dNu(e:) ,=1 We can decompose e: = WEw + e:~ where WEw represents its component parallel to wand e:~ represents its orthogonal component. Since JJwJJ = 1, we have W • e: = Ew. After integrating over e: ~ we are left with the following expression: The latter integral can be seen as the convolution of the Gaussian Nu (x) with the step function Sign(x), which is a sigmoid shaped function with asymptotes at ±1. Using notation rp(x) = 2 erf(x) - 1, we can write: 1 n (w. Xi + b) RVic(W, b) = :;;: ?= -Yi rp 0' ,=1 By rescaling wand b by a factor 1/0', we can write the following equivalent formulation of the VRM: Arg Min - L Yi rp(w· Xi + b) { In :i': constra:til~wJJ = 1/0' (5) Except for the minor shape difference between sigmoid functions, the above formulation describes a Logistic Classifier with a constraint on the weights. This formulation is also very close to using a single artificial neuron with a sigmoid transfer function and a weight decay. The above proof illustrates a general identity. Transforming the empirical probability estimate (2) by convolving it with a kernel function is equivalent to transforming the loss function £(f (x), y) by convolving it with the same kernel function. This is summarized in the following equality, where * represents the convolution operator. f £(f(x),y) [NuO * dPemp(',y)] (x) = f [£(f(.),y) *NuO] (x) dPemp(x,y) d) VRM Classifier and SVM (Tong and Koller, 2000) Consider again the case of VRM for classification with spherical Parzen windows with standard deviation 0' and with a family F of linear functions fw,b(X) = W . x + b. The resulting algorithm is in fact a Restricted Bayesian Classifier (Tong and Koller, 2000). Assuming that the examples are separable, Tong and Koller have shown that the resulting decision boundary tends towards the hard margin SVM decision boundary when a tends towards zero. The proof is based on the following observation: when a ~ 0, the vicinal risk (4) is dominated by the terms corresponding to the examples whose distance to the decision boundary is minimal. These examples in fact are the support vectors. On the other hand, choosing a > ° generates a decision boundary which depends on all the examples. The contribution of each example decreases exponentially when its distance to the decision boundary increases. This is only slightly different from a soft margin SVM whose boundary relies on support vectors that can be more distant than those selected by hard margin SVM. The difference here is just in the cost functions (sigmoid compared to linear loss). e) SVM and Constrained Logistic Classifiers The two previous paragraphs show that the same particular case of VRM is (a) equivalent to a Logistic Classifier with a constraint on the weights, and (b) tends towards the SVM classifier when a ~ ° and when the examples are separable. As a consequence, we can state that the Logistic Classifier decision boundary tends towards the SVM decision boundary when we relax the constraint on the weights. In practice we can find the SVM solution with a Logistic Classifier by simply using an iterative weight update algorithm such as gradient descent, choosing small initial weights, and letting the norm of the weights grow slowly while the iterative algorithm is running. Although this algorithm is not exact, it is fast and efficient. This is in fact similar to what is usually done with back-propagation neural networks (LeCun et aI., 1998). The same algorithm can be used for the VRM. In that context early stopping is similar to choosing the optimal a using cross-validation. 4 New Algorithms and Results 4.1 Adaptive Kernel Widths It is known in density estimation theory that the quality of the density estimate can be improved using variable kernel widths (Breiman, Meisel and Purcell, 1977). In regions of the space where there is little data, it is safer to have a smooth estimate of the density, whereas in the regions of the space there is more data one wants to be as accurate as possible via sharper kernel estimates. The VRM principle can take advantage of these improved density estimates for other problem domains. We consider here the following density estimate: 1 dPest(x, y) = - L 8Yi (y) NUi (x - Xi) dx n . ~ where the specific kernel width ai for each training example Xi is computed from the training set. a) Wisconsin Breast Cancer We made a first test of the method on the Wisconsin breast cancer dataset l which contains 589 examples on 30 dimensions. We compared VRM using the set of linear classifiers with various underlying density estimates. The minimization was achieved using gradient descent on the vicinal risk. All hyperparameters were determined using cross-validation. The following table reports results averaged on 100 runs. 1 h up:1 /horn. first. gmd .de/ ..... raetschl data/breast -cancer. SoftSVM VRM VRM Training Set HardSVM Beste Best fixed U Adaptive Ui 10 11.3% 11.1% 10.8% 9.6% 20 8.3% 7.5% 6.9% 6.6% 40 6.3% 5.5% 5.2% 4.8% 80 5.4% 4.0% 3.9% 3.7% The adaptive kernel width Ui were computed by multiplying a global factor by the average distance of the five closest training examples. The best global factor is determined by crossvalidation. These results suggest that VRM with adaptive kernel widths can outperform state of the art classifiers on small training sets. b) MNIST "I" versus other digits A second test was performed using the MNIST handwritten digits2• We considered the sub-problem of recognizing the ones versus all other digits. The testing set contains 10000 digits (5000 ones and 5000 non-ones). Two training set sizes were considered with 250 or 500 ones and an equal number of non-ones. Computations were achieved using the algorithm suggested in section (3.e). We simply trained a single linear unit with a sigmoid transfer function using stochastic gradient updates. This is appropriate for implementing an approximate VRM with a single kernel width. Adaptive kernel widths are implemented by simply changing the slope of the sigmoid for each example. For each example Xi, the kernel width Ui is computed from the training set using the 5/1000th quantile of the distances of all other examples to example Xi. The sigmoid slopes are then computed by renormalizing the Ui in order to make their mean equal to 1. Early stopping was achieved with cross-validation. Training Set HardSVM VRM VRM Fixed slope Adaptive slope 250+250 3.34% 2.79% 2.54% 500+500 3.11% 2.47% 2.27% 1000+1000 2.94% 2.08% 1.96% The statistical signifiance of these results can be asserted with very high probability by comparing the list of errors performed by each system (Bottou and Vapnik, 1992). Again these results suggest that VRM with adaptive kernel widths can be very useful with small training sets. 4.2 Unlabeled Data In some applications unlabeled data is in abundance whereas labeled data is not. The use of unlabeled data falls into the framework of VRM by simply making the same vicinal loss for unlabeled points. Given m unlabeled points xi, ... , x:n, one obtains the following formulation: 1 n f 1 m f Rvic(f) =;;: L l(f(X),Yi)dPXi(x) + m L l(f(x),f(xn)dPx;(x). i=l i=l To give an example of the usefulness of our approach consider the following example. Two normal distributions on the real line N( -1.6,1) and N(1.6, 1) model the patterns of two classes with equal probability; 20 labeled points and 100 unlabeled points are drawn. The following table compares the true generalization error of VRM with gaussian kernels and linear functions. Results are averaged over 100 runs. Two different kernel widths UL and Uu were used for kernels associated with labeled or unlabeled examples. Best kernel widths were obtained by cross-validation. We also studied the case UL -+ 0 in order to provide a result equivalent to a plain SVM. 2http://www.research.att.com/ ... yannlocr/index.html aL -+ 0 Best aL Best au Best au Labeled 6.5% 5.0% Labeled+Unlabeled 5.6% 4.3% Note that when both aL and au tend to zero, this algorithm reverts to a transduction algorithm due to Vapnik which was previously solved by the more difficult optimization procedure of integer programming (Bennet and Demiriz, 1999). 5 Conclusion In conclusion, the Vicinal Risk Minimization VRM principle provides a useful bridge between generative models and SRM methods such as SVM or Statistic Regularization. Several well known algorithms are in fact special cases of VRM. The VRM principle also suggests new algorithms. In this paper we proposed algorithms for dealing with unlabeled data and recognizing classes with very different pattern distributions, obtaining initial promising results. We hope that this approach can lead to further understanding of existing methods and also to suggest new ones. References Bennet, K. and Demiriz, A. (1999). Semi-supervised support vector machines. In Advances in Neural Information Processing Systems 11, pages 368-374. MIT Press. Bottou, L. and Vapnik, V. N. (1992). Local learning algorithms, appendix on confidence intervals. Neural Computation, 4(6):888- 900. Breiman, L., Meisel, W., and Purcell, E. (1977). Variable kernel estimates of multivariate densities. Technometrics, 19:135- 144. Drucker, H., Wu, D., and Vapnik, V. (1999). Support vector machines for spam categorization. Neural Networks, 10:1048- 1054. Hoed, A. and Kennard, R. W. (1970). Ridge regression: Biased estimation for non orthogonal problems. Technometrics, 12(1):55--67. Jaakkola, T., Meila, M., and Jebara, T. (2000). Maximum entropy discrimination. In Advances in Neural Information Processing Systems 12. MIT Press. LeCun, Y., Bottou, L., Orr, G., and Muller, K. (1998). Efficient backprop. In Orr, G. and K., M., editors, Neural Networks: Tricks of the Trade. Springer. Leen, T. K. (1995). Invariance and regularization in learning. In Advances in Neural Infonnation Processing Systems 7. MIT Press. Scholkopf, B., Simard, P., Smola, A., Vapnik, V. (1998). Prior knowledge in support vector kernels. In Advances in Neural Information Processing Systems 10. MIT Press. Schuurmans, D. and Southey, F. (2000). An adaptive regularization criterion for supervised learning. In Proceedings of the Seventeenth International Conference on Machine Learning (ICML2000). Simard, P., Victorri, B., Le Cun, Y., and Denker, J. (1992). Tangent prop: a formalism for specifying selected invariances in adaptive networks. In Advances in Neural Information Processing Systems 4, Denver, CO. Morgan Kaufman. Tong, S. and Koller, D. (2000). Restricted bayes optimal classifiers. Proceedings of the 17th National Conference on Artificial Intelligence (AAAI). Vapnik, V. (1999). The Nature of Statistical Learning Theory (Second Edition). Springer Verlag, New York.
|
2000
|
99
|
1,904
|
Means. Correlations and Bounds M.A.R. Leisink and H.J. Kappen Department of Biophysics University of Nijmegen, Geert Grooteplein 21 NL 6525 EZ Nijmegen, The Netherlands {martijn,bert}@mbfys.kun.nl Abstract The partition function for a Boltzmann machine can be bounded from above and below. We can use this to bound the means and the correlations. For networks with small weights, the values of these statistics can be restricted to non-trivial regions (i.e. a subset of [-1 , 1]). Experimental results show that reasonable bounding occurs for weight sizes where mean field expansions generally give good results. 1 Introduction Over the last decade, bounding techniques have become a popular tool to deal with graphical models that are too complex for exact computation. A nice property of bounds is that they give at least some information you can rely on. For instance, one may find that a correlation is definitely between 0.4 and 0.6. An ordinary approximation might be more accurate, but in practical situations there is absolutely no warranty for that. The best known bound is probably the mean field bound, which has been described for Boltzmann machines in [1] and later for sigmoid belief networks in [2]. Apart from its bounding properties, mean field theory is a commonly used approximation technique as well. Recently this first order bound was extended to a third order approximation for Boltzmann machines and sigmoid belief networks in [3] and [4], where it was shown that this particular third order expansion is still a bound. In 1996 an upper bound for Boltzmann machines was described in [5] and [6]. In the same articles the authors derive an upper bound for a special case of sigmoid belief networks: the two-layered networks. In this article we will focus solely on Boltzmann machines, but an extension to sigmoid belief networks is quite straightforward. This article is organized as follows: In section 2 we start with the general theory about bounding techniques. Later in that section the upper and lower bound are briefly described. For a full explanation we refer to the articles mentioned before. The section is concluded by explaining how these bounds on the partition function can be used to bound means and correlations. In section 3 results are shown for fully connected Boltzmann machines, where size of weights and thresholds as well as network size are varied. In section 4 we present our conclusions and outline possible extensions. 2 Theory There exists a general method to create a class of polynomials of a certain order, which all bound a function of interest, fo(x). Such a class of order 2n can be found if the 2n-th order derivative of fo(x), written as hn(x), can be bounded by a constant. When this constant is zero, the class is actually of order 2n-1. It turns out that this class is parameterized by n free parameters. Suppose we have a function b2k for some integer k which bounds the function 12k from below (an upper bound can be written as a lower bound by using the negative of both functions). Thus (1) Now construct the primitive functions 12k -1 and b2k -1 such that 12k - 1 (p) = b2k- 1(p) for a free to choose value for p. This constraint can always be achieved by adding an appropriate constant to the primitive function b2k- 1 . It is easy to prove that { 12k -1 (x) :S b2k -1 (x) 12k -1 (x) 2: b2k -1 (x) or in shorthand notation hk-1(x) § b2k- 1(X). for x < p for x 2: p (2) If we repeat this procedure and construct the primitive functions hk-2 and b2k- 2 such that hk-2(p) = b2k- 2(p) for the same p, one can verify that Vx hk-2(x) 2: b2k- 2(X) (3) Thus given a bound 12k (x) 2: b2k (x) we can construct a class of bounding functions for hk-2 parameterized by p. Since we assumed hn (x) can be bounded from below by a constant, we can apply the procedure n times and we finally find fa (x) 2: bo (x), where bo (x) is parameterized by n free parameters. This procedure can be found in more detail in [4]. 2.1 A third order lower bound for Boltzmann machines Boltzmann machines are stochastic neural networks with N binary valued neurons, Si, which are connected by symmetric weights Wij. Due to this symmetry the probability distribution is a Boltzmann-Gibbs distribution which is given by (see also [7]) p (siB, w) = ~ exp (~L. WijSiSj + L BiSi) = ~ exp (-E (s, B, w)) (4) 'J ' where the Bi are threshold values and Z (B , w) = L exp ( - E (s, B, w)) (5) all S is the normalization known as the partition function. This partition function is especially important, since statistical quantities such as means and correlations can be directly derived from it. For instance, the means can be computed as (sn) = LP (siB, w) Sn = L P (s, Sn =+l IB, w) - P (s, Sn = - l iB, w) all S all s/sn Z+ (B, w) - Z_ (B, w) Z (B, w) (6) where Z+ and Z_ are partition functions over a network with Sn clamped to +1 and -1 , respectively. This explains why the objective of almost any approximation method is the partition function given by equation 5. In [3] and [4] it is shown that the standard mean field lower bound can be obtained by applying the linear bound (7) to all exponentially many terms in the sum. Since J.l may depend on S, one can choose J.l (s) = J.li Si + J.lo , which leads to the standard mean field equations, where the J.li turn out to be the local fields. Moreover, the authors show that one can apply the procedure of 'upgrading bounds' (which is described briefly at the beginning of this section) to equation 7, which leads to the class of third order bounds for eX. This is achieved in the following way: 'r/X,V h(x) = eX 2': eV (1 + x - v) = b2(x) h(x)=ex'§ell-+ev ((1+J.l-v)(x-J.l)+~(x-J.l) 2) =bdx) (8) 'r/X,Il-,A fo(x) = eX 2': ell- { 1 + x - J.l + eA C ; >.. (x - J.l)2 + ~ (x - J.l)3) } = bo(x) with>" = v - J.l. In principle, this third order bound could be maximized with respect to all the free parameters, but here we follow the suggestion made in [4] to use a mean field optimization, which is much faster and generally almost as good as a full optimization. For more details we refer to [4]. 2.2 An upper bound An upper bound for Boltzmann machines has been described in [5] and [6]1. Basically, this method uses a quadratic upper bound on log cosh x, which can easily be obtained in the following way: h(x) = 1 - tanh2 x::; 1 = b2(x) h(x) = tanh x ~ x - J.l + tanhJ.l = bdx) (9) 1 2 fa (x) = log cosh x ::; "2 (x - J.l) + (x - J.l) tanh J.l + log cosh J.l = bo (x) Using this bound, one can derive Z (e , w) = L exp (~L WijSiSj + L eiSi) all s ij i = ~ 2exp (lOg cosh (L WniSi + en)) exp (~ .L WijSiSj + L eiSi) all sisn , 'J i'n ' i'n ::; L exp (~ L W~jSiSj + L e;Si + k) = ek . Z (e' , W') allsls n iji'n ii'n (10) INote: The articles referred to, use Si E {O, I} instead of the +1/-1 coding used here. where k is a constant and el and Wi are thresholds and weights in a reduced network given by I Wij = Wij + WniWnj e;j = ei + Wni (en J-Ln + tanhJ-Ln) (11) 1 2 1 2 k = "2 (en J-Ln + tanhJ-Ln) -"2 tanh J-Ln + log 2 cosh J-Ln Hence, equation 10 defines a recursive relation, where each step reduces the network by one neuron. Finally, after N steps, an upper bound on the partition function is found 2 . We did a crude minimization with respect to the free parameters J-L. A more sophisticated method can probably be found, but this is not the main objective of this article. 2.3 Bounding means and correlations The previous subsections showed very briefly how we can obtain a lower bound, ZL, and an upper bound, ZU , for any partition function. We can use this in combination with equation 6 to obtain a bound on the means: ZL _ ZU Zu _ ZL (sn)L = + X -::::; (sn)::::; + y - = (snt (12) where X = ZU if the nominator is positive and X = ZL otherwise. For Y it is the opposite. The difference, (sn)U - (sn)L, is called the bandwidth. Naively, we can compute the correlations similarly to the means using (13) where the partition function is computed for all combinations Sn Sm. Generally, however, this gives poor results, since we have to add four bounds together, which leads to a bandwidth which is about twice as large as for the means. We can circumvent this by computing the correlations using (14) where we allow the sum in the partition functions to be taken over Sn , but fix Sm either to Sn or its negative. Finally, the computation of the bounds (SnSm)L and (snsmt is analogue to equation 12. There exists an alternative way to bound the means and correlations. One can write ( ) _ Z+ - Z _ _ Z+/Z_ - 1 _ z - 1 - f ( ) Sn z Z+ + Z _ Z+/Z_ + 1 z + 1 with z = Z+/Z_ , which can be bounded by ZL Zu ----± < z < ----± Z~ Z~ (15) (16) Since f (z) is a monotonically increasing function of z, the bounds on (Sn) are given by applying this function to the left and right side of equation 16. The correlations can be bounded similarly. It is still unknown whether this algorithm would yield better results than the first one, which is explored in this article. 2The original articles show that it is not necessary to do all the N steps. However, since this is based on mixing approximation techniques with exact calculations, it is not used here as it would hide the real error the approximation makes. 13 ir==================i~----~----~ Exact Mean field lower bound 12.5 Upper bound Third order lower bound 12 11 10.5 10'--~·""'--: ' ' - ... ....... . _,_ ....... - ,'" o 0.2 0.4 a w 0.6 , , . ' 0.8 ,. , . , ,-,. ,. Figure 1: Comparison of 1) the mean field lower bound, 2) the upper bound and 3) the third order lower bound with the exact log partition function. The network was a fully connected Boltzmann machine with 14 neurons and (J'B = 0.2. The size of the weights is varied on the x-axis. Each point was averaged over 100 networks. 3 Results In all experiments we used fully connected Boltzmann machines of which the thresholds and weights both were drawn from a Gaussian with zero mean and standard deviation (J'B and (J'w/VN, respectively, where N is the network size. This is the so called sK-model (see also [8]). Generally speaking, the mean field approximation breaks down for (J'B = 0 and (J'w > 0.5, whereas it can be proven that any expansion based approximation is inaccurate when (J'w > 1 (which is the radius of convergence as in [9]). If (J'B #- 0 these maximum values are somewhat larger. In figure 1 we show the logarithm of the exact partition function , the first order or mean field bound, the upper bound (which is roughly quadratic) and the third order lower bound. The weight size is varied along the horizontal axis. One can see clearly that the mean field bound is not able to capture the quadratic form of the exact partition function for small weights due to its linear behaviour. The error made by the upper and third order lower bound is small enough to make non-trivial bounds on the means and correlations. An example of this bound is shown in figure 2 for the specific choice (J'B = (J'w = 0.4. For both the means and the correlations a histogram is plotted for the upper and lower bounds computed with equation 12. Both have an average bandwidth of 0.132, which is a clear subset of the whole possible interval of [-1 , 1]. In figure 3 the average bandwidth is shown for several values of (J'e and (J' w ' For bandwidths of 0.01,0.1 and 1 a line is drawn. We conclude that almost everywhere the bandwidth is non-trivially reduced and reaches practically useful values for (J'w less than 0.5. This is more or less equivalent to the region where the mean field approximation performs well. That approximation, however, gives no information on how close it actually is to the exact value, whereas the bounding method limits it to a definite region. Means 80 60 40 20 -0.2 0.2 Distance to exact Correlations 600,---------=----, ,------,6010 500 400 300 200 100 -0.2 -0.1 o Distance to exact 100 0.1 0.2 Figure 2: For the specific choice IJo = IJw = 0.4 thirty fully connected Boltzmann machines with 14 neurons were initialized and the bounds were computed. The two left panels show the distance between the lower bound and the exact means (left) and similarly for the upper bound (right). The right two panels show the distances of both bounds for the correlations. 0.8 1.5 1.5 0.6 o~ 0.4 0.5 0.5 0.2 00 Ow Figure 3: In the left panel the average bandwidth is colour coded for the means, where IJo and IJw are varied in ten steps along the axes. The right panel shows the same for the correlations. For each IJo , IJw thirty fully connected Boltzmann machines were initialized and the bounds on all the means and correlations were computed. For three specific bandwidths a line is drawn. 0.01 0.4 2 0.008 °e=0.3 °e=0.3 °e=0.3 0 =0.1 0.3 ° =0.3 1.5 ° =0.5 .r::; w w w ~0006 0.2 "0 @0.004 c:J 0.002 0.1 0.5 00 10 20 30 40 00 10 20 30 40 00 10 20 30 40 Network size Figure 4: For (Tw = 0.1, 0.3 and 0.5 the bandwidth for the correlations is shown versus the network size. (To = 0.3 in all cases, but the plots are nearly the same for other values. Please note the different scales for the y-axis. A similar graph for the means is not shown here, but it is roughly the same. The solid line is the average bandwidth over all correlations, whereas the dashed lines indicate the minimum and maximum bandwidth found. Unfortunately, the bounds have the unwanted property that the error scales badly with the size of the network. Although this makes the bounds unsuitable for very large networks, there is still a wide range of networks small enough to take advantage of the proposed method and still much too large to be treated exactly. The bandwidth versus network size is shown in figure 4 for three values of (T w' Obviously, the threshold of practical usefulness is reached earlier for larger weights. Finally, we remark that the computation time for the upper bound is (') (N4) and (') (N 3 ) for the mean field and third order lower bound. This is not shown here. 4 Conclusions In this article we combined two already existing bounds in such a way that not only the partition function of a Boltzmann machine is bounded from both sides, but also the means and correlations. This may seem superfluous, since there exist already several powerful approximation methods. Our method, however, can be used apart from any approximation technique and gives at least some information you can rely on. Although approximation techniques might do a good job on your data, you can never be sure about that. The method outlined in this paper ensures that the quantities of interest, the means and correlations, are restricted to a certain region. We have seen that , generally speaking, the results are useful for weight sizes where an ordinary mean field approximation performs well. This makes the method applicable to a large class of problems. Moreover, since many architectures are not fully connected, one can take advantage of that structure. At least for the upper bound it is shown already that this can improve computation speed and tightness. This would partially cancel the unwanted scaling with the network size. Finally, we would like to give some directions for further research. First of all, an extension to sigmoid belief networks can easily be done, since both a lower and an upper bound are already described. The upper bound, however, is only applicable to two layer networks. A more general upper bound can probably be found. Secondly one can obtain even better bounds (especially for larger weights) if the general constraint (17) is taken into account. This might even be extended to similar constraints, where three or more neurons are involved. Acknowledgelllents This research is supported by the Technology Foundation STW, applied science devision of NWO and the technology programme of the Ministry of Economic Affairs. References [1] C. Peterson and J. Anderson. A mean field theory learning algorithm for neural networks. Complex systems, 1:995- 1019, 1987. [2] S.K. Saul, T.S. Jaakkola, and M.l. Jordan. Mean field theory for sigmoid belief networks. Journal of Artificial Intelligence Research, 4:61- 76, 1996. [3] Martijn A.R. Leisink and Hilbert J. Kappen. A tighter bound for graphical models. In Todd K. Leen, Thomas G. Dietterich, and Volker Tresp, editors, Advances in Neural Information Processing Systems 13, pages 266- 272. MIT Press, 2001. [4] Martijn A.R. Leisink and Hilbert J. Kappen. A tighter bound for graphical models. Neural Computation, 13(9), 2001. To appear. [5] T. Jaakkola and M.l. Jordan. Recursive algorithms for approximating probabilities in graphical models. MIT Compo Cogn. Science Technical Report 9604, 1996. [6] Tommi S. Jaakkola and Michael 1. Jordan. Computing upper and lower bounds on likelihoods in intractable networks. In Proceedings of the Twelfth Annual Conference on Uncertainty in Artificial Intelligence (UAI- 96), pages 340- 348, San Francisco, CA, 1996. Morgan Kaufmann Publishers. [7] D. Ackley, G. Hinton, and T. Sejnowski. A learning algorithm for Boltzmann machines. Cognitive Science, 9:147-169, 1985. [8] D. Sherrington and S. Kirkpatrick. Solvable model of a spin-glass. Physical Review Letters, 35(26):1793-1796, 121975. [9] T. Plefka. Convergence condition of the TAP equation for the infinite-ranged ising spin glass model. J.Phys.A: Math. Gen., 15:1971-1978, 1981.
|
2001
|
1
|
1,905
|
Group Redundancy Measures Reveal Redundancy Reduction in the Auditory Pathway Gal Chechik Amir Globerson Naftali Tishby School of Computer Science and Engineering and The Interdisciplinary Center for Neural Computation Hebrew University of Jerusalem, Israel ggal@cs.huji.ac.il Michael J. Anderson Eric D. Young Department of Biomedical Engineering Johns Hopkins University, Baltimore, MD, USA Israel N elken Department of Physiology, Hadassah Medical School and The Interdisciplinary Center for Neural Computation Hebrew University of Jerusalem, Israel Abstract The way groups of auditory neurons interact to code acoustic information is investigated using an information theoretic approach. We develop measures of redundancy among groups of neurons, and apply them to the study of collaborative coding efficiency in two processing stations in the auditory pathway: the inferior colliculus (IC) and the primary auditory cortex (AI). Under two schemes for the coding of the acoustic content, acoustic segments coding and stimulus identity coding, we show differences both in information content and group redundancies between IC and AI neurons. These results provide for the first time a direct evidence for redundancy reduction along the ascending auditory pathway, as has been hypothesized for theoretical considerations [Barlow 1959,2001]. The redundancy effects under the single-spikes coding scheme are significant only for groups larger than ten cells, and cannot be revealed with the redundancy measures that use only pairs of cells. The results suggest that the auditory system transforms low level representations that contain redundancies due to the statistical structure of natural stimuli, into a representation in which cortical neurons extract rare and independent component of complex acoustic signals, that are useful for auditory scene analysis. 1 Introduction How do groups of sensory neurons interact to code information and how do these interactions change along the ascending sensory pathways? According to the a common view, sensory systems are composed of a series of processing stations, representing more and more complex aspects of sensory inputs. The changes in representations of stimuli along the sensory pathway reflect the information processing performed by the system. Several computational principles that govern these changes were suggested, such as information maximization and redundancy reduction [2, 3, 11]. In order to investigate such changes in practice, it is necessary to develop methods to quantify information content and redundancies among groups of neurons, and trace these measures along the sensory pathway. Interactions and high order correlations between neurons were mostly investigated within single brain areas on the level of pairs of cells (but also for larger groups of cells [9]) showing both synergistic and redundant interactions [8, 10, 21, 6, 7, 13]. The current study develops information theoretic redundancy measures for larger groups of neurons, focusing on the case of stimulus-conditioned independence. We then compare these measures in electro-physiological recordings from two auditory stations: the auditory mid-brain and the primary auditory cortex. 2 Redundancy measures for groups of neurons To investigate high order correlations and interactions within groups of neurons we start by defining information measures for groups of cells and then develop information redundancy measures for such groups. The properties of these measures are then further discussed for the specific case of stimulus-conditioned independence. Formally, the level of independence of two variables X and Y is commonly quantified by their mutual information (MI) [17,5]. This well known quantity, now widely used in analysis of neural data, is defined by J(X; Y) = DKL[P(X, Y)IIP(X)P(Y)] = ~p(x, y)log (:~~~~~)) (1) and measures how close the joint distribution P(X, Y) is to the factorization by the marginal distributions P(X)P(Y) (DKL is the Kullback Leiber divergence [5]). For larger groups of cells, an important generalized measure quantifies the information that several variables provide about each other. This multi information measure [18] is defined by (2) Similar to the mutual information case, the multi information measures how close the joint distribution is to the factorization by the marginals. It thus vanishes when variables are independent and is otherwise positive. We now turn to develop measures for group redundancies. Consider first the simple case of a pair of neurons (Xl, X 2 ) conveying information about the stimulus S. In this case, the redundancy-synergy index ([4, 7]) is defined by (3) Intuitively, RSpairs measures the amount of information on the stimulus S gained by observing the joint distribution of both Xl and X 2 , as compared with observing the two cells independently. In the extreme case where Xl = X 2 , the two cells are completely redundant and provide the same information about the stimulus, yielding RSpairs = I(Xl' X 2 ; S) - I(Xl ; S) - I(X2 ; S) = -I(Xl; S), which is always non-positive. On the other hand, positive RSpairs values testify for synergistic interaction between Xl and X 2 ([8, 7, 4]). For larger groups of neurons, several different measures of redundancy-synergy may be considered, that encompass different levels of interactions. For example, one can quantify the residual information obtained from a group of N neurons compared to all its N - 1 subgroups. As with inclusion-exclusion calculations this measure takes the form of a telescopic sum: RSNIN-l = I(XN; S) - L{XN-l} I(XN-\ S) + ... + (_l)N-l L{Xd I(Xi ; S), where {Xk} are all the subgroups of size k out of the N available neurons. Unfortunately, this measure involves 2N information terms, making its calculation infeasible even for moderate N values 1. A different RS measure quantifies the information embodied in the joint distribution of N neurons compared to that provided by N single independent neurons, and is defined by N RSNll = I(Xl ' ... , XN; S) - 2..: I(Xi ; S) (4) i=l Interestingly, this synergy-redundancy measure may be rewritten as the difference between two multi-information terms N I(Xl ' ... , XN; S) - 2..: I(Xi ; S) = (5) i = l N H(Xl' ... ,XN) - H(Xl' ... , XNIS) - 2..: H(Xi) - H(XiIS) = i=l I(Xl ; ... ; XNIS) - I(Xl ; ... ;XN) where H(X) = - L xP(x)log(P(x)) is the entropy of X 2 . We conclude that the index RSNll can be separated into two terms: one that is always non-negative, and measures the coding synergy, and the second which is always non-positive and quantifies the redundancy. These two terms correspond to two types of interactions between neurons: The first type are within-stimulus correlations (sometimes termed noise correlations) that emerge from functional connections between neurons and contribute to synergy. The second type are between stimulus correlations (or across stimulus correlations) that reflect the fact that the cells have similar responses per stimulus, and contribute to redundancy. Being interested in the latter type of correlations, we limit the discussion to the redundancy term -I(Xl; ... ; XN)' Formulating RSNll as in equation 5 proves highly useful when neural activities are independent given the stimulus P(XIS) = II~l P(XiIS). In this case, the first (synergy) term vanishes, thus limiting neural interactions to the redundant lOur results below suggest that some redundancy effects become significant only for groups larger than 10-15 cells. 2When comparing redundancy in different processing stations, one must consider the effects of the baseline information conveyed by single neurons. We thus use the normalized redundancy (compare with [15] p.315 and [4]) defined by !iSNll = RSNldI(Xl; ... ; X N; S) regime. More importantly, under the independence assumption we only have to estimate the marginal distributions P(XiIS = s) for each stimulus s instead of the full distribution P(XIS = s). It thus allows to estimate an exponentially smaller number of parameters, which in our case of small sample sizes, provides more accurate information estimates. This approximation makes it possible to investigate redundancy among considerably larger groups of neurons than the 2-3 neuron groups considered previously in the literature. How reasonable is the conditional-independence approximation ? It is a good approximation whenever neuronal activity is mostly determined by the presented stimulus and to a lesser extent by interactions with nearby neurons. A possible example is the high input regime of cortical neurons receiving thousands of inputs, where a single input has only a limited influence on the activity of the target cell. The experimental evidence in this regard is however mixed (see e.g.[9]). One should note however, that stimulus-conditioned independence is implicitly assumed in analysis of non-simultaneously recorded data. To summarize, the stimulus-conditioned independence assumption limits interactions to the redundant regime, but allows to compare the extent of redundancy among large groups of cells in different brain areas. 3 Experimental Methods To investigate redundancy in the auditory pathway, we analyze extracellular recordings from two brain areas of gas-anesthetized cats: 16 cells from the Inferior Colliculus (Ie) - the third processing station of the ascending auditory pathway - and 19 cells from the Primary Auditory Cortex (AI) - the fifth station. Neural activity was recorded non-simultaneously from a total of 6 different animals responding to a set of complex natural and modified stimuli. Because cortical auditory neurons respond differently to simple and complex stimuli [12, 1], we refrain from using artificial over-simplified acoustic stimuli but instead use a set of stimuli based on bird vocalizations which contains complex 'real-life' acoustic features. A representative example is shown in figure 1. Q) "0 . .e "1i E '" 20 40 60 80 time (milliseconds) 100 7 6 20 40 60 80 100 time (milliseconds) Figure 1: A representative stimulus containing a short bird vocalization recorded in a natural environment. The set of stimuli consisted of similar natural and modified recordings. A. Signal in time domain B. Signal in frequency domain. 4 Experimental Results In practice, in order to estimate the information conveyed by neural activity from limited data, one must assume a decoding procedure, such as focusing on a simple statistic of the spike trains that encompasses some of its informative properties. In this paper we consider two extreme cases: coding short acoustic segments with single spikes and coding the stimulus identity with spike counts in a long window. In addition, we estimated information and redundancy obtained with two other statistics. First, the latency of the first spike after stimulus onset, and secondly, a statistic which generalizes the counts statistics for a general renewal process [19]. These calculations yielded higher information content on average, but similar redundancies as presented below. Their detailed results will be reported elsewhere. 1.2 ~0 . 8 .$ :0 iO.6 0.4 Auditory Cortex (AI) o~========== 5 10 15 no of cells 0.15 ~ 0.1 '" c u C :::l U ~ 0.05 <ii c o t5 jg 0 Inlerior Colliculus (IC) -0.05 L--".--'--~---~--~--o 5 10 15 20 no of cells Figure 2: A. Information about stimulus frames as a function of number of cells. Information calculation was repeated for several subgroups of each size, and with several random seed initializations. The dark curve depicts the expected information provided by independent neurons (this expected curve is corrected for saturation effects [16] and is thus sub linear). The curved line depicts average information from joint distribution of sets of neurons Mean[J(Xl' ... Xk; S)]. All information estimations were corrected for small-samples bias by shuffling methods [14] . B. Fractional redundancy (difference of the mutual information from the expected baseline information divided by the baseline) as a function of number of neurons. 4.1 Coding acoustics with single spikes The current section focuses on the relation between single spikes and short windows of the acoustic stimuli shortly preceding them (which we denote as frames). As the set of possible frames is very large and no frame actually repeats itself, we must first pre-process the stimuli to reduce frames dimensionality. To this end, we first transformed the stimuli into the frequency domain (roughly approximating the cochlear transformation) and then extracted overlapping windows of 50 millisecond length, with 1 millisecond spacing. This set was clustered into 32 representatives, using a metric that groups together acoustic segments with the same spectro-temporal energy structure. This representation allowed us to estimate the joint distribution (under the stimulus-conditioned independence assumption) of cells' activity and stimuli, for groups of cells of different sizes. Figure 2A shows the mutual information between spikes and stimulus frames as a function of the number of cells for both AI and Ie neurons. Ie neurons convey high information but largely deviate from the information expected for independent neurons. On the other hand, AI neurons provide an order of magnitude less information than Ie cells but their information sums almost linearly, as expected from independent neurons. The difference between an information curve and its linear baseline measures the redundancy RSNII of equation 5. Figure 2B presents the normalized redundancy as a function of number of cells, showing that Ie cells are significantly more redundant than AI cells. Q) u c 0.6rr==--------,--------; D Primary Auditory Cortex A 1 _ Inferior Colliculus IC 0.5 ~'-"'=-c.:..::..::.:..:.:::.'--===::...:..::-----" 504 u u o 00.3 :§"' ~02 o 0. 0.1 o -0.5 -04 -0.3 -0.2 -0.1 pairwise redundancy -I(X;Y)/I(X;Y;S) o 0.6rr==--------,--------; D Primary Auditory Cortex A 1 _ Inferior Colliculus IC 0.5 ~'-"'=--"-"-':.:..:.:::.'--===::...:..::'---------" Q) u c 504 u u o 00.3 :§"' ~0.2 o 0. 0.1 -8.8 -0.6 -04 -0.2 0 triplets fractional redundancy -I(X;Y;Z)/I(X;Y;Z;S) Figure 3: Distribution of pairs (A.) and triplets (B.) normalized redundancies. AI cells (light bars) are significantly more independent than Ie cells (dark bars). Spike counts were collected over a window that maximizes mean single cells MI. Number of bins in counts-histogram was optimized separately for every cell. Information estimations were corrected for small-samples bias by shuffling methods [14]. 4.2 Coding stimuli by spike counts We now turn to investigate a second coding paradigm, and calculate the information conveyed by AI and Ie spike counts about the identity of the presented stimulus. To this end, we calculate a histogram of spike counts and estimate the counts' distribution as obtained from repeated presentations of the stimuli. The distribution of fractional redundancy in pairs of AI and Ie neurons is presented in figure 3A, and that of triplets in figure 3B 3 . As in the case of coding with single spikes, single AI cells convey on average less information about the stimulus. However, they are also more independent, thus making it possible to gain more information from groups of neurons. Ie neurons on the other hand, provide more information when considered separately but are more redundant. As in the case of coding acoustics with single spikes, single Ie cells provide more information than AI cells (data not shown) but this time AI cells convey half the information that Ie cells provide, while they convey ten times less information than Ie cells about acoustics. This suggests that AI cells poorly code the physical characteristics of the sound but convey information about its global properties. To illustrate the high information provided by both sets, we trained a neural network classifier that predicts the identity of the presented stimulus according to spike counts of a limited set of neurons. Figure 4 shows that both sets of neurons achieve considerable prediction accuracy, but Ie neurons obtain average accuracy of more than 90 percent already with five cells, while the average prediction accuracy using cortical neurons rises continuously 4. 3Unlike the binary case of single spikes, the limited amount of data prevents a robust estimation of information from spike counts for more than triplets of cells. 4The probability of accurate prediction is exponentially related to the input-output mutual information, via the relation Pcorrect = exp( -missing nats) yielding Mlnats = In(no. of stimuli) + In(Pcorrect). Classification thus provides lower bounds on information content. Figure 4. Prediction accuracy of stimulus identity as a function of number of Ie (upper curve) and AI (lower curve) cells used by the classifier. Error bars denote standard deviation across several subgroups of the same size. For each subgroup, a one-hidden layer neural network was trained separately for each stimulus using some stimulus presentations as a training set and the rest for testing. Performance reported is for the testing set. 5 Discussion >-" 0.95 [IS 0.9 :0 " al <:0.85 o "u ~ 0.8 Q. 0.75 I I I I I I • I I I I I I I I I I I I I I I I jI" I 0.7 '-------=---~--~---~--~ 5 10 15 20 number of cells We have developed information theoretic measures of redundancy among groups of neurons and applied them to investigate the collaborative coding efficiency in the auditory modality. Under two different coding paradigms, we show differences in both information content and group redundancies between Ie and cortical auditory neurons. Single Ie neurons carry more information about the presented stimulus, but are also more redundant. On the other hand, auditory cortical neurons carry less information but are more independent, thus allowing information to be summed almost linearly when considering groups of few tens of neurons. The results provide for the first time direct evidence for redundancy reduction along the ascending auditory pathway, as has been hypothesized by Barlow [2, 3]. The redundancy effects under the single-spikes coding paradigm are significant only for groups larger than ten cells, and cannot be revealed with the standard redundancy measures that use only pairs of cells. Our results suggest that transformations leading to redundancy reduction are not limited to low level sensory processing (aimed to reduce redundancy in input statistics) but are applied even at cortical sensory stations. We suggest that an essential experimental prerequisite to reveal these effects is the use of complex acoustic stimuli whose processing occurs at high level processing stations. The above findings are in agreement with the view that along the ascending sensory pathways, the number of neurons increase, their firing rates decrease, and neurons become tuned to more complex and independent features. Together, these suggest that the neural representation is mapped into a representation with higher effective dimensionality. Interestingly, recent advances in kernel-methods learning [20] have shown that nonlinear mapping into higher dimension and over-complete representations may be useful for learning of complex classifications. It is therefore possible that such mappings provide easier readout and more efficient learning in the brain. Acknowledgements This work supported in part by a Human Frontier Science Project (HFSP) grant RG 0133/1998 and by a grant from the Israeli Ministry of Science. References [1] O. Bar-Yosef and I. Nelken. Responses of neurons in cat primary auditory cortex to bird chirps: Effects of temporal and spectral context. J. Neuroscience, in press, 2001. [2] H.B. Barlow. Sensory mechanisms, the reduction of redundancy, and intelligence. In Mechanisation of thought processes, pages 535- 539. Her Majesty's stationary office, London, 1959. [3] H.B. Barlow. Redundancy reduction revisited. Network: Computation in neural systems, 12:241-253, 200l. [4] N. Brenner, S.P. Strong, R . Koberle, R. de Ruyter van Steveninck, and W. Bialek. Synergy in a neural code. Neural Computation, 13(7):1531, 2000. [5] T.M. Cover and J.A. Thomas. The elements of information theory. Plenum Press, New York, 1991. [6] Y. Dan, J.M. Alonso, W.M. Usrey, and R.C. Reid. Coding of visual information by precisely correlated spikes in the lateral geniculate nucleus. Nature Neuroscience, 1(6):501- 507, 1998. [7] I. Gat and N. Tishby. Synergy and redundancy among brain cells of behaving monkeys. In M.S. Kearns, S.A. Solla, and D.A.Cohn, editors, Advances in Neural Information Proceedings systems, volume 11, Cambridge, MA, 1999. MIT Press. [8] T.J. Gawne and B.J. Richmond. How independent are the messages carried by adjacent inferior temporal cortical neurons? J. Neurosci., 13(7):2758- 2771, 1993. [9] P.M. Gochin, M. Colombo, G. A. Dorfman, G.L. Gerstein, and C.G. Gross. Neural ensemble coding in inferior temporal cortex. J. Neurophysiol., 71:2325- 2337, 1994. [10] M. Meister. Multineural codes in retinal signaling. Proc. Natl. Acad. Sci., 93:609- 614, 1996. [11] J .P. Nadal, N. Brunei, and N. Parga. Nonlinear feedforward networks with stochastic outputs: infomax implies redundancy reduction. Network: Computation in neural systems, 9:207- 217, 1998. [12] I. Nelken, Y. Rotman, and O. Bar-Yosef. Specialization of the auditory system for the analysis of natural sounds. In J. Brugge and P.F. Poon, editors, Central Auditory Processing and Neural Modeling. Plenum, New York, 1997. [13] S. Nirenberg, S.M. Carcieri, A.L. Jacobs, and P.E. Latham. Retinal ganglion cells act largely as independent encoders. Nature, 411:698- 701, 200l. [14] LM. Optican, T.J. Gawne, B.J. Richmond, and P.J . Joseph. Unbiased measures of transmitted information and channel capacity from multivariate neuronal data. Bioi. Cyber, 65:305- 310, 1991. [15] E. T. Rolls and A. Treves. Neural Networks and Brain Function. Oxford Univ. Press, 1998. [16] I. Samengo. Independent neurons representing a fintie set of stimuli: dependence of the mutual information on the number of units sampled. Network: Comput. Neural Syst., 12:21- 31, 200l. [17] C.E. Shanon. A mathematical theory of communication. The Bell systems technical journal, 27:379- 423,623- 656, 1948. [18] M. Studenty and J. Vejnarova. The multiinformation function as a tool for measuring stochastic dependence. In M.I. Jordan, editor, Learning in Graphical Models, pages 261-297. Dordrecht: Kluwer, 1998. [19] C. van Vreeswijk. Information trasmission with renewal neurons. In J.M. Bower, editor, Computational Neuroscience: Trends in Research. Elsevier Press, 200l. [20] V.N. Vapnik. The nature of statistical learning theory. Springer-Verlag, Berlin, 1995. [21] DK. Warland, P. Reinagel, and M. Meister. Decoding visual information from a population of retinal ganglion cells. J. Neurophysiol., 78:2336- 2350, 1997.
|
2001
|
10
|
1,906
|
Global Coordination of Local Linear Models Sam Roweis , Lawrence K. Saul , and Geoffrey E. Hinton Department of Computer Science, University of Toronto Department of Computer and Information Science, University of Pennsylvania Abstract High dimensional data that lies on or near a low dimensional manifold can be described by a collection of local linear models. Such a description, however, does not provide a global parameterization of the manifold—arguably an important goal of unsupervised learning. In this paper, we show how to learn a collection of local linear models that solves this more difficult problem. Our local linear models are represented by a mixture of factor analyzers, and the “global coordination” of these models is achieved by adding a regularizing term to the standard maximum likelihood objective function. The regularizer breaks a degeneracy in the mixture model’s parameter space, favoring models whose internal coordinate systems are aligned in a consistent way. As a result, the internal coordinates change smoothly and continuously as one traverses a connected path on the manifold—even when the path crosses the domains of many different local models. The regularizer takes the form of a Kullback-Leibler divergence and illustrates an unexpected application of variational methods: not to perform approximate inference in intractable probabilistic models, but to learn more useful internal representations in tractable ones. 1 Manifold Learning Consider an ensemble of images, each of which contains a face against a neutral background. Each image can be represented by a point in the high dimensional vector space of pixel intensities. This representation, however, does not exploit the strong correlations between pixels of the same image, nor does it support many useful operations for reasoning about faces. If, for example, we select two images with faces in widely different locations and then average their pixel intensities, we do not obtain an image of a face at their average location. Images of faces lie on or near a low-dimensional, curved manifold, and we can represent them more usefully by the coordinates on this manifold than by pixel intensities. Using these “intrinsic coordinates”, the average of two faces is another face with the average of their locations, poses and expressions. To analyze and manipulate faces, it is helpful to imagine a “magic black box” with levers or dials corresponding to the intrinsic coordinates on this manifold. Given a setting of the levers and dials, the box generates an image of a face. Given an image of a face, the box deduces the appropriate setting of the levers and dials. In this paper, we describe a fairly general way to construct such a box automatically from an ensemble of high-dimensional vectors. We assume only that there exists an underlying manifold of low dimensionality and that the relationship between the raw data and the manifold coordinates is locally linear and smoothly varying. Thus our method applies not only to images of faces, but also to many other forms of highly distributed perceptual and scientific data (e.g., spectrograms of speech, robotic sensors, gene expression arrays, document collections). 2 Local Linear Models The global structure of perceptual manifolds (such as images of faces) tends to be highly nonlinear. Fortunately, despite their complicated global structure, we can usually characterize these manifolds as locally linear. Thus, to a good approximation, they can be represented by collections of simpler models, each of which describes a locally linear neighborhood[3, 6, 8]. For unsupervised learning tasks, a probabilistic model that nicely captures this intuition is a mixture of factor analyzers (MFA)[5]. The model is used to describe high dimensional data that lies on or near a lower dimensional manifold. MFAs parameterize a joint distribution over observed and hidden variables:
(1) where the observed variable, ! , represents the high dimensional data; the discrete hidden variables, "$# % &')()(*(+-, , indexes different neighborhoods on the manifold; and the continuous hidden variables, ./0!1 , represent low dimensional local coordinates. The model assumes that data is sampled from different neighborhoods on the manifold with prior probabilities 243 , and that within each neighborhood, the data’s local coordinates are normally distributed1 as: 256&87.:9 1;<
=?>A@BDC % & *E F ( (2) Finally, the model assumes that the data’s high and low dimensional coordinates are related by linear processes parameterized by centers G , loading matrices H and noise levels I :
J &7 I 8 9K ;< =?>A@LBDC % &M C G C H : N E I 9
K M C G C H : N F ( (3) The marginal data distribution, O , is obtained by summing/integrating out the model’s discrete and continuous latent variables. The result is a mixture of Gaussian distributions with parameterized covariance matrices of the form: 2QP 3R- &872 H H E "S I ) 9
K ;< =)>T@ B C % &LM C G N E H H E S I 9
K M C G N:FU( (4) The learning problem for MFAs is to estimate the centers G , transformations H , and noise levels I of these linear processes, as well as the prior probabilities 3 of sampling data from different parts of the manifold. Parameter estimation in MFAs can be handled by an Expectation-Maximization (EM) algorithm[5] that attempts to maximize the logprobability, VXWY Z , averaged over training examples. Note that the parameter space of this model exhibits an invariance: taking H -[ H :\/ , where \ are ]_^`] orthogonal matrices ( \ \ E ba ), does not change the marginal distribution, 2 . The transformations H [ H \ correspond to arbitrary rotations and reflections of the local coordinates in each linear model. The objective function for the EM algorithm is unchanged by these transformations. Thus, maximum likelihood estimation in MFAs does not favor any particular alignment; instead, it produces models whose internal representations change unpredictably as one traverses connected paths on the manifold. Can we encourage models whose local coordinate systems are aligned in a consistent way? 3 Global Coordination Suppose the data lie near a smooth manifold with a locally flat (developable) structure. Then there exist a single set of “global coordinates” c which parametrize the manifold 1Although in principle each neighborhood could have a different prior on its local coordinates, without loss of generality we have made the standard assumption that d/egfih*j k?l is the same for all settings of k and absorbed the shape of each local Gaussian model into the matrices m2h . s,z g x global hidden variables coordinates data Figure 1: Graphical model for globally coordinated MFAs. Although global coordinates are unobserved, they affect the learning through a regularization term. After learning, inferences about the global variables are made by computing posterior distributions, d/e j .l . Likewise, data can easily be generated by sampling from the conditional distribution, d/eZj l . All these operations are particularly tractable due to the conditional independencies of the model. everywhere. Furthermore, to a good approximation, these global coordinates can be related to the local coordinates of different neighborhoods (in their region of validity) by linear2 transformations: c 6
S ( (5) What does it mean to say that the coordinates c ? provide a global parameterization of the manifold? Intuitively, if a data point belongs to overlapping neighborhoods, then the global coordinates computed from their local coordinate systems, given by eq. (5), should agree. We can formalize this “global coordination” of different local models by treating the coordinates c as unobserved variables and incorporating them into the probabilistic model: c
$ A c C C ) (6) (Here we posit a deterministic relationship between local and global coordinates, although it is possible to add noise to this mapping as well.) The globally coordinated MFA is represented by the graphical model in Fig. 1. We can appeal to its conditional independencies to make other useful inferences. In particular: c R: ] RR c
' (7) c ' P 6 ' c R: ?( (8) Now, if two or more mixture components—say, K and < —explain a data point with non-negligible probability, then the posterior distributions for the global coordinates of this data point, as induced by eq. (8), should be nearly identical: that is, c R: K c < . To enforce this criterion of agreement, we need to penalize models whose posterior distributions c ' given by eq. (8) are multimodal, since multiple modes only arise when different mixture components give rise to inconsistent global coordinates. While directly penalizing multimodality of c ' is difficult, a penalty which encourages consistency can be easily incorporated into the learning algorithm. We introduce a family of unimodal distributions over both c and , and encourage the true posteriors, c T ' , to be close to some member, c : ' , of this family. Developing this idea further, we introduce a new objective function for unsupervised learning in MFAs. The new objective function incorporates a regularizer to encourage the global consistency of local models: P VXWY ' C P c c T
i VXWY c :
i c T (9) The first term in this objective function computes the log-probability of the data. The second term computes a sum of Kullback-Leibler (KL) divergences; these are designed to 2Without loss of generality, the matrices h can be taken to be symmetric and positive-definite, by exploiting the polar factorization and absorbing reflection and rotation into the local coordinate systems. (In practice, though, it may be easier to optimize the objective function without constraining the matrices to be of this form.) In the experiments reported below, we have further restricted them to be diagonal. Together, then, the coordination matrices h and vectors
h account for an axis-aligned scaling and uniform translation between the global and local coordinate systems. penalize MFAs whose posterior distributions over global coordinates are not unimodal. The twin goals of density estimation and manifold learning in MFAs are pursued by attempting to balance these terms in the objective function. The factor controls the tradeoff between density modeling and global coordination: as [ only strict invariances (which do not affect likelihood) are exploited in order to achieve submodel agreement. In what follows we have set % arbitrarily; further optimization is possible. The most convenient way to parameterize the family of unimodal distributions is a factorized form involving a Gaussian density and a multinomial: c T '2 c ' T ' c ' c ' 6T '2 (10) Note that the distribution c T ' in eq. (10) factorizes over and c , implying that— according to this family of models—the global coordinate c is independent of the mixture component given the data point . Also, c is Gaussian, and thus unimodal. These are exactly the constraints we wish to impose on the posterior c :T i . At each iteration of learning, the means c , covariance matrices , and mixture weights are determined separately for each data point, so as to maximize the objective function in eq. (9): this amounts to computing the unimodal distributions, c :T , best matched to the true posterior distributions, c : ' . 4 Learning Algorithm Latent variable models are traditionally estimated by maximum likelihood or Bayesian methods whose objective functions do not reward the interpretability of their internal representations. Note how the goal of developing more useful internal representations has changed the learning problem in a fundamental way. Now we have additional “coordination” parameters–the offsets and weights –that must also be learned from examples. We also have auxiliary parameters for each data point—the means c , covariance matrices , and mixture weights —that determine the target distributions, c T ' . All these parameters, as well as the MFA model parameters #3 H G I , , must be chosen to “stitch together” the local coordinates systems in a smooth way and to learn internal representations easily coordinated by the local-to-global mapping in eq. (6). Optimization of the objective function in eq. (9) is reminiscent of so-called “variational” methods for approximate learning[7]. In these methods, an approximation to an exact (but intractable) posterior distribution is fitted by minimizing a KL divergence between the two distributions. The auxiliary parameters of the approximating distribution are known as variational parameters. Our objective function illustrates an unexpected application of such variational methods: not to perform approximate inference in intractable probabilistic models, but to learn more useful internal representations in tractable ones. We introduce the unimodal and factorized distributions c T ' to regularize the multimodal distributions c T . Penalizing the KL divergence between these distributions lifts a degeneracy in the model’s parameter space and favors local linear models that can be globally aligned. 4.1 Computing and optimizing the objective function Evaluating the objective function in eq. (9) requires a sum and integral over the latent variables of the model. These operations are simplified by rewriting the objective function as: P c c : ' M C V W Y c T
' S VXWY R c N ( (11) The factored form of the distributions c :T ' makes it straightforward to perform the required sums and integrals. The final result is a simple form in terms of entropies
and energies associated with the th data point: P O
C (12)
% & VXWY C V W Y S ] & V W Y &7. (13) % & c E c S % & E I 9
K C c E E H E I 9
K S % & M
N S % & VXWY I 8 S V W Y C VXWY 3R S S ] & VXWY 6&87.? (14) where we have introduced simplifying notation for the vector differences C G and c c C and the local precision matrices 9K a S H E I 9
K H 9
K . Iteratively maximizing the objective function by coordinate ascent now leads to a learning algorithm of the same general style as EM. 4.2 E-step Maximizing the objective function, eq. (9), with respect to the regularizing parameters # c R , (and subject to the constraint % ) leads to the fixed point equations: P 9K c P 9! #" %$ 9 &" $ ( (15) where " S E H E I 9
K . These equations can be solved by iteration with initialization 43 . Notice that and only need to be computed once before iterating the fixed point equations. The objective function is completely invariant to translation and rescaling of c and (since , and c appear only in the form c C E ). To remove this degeneracy, after solving the equations above we further constrain the global coordinates to have mean zero and unit variance in each direction. These constraints are enforced without changing the value of the objective function by simply translating the offsets ' and rescaling the diagonal matrices . 4.3 M-step The M-step consists of maximizing the objective function, eq. (9), with respect to the generative model parameters. Let us denote the updated parameter estimates by #)( 3R +* ,* G ( H ( I ( *, . Letting , the M-step updates for the first three of these are: ( 3 .*!/ P %$ $ * , 9
K P c * G 9K P Z R( (16) The remaining updates, to be performed in the order shown, are given in terms of updated difference vectors *
C * G 0* c c C * , the correlations 1 .* +* c E , and the variances 2 M S * c * c E N . ( H 1 2 9
K (17) 3 ( I 465 9
K P B 3 * C ( H 9
K * c 4 < 5 S 3 ( H 9K E ( H E 465?F (18) ( 9
K a S H E I 9
K H 9
K87 E * S H E ( I 9
K 1 +9 2 9K (19) At the optimum, the coordination weights satisfy an algebraic Riccati equation which can be solved by iterating the update shown above. (Such equations can also be solved by much more sophisticated methods well known in the engineering community. Most approaches involve inverting the previous value of which may be expensive for full matrices but is fast in our diagonal implementation.) Figure 2: Global coordination of local linear models. (left) A model trained using maximum likelihood, with the arrows indicating the direction of increase for each factor analyzer’s local coordinate system. (right) A coordinated model; arrows indicate the direction in the data space corresponding to increasing the global coordinate as inferred by the algorithm. The ellipses show the one standard deviation contour of the density of each analyzer. 5 Experiments We have tested our model on simple synthetic manifolds whose structure is known as well as on collections of images of handwritten digits and faces. Figure 2 illustrates the basic concept of coordination, as achieved by our learning rule. In the coordinated model, the global coordinate always points in the same direction along the data manifold, as defined by the composition of the transformations H and . In the model trained with maximum likelihood, the density is well captured but each local latent variable has a random orientation along the manifold. We also applied the algorithm to collections of images of handwritten digits and of faces. The representation of was an unprocessed vector of raw 8-bit grayscale pixel intensities for each image (of dimensionality 256 for the % ^ % digits and 560 for the & ^ & faces.) The MFAs had 64 local models and the global coordinates were two dimensional. After training, the coordinated MFAs had learned a smooth, continuous mapping from the plane to images of digits or of faces. This allows us both to infer a two-dimensional location given any image by computing c 2 and to generate new images from any point in the plane by computing c . (Precisely what we wanted from the magic box.) In general, both of these conditional distributions have the form of a mixture of Gaussians. Figure 3 shows the inferred global coordinates c (i.e. the means of the unimodal distributions c ) of the training points after the last iteration of training as well as examples of new images from the generative model, created by evaluating the mean of c along straight line paths in the global coordinate space. In the case of digits, it seems as though our models have captured tilt/shape and identity and represented them as the two axes of the c space; in the case of the faces the axes seem to capture pose and expression. (For the faces, the final c space was rotated by hand to align interpretable directions with the coordinate axes.) As with all EM algorithms, the coordinated MFA learning procedure is susceptible to local optima. Crucial to the success of our experiments is a good initialization, which was provided by the Locally Linear Embedding algorithm[9]. We clamped c equal to the embedding coordinate provided by LLE and to a small value and trained until convergence (typically 30-100 iterations). Then we proceeded with training using the full EM equations to update c , again until convergence (usually 5-10 more iterations). Note, however, that LLE and other embedding algorithms such as Isomap[10] are themselves unsupervised, so the overall procedure, including this initial phase, is still unsupervised. 6 Discussion Mixture models provide a simple way to approximate the density of high dimensional data that lies on or near a low dimensional manifold. However, their hidden representations do not make explicit the relationship between dissimilar data vectors. In this paper, we have shown how to learn global coordinates that can act as an encapsulating interface, so that other parts of a learning system do not need to interact with the individual components of a mixture. This should improve generalization as well as facilitate the propagation and exchange of information when these models are incorporated into a larger (perhaps Figure 3: Automatically constructed two dimensional global parameterizations of manifolds of digits and faces. Each plot shows the global coordinate space discovered by the unsupervised algorithm; points indicate the inferred means for each training item at the end of learning. The image stacks on the borders are not from the training set but are generated from the model itself and represent the mean of the predictive distribution d/eZj l at the corresponding open circles (sampled along the straight lines in the global space). The models provide both a two degree-of-freedom generator for complex images via d/eZj l as well as a pose/slant recognition system via d/e j .l . For the handwritten digits, the training set consisted of 1100 examples of the digit “2” (shown as crosses above) mixed with 1100 examples of “3”s (shown as triangles). The digits are from the NIST dataset, digitized at 16x16 pixels. For the faces, we used 2000 images of a single person with various poses and expressions taken from consecutive frames of a video digitized at 20x20 pixels. Brendan Frey kindly provided the face data. hierarchical) architecture for probabilistic reasoning. Two variants of our purely unsupervised proposal are possible. The first is to use an embedding algorithm (such as LLE or Isomap) not only as an initialization step but to provide clamped values for the global coordinates. While this supervised approach may work in practice, unsupervised coordination makes clear the objective function that is being optiFigure 4: A situation in which an un-coordinated mixture model–trained to do density estimation–cannot be “postcoordinated”. Noise has caused one of the local density models to orient orthogonal to the manifold. In globally coordinated learning, there is an additional pressure to align with neighbouring models which would force the local model to lie in the correct subspace. mized, which unifies the goals of manifold learning and density estimation. Another variant is to train an unsupervised mixture model (such as a MFA) using a traditional maximum likelihood objective function and then to “post-coordinate” its parameters by applying local reflections/rotations and translations to create global coordinates. As illustrated in figure 4, however, this two-step procedure can go awry because of noise in the original training set. When both density estimation and coordination are optimized simultaneously there is extra pressure for local experts to fit the global structure of the manifold. Our work can be viewed as a synthesis of two long lines of research in unsupervised learning. In the first are efforts at learning the global structure of nonlinear manifolds [1, 4, 9, 10]; in the second are efforts at developing probabilistic graphical models for reasoning under uncertainty[5, 6, 7]. Our work proposes to model the global coordinates on manifolds as latent variables, thus attempting to combine the representational advantages of both frameworks. It differs from embedding by providing a fully probabilistic model valid away from the training set, and from work in generative topographic mapping[2] by not requiring a uniform discretized gridding of the latent space. Moreover, by extending the usefulness of mixture models,it further develops an architecture that has already proved quite powerful and enormously popular in applications of statistical learning. Acknowledgements We thank Mike Revow for sharing his unpublished work (at the University of Toronto) on coordinating mixtures, and Zoubin Ghahramani, Peter Dayan, Jakob Verbeek and two anonymous reviewers for helpful comments and corrections. References [1] D. Beymer & T. Poggio. Image representations for visual learning. pringerScience 272 (1996). [2] C. Bishop, M. Svensen, and C. Williams. GTM: The generative topographic mapping. Neural Computation 10 (1998). [3] C. Bregler & S. Omohundro. Nonlinear image interpolation using manifold learning. Advances in Neural Information Processing Systems 7 (1995). [4] D. DeMers & G.W. Cottrell. Nonlinear dimensionality reduction. Advances in Neural Information Processing Systems 5 (1993). [5] Ghahramani, Z. and Hinton, G. The EM algorithm for mixtures of factor analyzers. University of Toronto Technical Report CRG-TR-96-1 (1996). [6] Hinton, G., Dayan, P., and Revow, M. Modeling the manifolds of images of handwritten digits. IEEE Transactions on Neural Networks 8 (1997). [7] M. Jordan, Z. Ghahramani, T. Jaakkola, and L. Saul. An introduction to variational methods for graphical models. Machine Learning 37(2) (1999). [8] N. Kambhatla and T. K. Leen. Dimension reduction by local principal component analysis. Neural Computation 9 (1997). [9] S. T. Roweis & L. K. Saul. Nonlinear dimensionality reduction by locally linear embedding. Science 290 (2000). [10] J. B. Tenenbaum, V. de Silva, and J. C. Langford. A global geometric framework for nonlinear dimensionality reduction. Science 290 (2000).
|
2001
|
100
|
1,907
|
Correlation Codes in Neuronal Populations Maoz Shamir and Haim Sompolinsky Racah Institute of Physics and Center for Neural Computation, The Hebrew University of Jerusalem, Jerusalem 91904, Israel
Abstract Population codes often rely on the tuning of the mean responses to the stimulus parameters. However, this information can be greatly suppressed by long range correlations. Here we study the efficiency of coding information in the second order statistics of the population responses. We show that the Fisher Information of this system grows linearly with the size of the system. We propose a bilinear readout model for extracting information from correlation codes, and evaluate its performance in discrimination and estimation tasks. It is shown that the main source of information in this system is the stimulus dependence of the variances of the single neuron responses. 1 Introduction Experiments in the last years have shown that in many cortical areas, the fluctuations in the responses of neurons to external stimuli are significantly correlated [1, 2, 3, 4], raising important questions regarding the computational implications of neuronal correlations. Recent theoretical studies have addressed the issue of how neuronal correlations affect the efficiency of population coding [4, 5, 6]. It is often assumed that the information about stimuli is coded mainly in the mean neuronal responses, e.g., in the tuning of the mean firing rates, and that by averaging the tuned responses across large populations, an accurate estimate can be obtained despite the significant noise in the single neuron responses. Indeed, for uncorrelated neurons the Fisher Information of the population is extensive [7]; namely, it increases linearly with the number of neurons in the population. Furthermore, it has been shown that this extensive information can be extracted by relatively simple linear readout mechanisms [7, 8]. However, it was recently shown [6] that positive correlations which vary smoothly with space may drastically suppress the information in the mean responses. In particular, the Fisher Information of the system saturates to a finite value as the system size grows. This raises questions about the computational utility of neuronal population codes. Neuronal population responses can represent information in the higher order statistics of the responses [3], not only in their means. In this work, we study the accuracy of coding information in the second order statistics. We call such schemes correlation codes. Specifically, we assume that the neuronal responses obey multivariate Gaussian statistics governed by a stimulus-dependent correlation matrix. We ask whether the Fisher Information of such a system is extensive even in the presence of strong correlations in the neuronal noise. Secondly, we inquire how information in the second order statistics can be efficiently extracted. 2 Fisher Information of a Correlation Code Our model consists of a system of neurons that code a 2D angle , . Their stochastic response is given by a vector of activities
where
is the activity of the
-th neuron in the presence of a stimulus , and is distributed according to a multivariate Gaussian distribution "!$#&% ')( (+* ,.-0/ 21 (3* 04 (1) Here
is the mean activity of the
-th neuron and its dependence on is usually referred to as the tuning curve of the neuron; / is the correlation matrix; and is a normalization constant. Here we shall limit ourselves to the case of multiplicative modulation of the correlations. Specifically we use 5
76 8 9
09 6 :<; 5
=6 (2) ; 5
76 ; 5 ?>
( > 6 @
76)A ( @
=6 !B#%DC ( >
( > 6 E F (3) 9
8 9 ?>
( :G !$#&% CGHBIJ K>
( LNM F (4) where and E are the correlation strength and correlation length respectively; L defines the tuning width of the correlations; and >
denotes the angle at which the variance of the
-th neuron, 9 M
, is maximal. An example is shown in Fig. 1. It is important to note that the variance adds a contribution to 5
76 which is larger than the contribution of the smooth part of the correlations. For reasons that will become clear below, we write, 5
76 :G 5PO
76 : A 5PQ
.@
=6 (5) where 5 O
=6 denotes the smooth part of the correlation matrix and 5 Q
the discontinuous diagonal part, which in the example of Eqs. (2)-(4) is 5PQ
G ( ,9 M
(6) A useful measure of the accuracy of a population code is the Fisher Information (FI). In the case of uncorrelated populations it is well known that FI increases linearly with system size [7], indicating that the accuracy of the population coding improves as the system size is increased. Furthermore, it has been shown that relatively simple, linear schemes can provide reasonable readout models for extracting the information in uncorrelated populations [8]. In the case of a correlated multivariate Gaussian distribution, FI is given as R RSUT.V2W A R:XZY0[0[ , where RSUT.V2W * .\ / : 1 * ,\ (7) R:XZY0[0[ N] ^ / 1 / 0.\_ M (8) where * \ and / \ denote derivatives of * and / with respect to , respectively. The form of these terms reveals that in general the correlations play two roles. First they control the efficiency of the information encoded in the mean activities * : (note the dependence of R SUT.V2W on 5 ). Secondly, / provides an additional source of information about the stimulus ( R&XZY0[0[ ). When the correlations are independent of the stimulus,
` 9
a bdce , it was shown [6] that positive correlations, gf , with long correlation length, E ih , −180 −120 −60 0 60 120 180 ψ [deg] C(φ,ψ) φ=−120o φ=−60o φ=0o φ=60o φ=120o Figure 1: The stimulus-dependent correlation matrix, Eqs. (2)-(4), depicted as a function of two angles, 5 K> , where > >
( and >
( . Here, i , E and L . cause the saturation of FI to a finite limit at large . This implies that in the presence of such correlations, population averaging cannot overcome the noise even in large networks. This analysis however, [6], did not take into account stimulus-dependent correlations, which is the topic of the present work. Analyzing the dependence of RNXZY,[ [ , Eq. (8), we find it useful to write R XZY0[0[ R Q A R O (9) where R Q
C 5 Q
, \ 5 Q
F M (10) is FI of an uncorrelated population with stimulus-dependent variance which equals 5 Q
, and scales linearly with ; R O R:XZY0[0[ ( R Q . Evaluating these terms for the multiplicative model, Eq. (2), we find that R O is positive, so that R R Q . Furthermore, numerical evaluation of this term shows that R O saturates at large to a small finite value, so that for large R:XZY0[ [ R Q M
> C 9 \ ?> 9 ?> F M (11) as shown in Fig. 2. We thus conclude that R XZY0[0[ increases linearly with and is equal, for large , to the FI of variance coding namely to R of an independent population in which information is encoded in their activity variances. Since in our system the information is encoded in the second order statistics of the population responses, it is obvious that linear readouts are inadequate. This raises the question of whether there are relatively simple nonlinear readout models for such systems. In the next sections we will study bilinear readouts and show that they are useful models for extracting information from correlation codes. 3 A Bilinear Readout for Discrimination Tasks In a two-interval discrimination task the system is given two sets of neuronal activities M generated by two proximal stimuli and A @ and must infer which stimulus generated which activity. The Maximum-Likelihood (ML) discrimination yields the 0 200 400 600 800 1000 0 0.2 0.4 0.6 0.8 1 N Jcorr [deg−2] 0 200 400 600 800 1000 0 0.5 1x 10 −3 N Js [deg−2] Figure 2: (a) Fisher Information, R XZY0[0[ , of the stimulus-dependent correlations, Eqs. (2)(4), as a function of the number of neurons in the system. In (b) we show the difference between the full FI and the contribution of the diagonal term, R O - as defined by Eq. (9). Here , E and L . Note the different scales in (a) and (b). probability of error given by \ , where d 1 M ` 1
M and the discriminability \ equals \ @ R (12) It has been previously shown that in the case of uncorrelated populations with mean coding, the optimal linear readouts achieves the Maximum-Likelihood discrimination performance in large N [7]. In order to isolate the properties of correlation coding we will assume that no information is coded in the average firing rates of the neurons, and take * hereafter. We suggest a bilinear readout as a simple generalization of the linear readout to correlation codes. In a discrimination task the bilinear readout makes a decision according to the sign of
=6
76
6 ( M
M 6 (13) where a A ( decision refers to A @ . Maximizing the signal-to-noise ratio of this rule, the optimal bilinear discriminator (OBD) matrix is given by
76 5 1 \
76 (14) Using the optimal weights to evaluate the discrimination error we obtain that in large the performance of the OBD saturates the ML performance, Eq. (12). Thus, since FI of this model increases linearly with the size of the system, the discriminability increases as . Since the correlation matrix / depends on the stimulus, , the OBD matrix, Eq. (14), will also be stimulus dependent. Thus, although the OBD is locally efficient, it cannot be used as such as a global efficient readout. 4 A Bilinear Readout for Angle Estimation 4.1 Optimal bilinear readout for estimation To study the global performance of bilinear readouts we investigate bilinear readouts which minimize the square error of estimating the angle averaged over the whole range of . For convenience we use complex notation for the encoded angle, and write as the estimator of `
. Let
76
=6
K26 (15) where
76 are stimulus independent complex weights. We define the optimal bilinear estimator (OBE) as the set of weights that minimizes on average the quadratic estimation error of an unbiased estimator. This error is given by G @ M (
(16) where is the Lagrange multiplier of the constraint . In general, it is impossible to find a perfectly unbiased estimator for a continuously varied stimulus, using a finite number of weights. However, in the case of angle estimation, we can employ the underlying rotational symmetry to generate such an estimator. For this we use the symmetry of the correlation matrix, Eq. (2). In this case one can show that the Lagrange multipliers have the simple form of ) `
, and the OBE weight matrix is in the form of
76 ?>
( > 6 !B#%DC
>
A > 6 F (17) where ?> G ( > and ?> ) ( ?> A d . This form of a readout matrix, Eq. (17), guarantees that the estimator will be unbiased. Using these symmetry properties, K> can be written in the following form (for even ) ?> @
A M 1 W W HBIJ ^ b ( > _ ( > " (18) Figure 3 (a) presents an example of the function K> . These numerical results (Fig. 3 (a)) also suggest that the function K> is mainly determined by a few harmonics plus a delta peak at > . Below we will use this fact to study simpler forms of bilinear readout. Further analysis of the OBE performance in the large limit yields the following asymptotic result @ : M 1 Q M 5 Q
?> `
M Q M 5 Q
?> 0 M ( c > , (19) Figure 3 (b) shows the numerical calculation of the OBE error (open circles) as a function of . The dashed line is the asymptotic behavior, given by Eq. (19). The dotted line is the Creamer-Rao bound. From the graph one can see that the estimation efficiency of this readout grows linearly with the size of the system, , but is lower than the bound. 4.2 Truncated bilinear readout Motivated by the simple structure of the optimal readout matrix observed in Fig. 3 (a), we studied a bilinear readout of the form of Eqs. (17) and (18) with K> which has a delta function peak at the origin plus a few harmonics. Restricting the number of harmonics to relatively small integers, we evaluated numerically the optimal values of the coefficients W for large systems. Surprisingly we found that for small and large , these coefficients approach a value which is independent of the specifics of the model and equals W ( B , yielding a bilinear weight matrix of the form
76 @
=6 ( W HBIJ ^ b ( K>
( > 6 _ !B#%DC
>
A > 6 F (20) Figure 4 shows the numerical results for the squared average error of this readout for several values of ! and 8 . The results of Fig. 4 show that for a given the 0 100 200 300 400 0 0.05 0.1 N ∆θ−2 [deg−2] J (b) −2 0 2 0 φ w(φ) (a) Figure 3: (a) Profile of ?> , Eq. (17), for the OBE with . (b) Numerical evaluation of one over the squared estimation error, for the optimal bilinear readout in the multiplicative modulation model (open circles). The dashed line is the asymptotic behavior, given by Eq. (19). Here @ & M M , for the optimal bilinear readout in the multiplicative modulation model. The dotted line is the FI bound. In these simulations , E and L were used. inverse square error initially increases linearly with but saturates in the limit of large . However, the saturation size O V increases rapidly with . The precise form of O V depends on the specifics of the correlation model. For the exponentially decaying correlations assumed in Eq. (2), we find O V . Figure 4 shows that for this range of , and the deviations of the inverse square error from linearity are small. Thus, in the regime O V , @ : M is given by the asymptotic behavior, Eq. (19), shown by the dashed line. We thus conclude that the OBE (with unlimited ) will generate an inverse square estimation error which increases linearly with with a coefficient given by Eq. (19), and that this value can be achieved for reasonable values of by an approximate bilinear weight matrix, of the form of Eq. (20), with small . The asymptotic result, Eq. (19), is smaller than the optimal value given by the full FI, Eq. (11), see Fig. 4 (dotted line). In fact, it is equal to the error of an independent population with a variance which equals 5 Q
and a quadratic population vector readout of the form
M
`
(21) It is important to note that in the presence of correlations, the quadratic readout of Eq. (21) is very inefficient, yielding a finite error for large as shown in Fig. 4 (line marked ‘quadratic’). 5 Discussion To understand the reason for the simple form of the approximately optimal bilinear weight matrix, Eq. (20), we rewrite Eq. (15) with of Eq. (20) as
?`
(22)
60 @
76 ( W 1 `
W 1 26 (23) 0 500 1000 1500 2000 0 0.1 0.2 0.3 0.4 0.5 N 1/ δθ2 deg−2 quadratic p=1 p=2 p=3 J Figure 4: Inverse square estimation error of the finite approximation for the OBE, Eq. (20). Solid curves from the bottom . The bottom curve is . The dashed line is the asymptotic behavior, given by Eq. (19). The FI bound is shown by the dotted line. For the simulations , E and L were used. Comparing this form with Eq. (21) it can be seen that our readout is in the form of a bilinear population vector in which the lowest Fourier modes of the response vector have been removed. Retaining only the high Fourier modes in the response profile suppresses the cross-correlations between the different components of the residual responses
because the underlying correlations have smooth spatial dependence, whose power is concentrated mostly in the low Fourier modes. On the other hand, the information contained in the variance is not removed because the variance contains a discontinuous spatial component, 5 Q
. In other words, the variance of a correlation profile which has only high Fourier modes can still preserve its slowly varying components. Thus, by projecting out the low Fourier modes of the spatial responses the spatial correlations are suppressed but the information in the response variance is retained. This interpretation of the bilinear readout implies that although all the elements of the correlation matrix depend on the stimulus, only the stimulus dependence of the diagonal elements is important. This important conclusion is borne out by our theoretical results concerning the performance of the system. As Eqs. (11) and (19) show, the asymptotic performance of both the full FI as well as that of the OBE are equivalent to those of an uncorrelated population with a stimulus dependent variance which equals 5 Q
. Although we have presented results here concerning a multiplicative model of correlations, we have studied other models of stimulus dependent correlations. These studies indicate that the above conclusions apply to a broad class of populations in which information is encoded in the second order statistics of the responses. Also, for the sake of clarity we have assumed here that the mean responses are untuned, * . Our studies have shown that adding tuned mean inputs does not modify the picture since the smoothly varying positive correlations greatly suppress the information embedded in the first order statistics. The relatively simple form of the readout Eq. (22) suggests that neuronal hardware may be able to extract efficiently information embedded in local populations of cells whose noisy responses are strongly correlated, provided that the variances of their responses are significantly tuned to the stimulus. This latter condition is not too restrictive, since tuning of variances of neuronal firing rates to stimulus and motor variables is quite common in the nervous system. Acknowledgments This work was partially supported by grants from the Israel-U.S.A. Binational Science Foundation and the Israeli Science Foundation. M.S. is supported by a scholarship from the Clore Foundation. References [1] E. Fetz, K. Yoyoma and W. Smith, Cerebral Cortex (Plenum Press, New York, 1991). [2] D. Lee, N.L. Port, W. Kruse and A.P. Georgopoulos, J. Neurosci. , 1161 (1998). [3] E.M. Maynard, N.G. Hatsopoulos, C.L. Ojakangas, B.D. Acuna, J.N. Sanes, R.A. Normann, and J.P. Donoghue, J. Neurosci. 19, 8083 (1999). [4] E. Zohary, M.N. Shadlen and W.T. Newsome, Nature , 140 (1994). [5] L.F. Abbott and P. Dayan, Neural Computation , 91 (1999). [6] H. Sompolinsky, H. Yoon, K. Kang and M. Shamir, Phys. Rev. E, , 051904 (2001); H. Yoon and H. Sompolinsky, Advances in Neural Information Processing Systems 11 (pp. 167). Kearns M.J, Solla S.A and Cohn D.A, Eds., (Cambridge, MA: MIT Press, 1999). [7] S. Seung and H. Sompolinsky, Proc. Natl. Acad. Sci. USA , 10794 (1993). [8] E. Salinas and L.F. Abbott, J. Comp. Neurosci. , 89 (1994).
|
2001
|
101
|
1,908
|
Grammatical Bigrams Mark A. Paskin Computer Science Division University of California, Berkeley Berkeley, CA 94720 paskin@cs.berkeley.edu Abstract Unsupervised learning algorithms have been derived for several statistical models of English grammar, but their computational complexity makes applying them to large data sets intractable. This paper presents a probabilistic model of English grammar that is much simpler than conventional models, but which admits an efficient EM training algorithm. The model is based upon grammatical bigrams, i.e., syntactic relationships between pairs of words. We present the results of experiments that quantify the representational adequacy of the grammatical bigram model, its ability to generalize from labelled data, and its ability to induce syntactic structure from large amounts of raw text. 1 Introduction One of the most significant challenges in learning grammars from raw text is keeping the computational complexity manageable. For example, the EM algorithm for the unsupervised training of Probabilistic Context-Free Grammars- known as the Inside-Outside algorithm- has been found in practice to be "computationally intractable for realistic problems" [1]. Unsupervised learning algorithms have been designed for other grammar models (e.g., [2, 3]). However, to the best of our knowledge, no large-scale experiments have been carried out to test the efficacy of these algorithms; the most likely reason is that their computational complexity, like that of the Inside-Outside algorithm, is impractical. One way to improve the complexity of inference and learning in statistical models is to introduce independence assumptions; however, doing so increases the model's bias. It is natural to wonder how a simpler grammar model (that can be trained efficiently from raw text) would compare with conventional models (which make fewer independence assumptions, but which must be trained from labelled data) . Such a model would be a useful tool in domains where partial accuracy is valuable and large amounts of unlabelled data are available (e.g., Information Retrieval, Information Extraction, etc.) . In this paper, we present a probabilistic model of syntax that is based upon grammatical bigrams, i.e., syntactic relationships between pairs of words. We show how this model results from introducing independence assumptions into more conventhe quick brown fox jumps over the lazy dog Figure 1: An example parse; arrows are drawn from head words to their dependents. The root word is jumps; brown is a predependent (adjunct) of fox; dog is a postdependent (complement) of over. tional models; as a result, grammatical bigram models can be trained efficiently from raw text using an O(n3 ) EM algorithm. We present the results of experiments that quantify the representational adequacy of the grammatical bigram model, its ability to generalize from labelled data, and its ability to induce syntactic structure from large amounts of raw text. 2 The Grammatical Bigram Model We first provide a brief introduction to the Dependency Grammar formalism used by the grammatical bigram model; then, we present the probability model and relate it to conventional models; finally, we sketch the EM algorithm for training the model. Details regarding the parsing and learning algorithms can be found in a companion technical report [4]. Dependency Grammar Formalism. 1 The primary unit of syntactic structure in dependency grammars is the dependency relationship, or link- a binary relation between a pair of words in the sentence. In each link, one word is designated the head, and the other is its dependent. (Typically, different types of dependency are distinguished, e.g, subject, complement, adjunct, etc.; in our simple model, no such distinction is made.) Dependents that precede their head are called pre dependents, and dependents that follow their heads are called postdependents. A dependency parse consists of a set of links that, when viewed as a directed graph over word tokens, form an ordered tree. This implies three important properties: 1. Every word except one (the root) is dependent to exactly one head. 2. The links are acyclic; no word is, through a sequence of links, dependent to itself. 3. When drawn as a graph above the sentence, no two dependency relations cross-a property known as projectivity or planarity. The planarity constraint ensures that a head word and its (direct or indirect) dependents form a contiguous subsequence of the sentence; this sequence is the head word's constituent. See Figure 1 for an example dependency parse. In order to formalize our dependency grammar model, we will view sentences as sequences of word tokens drawn from some set of word types. Let V = {tl' t2, ... , t M } be our vocabulary of M word types. A sentence with n words is therefore represented as a sequence S = (Wl, W2 , ... ,wn ), where each word token Wi is a variable that ranges over V. For 1 :S i ,j :S n, we use the notation (i,j) E L to express that Wj is a dependent of Wi in the parse L. IThe Dependency Grammar formalism described here (which is the same used in [5, 6]) is impoverished compared to the sophisticated models used in Linguistics; refer to [7] for a comprehensive treatment of English syntax in a dependency framework. Because it simplifies the structure of our model, we will make the following three assumptions about Sand L (without loss of generality): (1) the first word WI of S is a special symbol ROOT E V; (2) the root of L is WI; and, (3) WI has only one dependent. These assumptions are merely syntactic sugar: they allow us to treat all words in the true sentence (i.e., (W2, ... ,Wn )) as dependent to one word. (The true root of the sentence is the sole child of WI.) Probability Model. A probabilistic dependency grammar is a probability distribution P(S, L) where S = (WI,W2, .. . ,wn ) is a sentence, L is a parse of S, and the words W2, ... ,Wn are random variables ranging over V. Of course, S and L exist in high dimensional spaces; therefore, tractable representations of this distribution make use of independence assumptions. Conventional probabilistic dependency grammar models make use of what may be called the head word hypothesis: that a head word is the sole (or primary) determinant of how its constituent combines with other constituents. The head word hypothesis constitutes an independence assumption; it implies that the distribution can be safely factored into a product over constituents: n P(S,L) = II P((Wj: (i,j) E L) is the dependent sequencelwi is the head) i=1 For example, the probability of a particular sequence can be governed by a fixed set of probabilistic phrase-structure rules, as in [6]; alternatively, the predependent and postdependent subsequences can be modeled separately by Markov chains that are specific to the head word, as in [8]. Consider a much stronger independence assumption: that all the dependents of a head word are independent of one another and their relative order. This is clearly an approximation; in general, there will be strong correlations between the dependents of a head word. More importantly, this assumption prevents the model from representing important argument structure constraints. (For example: many words require dependents, e.g. prepositions; some verbs can have optional objects, whereas others require or forbid them.) However, this assumption relieves the parser of having to maintain internal state for each constituent it constructs, and therefore reduces the computational complexity of parsing and learning. We can express this independence assumption in the following way: first, we forego modeling the length of the sentence, n, since in parsing applications it is always known; then, we expand P(S, Lin) into P(S I L)P(L I n) and choose P(L I n) as uniform; finally, we select P(S I L) II P( Wj is a [pre/post]dependent I Wi is the head) (i,j)EL This distribution factors into a product of terms over syntactically related word pairs; therefore, we call this model the "grammatical bigram" model. The parameters of the model are <"(xy P(predependent is ty I head is tx ) 6. "(~ P(postdependent is ty I head is tx ) We can make the parameterization explicit by introducing the indicator variable wi, whose value is 1 if Wi = tx and a otherwise. Then we can express P(S I L) as P(S I L) (i,j)EL x=1 y=1 j<i (i,j)EL x=1 y=1 i<j Parsing. Parsing a sentence S consists of computing L* L:, argmaxP(L I S,n) = argmaxP(L, Sin) = argmaxP(S I L) L L L Yuret has shown that there are exponentially many parses of a sentence with n words [9], so exhaustive search for L * is intractable. Fortunately, our grammar model falls into the class of "Bilexical Grammars" , for which efficient parsing algorithms have been developed. Our parsing algorithm (described in the tech report [4]) is derived from Eisner's span-based chart-parsing algorithm [5], and can find L* in O(n3 ) time. Learning. Suppose we have a labelled data set where Sk = (Wl,k, W2,k,·· · , Wnk,k) and Lk is a parse over Sk. The maximum likelihood values for our parameters given the training data are where the indicator variable et is equal to 1 if (i,j) E Lk and 0 otherwise. As one would expect, the maximum-likelihood value of ,;; (resp. ,~ ) is simply the fraction of tx's predependents (resp. postdependents) that were ty. In the unsupervised acquisition problem, our data set has no parses; our approach is to treat the Lk as hidden variables and to employ the EM algorithm to learn (locally) optimal values of the parameters ,. As we have shown above, the et are sufficient statistics for our model; the companion tech report [4] gives an adaptation of the Inside-Outside algorithm which computes their conditional expectation in O(n3 ) time. This algorithm effectively examines every possible parse of every sentence in the training set and calculates the expected number of times each pair of words was related syntactically. 3 Evaluation This section presents three experiments that attempt to quantify the representational adequacy and learnability of grammatical bigram models. Corpora. Our experiments make use of two corpora; one is labelled with parses, and the other is not. The labelled corpus was generated automatically from the phrase-structure trees in the Wall Street Journal portion of the Penn Treebank-III [10].2 The resultant corpus, which we call C, consists of 49,207 sentences (1,037,374 word tokens). This corpus is split into two pieces: 90% of the sentences comprise corpus Ctrain (44,286 sentences, 934,659 word tokens), and the remaining 10% comprise Ctest (4,921 sentences, 102,715 word tokens). The unlabelled corpus consists of the 1987- 1992 Wall Street Journal articles in the TREC Text Research Collection Volumes 1 and 2. These articles were segmented on sentence boundaries using the technique of [11], and the sentences were postprocessed to have a format similar to corpus C. The resultant corpus consists of 3,347,516 sentences (66,777,856 word tokens). We will call this corpus U. 2This involved selecting a head word for each constituent, for which the head-word extraction heuristics described in [6] were employed. Additionally, punctuation was removed, all words were down-cased, and all numbers were mapped to a special <#> symbol. The model's vocabulary is the same for all experiments; it consists of the 10,000 most frequent word types in corpus U; this vocabulary covers 94.0% of word instances in corpus U and 93.9% of word instances in corpus L. Words encountered during testing and training that are outside the vocabulary are mapped to the <unk> type. Performance metric. The performance metric we report is the link precision of the grammatical bigram model: the fraction of links hypothesized by the model that are present in the test corpus Ltest. (In a scenario where the model is not required to output a complete parse, e.g., a shallow parsing task, we could similarly define a notion of link recall; but in our current setting, these metrics are identical.) Link precision is measured without regard for link orientation; this amounts to ignoring the model's choice of root, since this choice induces a directionality on all of the edges. Experiments. We report on the results of three experiments: I. Retention. This experiment represents a best-case scenario: the model is trained on corpora Ltrain and Ltest and then tested on Ltest. The model's link precision in this setting is 80.6%. II. Generalization. In this experiment, we measure the model's ability to generalize from labelled data. The model is trained on Ltrain and then tested on Ltest. The model's link precision in this setting is 61.8%. III. Induction. In this experiment, we measure the model's ability to induce grammatical structure from unlabelled data. The model is trained on U and then tested on Ltest . The model's link precision in this setting is 39.7%. Analysis. The results of Experiment I give some measure of the grammatical bigram model's representational adequacy. A model that memorizes every parse would perform perfectly in this setting, but the grammatical bigram model is only able to recover four out of every five links. To see why, we can examine an example parse. Figure 2 shows how the models trained in Experiments I, II, and III parse the same test sentence. In the top parse, syndrome is incorrectly selected as a postdependent of the first on token rather than the second. This error can be attributed directly to the grammatical bigram independence assumption: because argument structure is not modeled, there is no reason to prefer the correct parse, in which both on tokens have a single dependent, over the chosen parse, in which the first has two dependents and the second has none.3 Experiment II measures the generalization ability of the grammatical bigram model; in this setting, the model can recover three out of every five links. To see why the performance drops so drastically, we again turn to an example parse: the middle parse in Figure 2. Because the forces -+ on link was never observed in the training data, served has been made the head of both on tokens; ironically, this corrects the error made in the top parse because the planarity constraint rules out the incorrect link from the first on token to syndrome. Another error in the middle parse is a failure to select several as a predependent of forces; this error also arises because the combination never occurs in the training data. Thus, we can attribute this drop in performance to sparseness in the training data. We can compare the grammatical bigram model's parsing performance with the results reported by Eisner [8]. In that investigation, several different probability models are ascribed to the simple dependency grammar described above and 3 Although the model's parse of acquired immune deficiency syndrome agrees with the labelled corpus, this particular parse reflects a failure of the head-word extraction heuristics; acquired and immune should be predependents of deficiency, and deficiency should be a predependent of syndrome. r 1. 843 '.88' 3.2",-d fir hr-. n~ 14 . 383 I. <root> she has also served on several task forces on acquired immune deficiency syndrome 1. 528 r 9 . 630 1 '. 803 r,m 1.358 1 12 . 527 fir hn fO ~8~\ A , 14 . 264 II. <root> she has also served on several task forces on acquired immune deficiency syndrome 1. 990 0.913 4 . 124 - 1.709 (~ 13 . 585 k 0 .14 9 tI 1--' f' ~ ( III. <root> she has also served on several task forces on acquired immune deficiency syndrome Figure 2: The same test sentence, parsed by the models trained in each of the three experiments. Links are labelled with -log2 IXY I I:~1 IXY, the mutual information of the linked words; dotted edges are default attachments. are compared on a task similar to Experiment 11.4 Eisner reports that the bestperforming dependency grammar model (Model D) achieves a (direction-sensitive) link precision of 90.0%, and the Collins parser [6] achieves a (direction-sensitive) link precision of 92.6%. The superior performance of these models can be attributed to two factors: first, they include sophisticated models of argument structure; and second, they both make use of part-of-speech taggers, and can "back-off" to non-lexical distributions when statistics are not available. Finally, Experiment III shows that when trained on unlabelled data, the grammatical bigram model is able to recover two out of every five links. This performance is rather poor, and is only slightly better than chance; a model that chooses parses uniformly at random achieves 31.3% precision on L\est . To get an intuition for why this performance is so poor, we can examine the last parse, which was induced from unlabelled data. Because Wall Street Journal articles often report corporate news, the frequent co-occurrence of has -+ acquired has led to a parse consistent with the interpretation that the subject she suffers from AIDS, rather than serving on a task force to study it. We also see that a flat parse structure has been selected for acquired immune deficiency syndrome; this is because while this particular noun phrase occurs in the training data, its constituent nouns do not occur independently with any frequency, and so their relative co-occurrence frequencies cannot be assessed. 4 Discussion Future work. As one would expect, our experiments indicate that the parsing performance of the grammatical bigram model is not as good as that of state-ofthe-art parsers; however, its performance in Experiment II suggests that it may be useful in domains where partial accuracy is valuable and large amounts of unlabelled data are available. However, to realize that potential, the model must be improved so that its performance in Experiment III is closer to that of Experiment II. To that end, we can see two obvious avenues of improvement. The first involves increasing the model's capacity for generalization and preventing overfitting. The 4The labelled corpus used in that investigation is also based upon a transformed version of Treebank-III, but the head-word extraction heuristics were slightly different, and sentences with conjunctions were completely eliminated. However, the setup is sufficiently similar that we think the comparison we draw is informative. model presented in this paper is sensitive only to pairwise relationships to words; however, it could make good use of the fact that words can have similar syntactic behavior. We are currently investigating whether word clustering techniques can improve performance in supervised and unsupervised learning. Another way to improve the model is to directly address the primary source of parsing error: the lack of argument structure modeling. We are also investigating approximation algorithms that reintroduce argument structure constraints without making the computational complexity unmanageable. Related work. A recent proposal by Yuret presents a "lexical attraction" model with similarities to the grammatical bigram model [9]; however, unlike the present proposal, that model is trained using a heuristic algorithm. The grammatical bigram model also bears resemblance to several proposals to extend finite-state methods to model long-distance dependencies (e.g., [12, 13]), although these models are not based upon an underlying theory of syntax. References [1] K. Lari and S. J. Young. The estimation of stochastic context-free grammars using the Inside-Outside algorithm. Computer Speech and Language, 4:35- 56, 1990. [2] John Lafferty, Daniel Sleator, and Davy Temperley. Grammatical trigrams: A probabilistic model of link grammar. In Proceedings of the AAAI Conference on Probabilistic Approaches to Natural Language, October 1992. [3] Yves Schabes. Stochastic lexicalized tree-adjoining grammars. In Proceedings of the Fourteenth International Conference on Computational Linguistics, pages 426-432, Nantes, France, 1992. [4] Mark A. Paskin. Cubic-time parsing and learning algorithms for grammatical bigram models. Technical Report CSD-01-1148, University of California, Berkeley, 2001. [5] Jason Eisner. Bilexical grammars and their cubic-time parsing algorithms. In Harry Bunt and Anton Nijholt, editors, Advances in Probabilistic and Other Parsing Technologies, chapter 1. Kluwer Academic Publishers, October 2000. [6] Michael Collins. Head-driven Statistical Models for Natural Language Parsing. PhD thesis, University of Pennsylvania, Philadelphia, Pennsylvania, 1999. [7] Richard A. Hudson. English Word Grammar. B. Blackwell, Oxford, UK, 1990. [8] Jason M. Eisner. An empirical comparison of probability models for dependency grammars. Technical Report ICRS-96-11, CIS Department, University of Pennsylvania, 220 S. 33rd St. Philadelphia, PA 19104- 6389, 1996. [9] Deniz Yuret. Discovery of Linguistic Relations Using Lexical Attraction. PhD thesis, Massachusetts Institute of Technology, May 1998. [10] M. Marcus, B. Santorini, and M. Marcinkiewicz. Building a large annotated corpus of english: The penn treebank. Computational Linguistics, 19:313- 330, 1993. [11] Jeffrey C. Reynar and Adwait Ratnaparkhi. A maximum entropy approach to identifying sentence boundaries. In Proceedings of the Fifth Conference on Applied Natural Language Processing, Washington, D.C., March 31 - April 3 1997. [12] S. Della Pietra, V. Della Pietra, J. Gillett, J. Lafferty, H. Printz, and L. Ures. Inference and estimation of a long-range trigram model. In Proceedings of the Second International Colloquium on Grammatical Inference and Applications, number 862 in Lecture Notes in Artificial Intelligence, pages 78- 92. Springer-Verlag, 1994. [13] Ronald Rosenfeld. Adaptive Statistical Language Modeling: A Maximum Entropy Approach. PhD thesis, Carnegie Mellon University, 1994.
|
2001
|
102
|
1,909
|
The Steering Approach for Multi-Criteria Reinforcement Learning Shie Mannor and Nahum Shimkin Department of Electrical Engineering Technion, Haifa 32000, Israel {shie,shimkin}@{tx,ee}.technion.ac.il Abstract We consider the problem of learning to attain multiple goals in a dynamic environment, which is initially unknown. In addition, the environment may contain arbitrarily varying elements related to actions of other agents or to non-stationary moves of Nature. This problem is modelled as a stochastic (Markov) game between the learning agent and an arbitrary player, with a vector-valued reward function. The objective of the learning agent is to have its long-term average reward vector belong to a given target set. We devise an algorithm for achieving this task, which is based on the theory of approachability for stochastic games. This algorithm combines, in an appropriate way, a finite set of standard, scalar-reward learning algorithms. Sufficient conditions are given for the convergence of the learning algorithm to a general target set. The specialization of these results to the single-controller Markov decision problem are discussed as well. 1 Introduction This paper considers an on-line learning problem for Markov decision processes with vector-valued rewards. Each entry of the reward vector represents a scalar reward (or cost) function which is of interest. Focusing on the long-term average reward, we assume that the desired performance is specified through a given target set, to which the average reward vector should eventually belong. Accordingly, the specified goal of the decision maker is to ensure that the average reward vector will converge to the target set. Following terminology from game theory, we refer to such convergence of the reward vector as approaching the target set. A distinctive feature of our problem formulation is the possible incorporation of arbitrarily varying elements of the environment, which may account for the influence of other agents or non-stationary moves of Nature. These are collectively modelled as a second agent, whose actions may affect both the state transition and the obtained rewards. This agent is free to choose its actions according to any control policy, and no prior assumptions are made regarding its policy. This problem formulation is derived from the so-called theory of approachability that was introduced in [3] in the context of repeated matrix games with vector payoffs. Using a geometric viewpoint, it characterizes the sets in the reward space that a player can guarantee for himself for any possible policy of the other player, and provides appropriate policies for approaching these sets. Approachability theory has been extended to stochastic (Markov) games in [14], and the relevant results are briefly reviewed in Section 2. In this paper we add the learning aspect, and consider the problem of learning such approaching policies on-line, using Reinforcement Learning (RL) or similar algorithms. Approaching policies are generally required to be non-stationary. Their construction relies on a geometric viewpoint, whereby the average reward vector is “steered” in the direction of the target set by the use of direction-dependent (and possibly stationary) control policies. To motivate the steering viewpoint, consider the following one dimensional example of an automatic temperature controlling agent. The measured property is the temperature which should be in some prescribed range [T, T], the agent may activate a cooler or a heater at will. An obvious algorithm that achieves the prescribed temperature range is – when the average temperature is higher than T choose a “policy” that reduces it, namely activate the cooler; and if the average temperature is lower than T use the heater. See Figure 1(a) for an illustration. Note that this algorithm is robust and requires little knowledge about the characteristics of the processes, as would be required by a procedure that tunes the heater or cooler for continuous operation. A learning algorithm needs only determine which element to use at each of the two extreme regions. a b T T Temperature Heating policy Cooling policy Temperature Humidity Target Figure 1: (a) The single dimensional temperature example. If the temperature is higher than T the control is to cool, and if the temperature is lower than T the control is to heat. (b) The two dimensional temperature-humidity example. The learning directions are denoted by arrows, note that an infinite number of directions are to be considered. Consider next a more complex multi-objective version of this controlling agent. The controller’s objective is as before to have the temperature in a certain range. One can add other parameters such as the average humidity, frequency of switching between policies, average energy consumption and so on. This problem is naturally characterized as a multi-objective problem, in which the objective of the controller is to have the average reward in some target set. (Note that in this example, the temperature itself is apparently the object of interest rather than its long-term average. However, we can reformulate the temperature requirement as an average reward objective by measuring the fraction of times that the temperature is outside the target range, and require this fraction to be zero. For the purpose of illustration we shall proceed here with the original formulation). For example, suppose that the controller is also interested in the humidity. For the controlled environment of, say, a greenhouse, the allowed level of humidity depends on the average temperature. An illustrative target set is shown in Figure 1(b). A steering policy for the controller is not as simple anymore. In place of the two directions (left/right) of the one-dimensional case, we now face a continuum of possible directions, each associated with a possibly different steering policy. For the purpose of the proposed learning algorithm we shall require to consider only a finite number of steering policies. We will show that this can always be done, with negligible effect on the attainable performance. The analytical basis for this work relies on three elements: stochastic game models, which capture the Markovian system dynamics while allowing arbitrary variation in some elements of the environment; the theory of approachability for vector-valued dynamic games, which provides the basis for the steering approach; and RL algorithms for (scalar) average reward problems. For the sake of brevity, we do not detail the mathematical models and proofs and concentrate on concepts. Reinforcement Learning (RL) has emerged in the last decade as a unifying discipline for learning and adaptive control. Comprehensive overviews may be found in [2, 7]. RL for average reward Markov Decision Processes (MDPs) was suggested in [13, 10] and later analyzed in [1]. Several methods exist for average reward RL, including Q-learning [1] the E3 algorithm [8], actor-critic schemes [2] and more. The paper is organized as follows: In Section 2 we describe the stochastic game setup, recall approachability theory, and mention a key theorem that allows to consider only a finite number of directions for approaching a set. Section 3 describes the proposed multi-criteria RL algorithm and outlines its convergence proof. We also briefly discuss learning in multi-criteria single controller environments, as this case is a special case of the more general game model. An illustrative example is briefly described in Section 4 and concluding remarks are drawn in Section 5. 2 Multi-Criteria Stochastic Games In this section we present the multi-criteria stochastic game model. We recall some known results from approachability theory for stochastic games with vector-valued reward, and state a key theorem which decomposes the problem of approaching a target set into a finite number of scalar control problems. We consider a two-person average reward stochastic game model, with a vector-valued reward function. We refer to the players as P1 (the learning agent) and P2 (the arbitrary adversary). The game is defined by: the state space S; the sets of actions for P1 and P2, respectively, in each state s, A and B; the state transition kernel, P = (P(s′|s, a, b)); a vector-valued reward function m : S × A × B →IRk. The reward itself is allowed to be random, in which case it is assumed to have a bounded second moment. At each time epoch n ≥0, both players observe the current state sn, and then P1 and P2 simultaneously choose actions an and bn, respectively. As a result P1 receives the reward vector mn = m(sn, an, bn) and the next state is determined according to the transition probability P(·|sn, an, bn). More generally, we allow the actual reward mn to be random, in which case m(sn, an, bn) denotes its mean and a bounded second moment is assumed. We further assume that both players observe the previous rewards and actions (however, in some of the learning algorithms below, the assumption that P1 observes P2’s action may be relaxed). A policy π ∈Π for P1 is a mapping which assigns to each possible observed history a mixed action in ∆(A), namely a probability vector over P1’s action set A. A policy σ ∈Σ for P2 is defined similarly. A policy of either player is called stationary if the mixed action it prescribes depends only on the current state sn. Let ˆmn denote the average reward by time n: ˆmn △= 1 n Pn−1 t=0 mt. The following recurrence assumption will be imposed. Let state s∗denote a specific reference state to which a return is guaranteed. We define the hitting time of state s∗as: τ △= min{n > 0 : sn = s∗}. Assumption 1 (Recurrence) There exist a state s∗∈S and a finite constant N such that Es πσ(τ 2) < N for all π ∈Π, σ ∈Σ and s ∈S, where Es πσ is the expectation operator when starting from state s0 = s and using policies π and σ for P1 and P2, respectively. If the game is finite then this assumption is satisfied if state s∗is accessible from all other states under any pair of stationary deterministic policies [14]. We note that the recurrence assumption may be relaxed in a similar manner to [11]. Let u be a unit vector in the reward space IRk. We often consider the projected game in direction u as the zero-sum stochastic game with same dynamic as above, and scalar rewards rn := mn · u. Here “·” stands for the standard inner product in IRk. Denote this game by Γs(u), where s is the initial state. The scalar stochastic game Γs(u), has a value, denoted vΓs(u), if vΓs(u) = sup π inf σ lim inf n→∞Es πσ( ˆmn · u) = inf σ sup π lim sup n→∞Es πσ( ˆmn · u) . For finite games, the value exists [12]. Furthermore, under Assumption 1 the value is independent of the initial state and can be achieved in stationary policies [6]. We henceforth simply write vΓ(u) for this value. We next consider the task of approaching a given target set in the reward space, and introduce approaching policies for the case where the game parameters are fully known to P1. Let T ⊂IRk denote the target set. In the following, d is the Euclidean distance in IRk, and P s π,σ is the probability measure induced by the policies π and σ, with initial state s. Definition 2.1 The set T ⊂IRk is approachable (from initial state s) if there exists a Tapproaching policy π∗of P1 such that d( ˆmn, T) →0 P s π∗,σ-a.s., for every σ ∈Σ at a uniform rate over Σ. The policy π∗in that definition will be called an approaching policy for P1. A set is approachable if it is approachable from all states. Noting that approaching a set and its closure are the same, we shall henceforth suppose that the set T is closed. We recall the basic results from [14] regarding approachability for known stochastic games, which generalize Blackwell’s conditions for repeated matrix games. Let φ(π, σ) △= Es∗ π,σ(Pτ−1 t=0 mt) Es∗ π,σ(τ) (1) denote the average per-cycle reward vector, which is the expected total reward over the cycle that starts and ends in the reference state, divided by the expected duration of that cycle. For any x ̸∈T, denote by Cx a closest point in T to x, and let ux be the unit vector in the direction of Cx −x, which points from x to the goal set T, see Figure 2 for an illustration. Theorem 2.1 [14] Let Assumption 1 hold. Assume that for every point x ̸∈T there exists a policy π(x) such that: (φ(π(x), σ) −Cx) · ux ≥0 , ∀σ ∈Σ . (2) Then T is approachable by P1. An approaching policy is: If sn = s∗and ˆmn ̸∈T, play π( ˆmn) until the next visit to state s∗; otherwise, play arbitrarily. Figure 2: An illustration of approachability. π(x) brings P1 to the other side of the hyperplane perpendicular to the segment between Cx and x. Geometrically, the condition in (2) means that P1 can ensure, irrespectively of P2’s policy, that the average per-cycle reward will be on the other side (relative to x) of the hyperplane which is perpendicular to the line segment that points from x to Cx. We shall refer to the direction ux as the steering direction from point x, and to the policy π(x) as the steering policy from x. The approaching policy uses the following rule: between successive visits to the reference state, a fixed (possibly stationary) policy is used. When in the reference state, the current average reward vector ˆmn is inspected. If this vector is not in T, then the steering policy that satisfies (2) with x = ˆmn is selected for the next cycle. Consequently, the average reward is “steered” towards T, and eventually converges to it. Recalling the definition of the projected game in direction u and its value vΓ(u), the condition in (2) may be equivalently stated as vΓ(ux) ≥Cx · ux. Furthermore, the policy π(x) can always be chosen as the stationary policy which is optimal for P1 in the game Γ(ux). In particular, the steering policy π(x) needs to depend only on the corresponding steering direction ux. It can be shown that for convex target sets, the condition of the last theorem turns out to be both sufficient and necessary. Standard approachability results, as outlined above, require to consider an infinite number of steering directions whenever the reward in non-scalar. The corresponding set of steering policies may turn out to be infinite as well. For the purpose of our learning scheme, we shall require an approaching policy which relies on a finite set of steering directions and policies. The following results show that this can indeed be done, possibly requiring to slightly expand the target set. In the following, let M be an upper bound on the magnitude of the expected one-stage reward vector, so that ||m(s, a, b)|| < M for all (s, a, b) (|| · || denote the Euclidean norm). We say that a set of vectors (u1, . . . , uJ) is an ϵ-cover of the unit ball if for every vector in the unit ball u there exists a vector ui such that ||ui −u|| ≤ϵ . Theorem 2.2 Let Assumption 1 hold and suppose that the target set T ⊂IRk satisfies condition (2). Fix ϵ > 0. Let {u1, . . . , uJ} be an ϵ/M cover of the unit ball. Suppose that πi is an optimal strategy in the scalar game Γ(ui) (1 ≤i ≤J). Then the following policy approaches T ϵ, the ϵexpansion of T: If sn = s∗and ˆmn ̸∈T ϵ, then choose j so that u ˆmn is closest to uj (in Euclidean norm) and play πj until the next visit to state s∗; otherwise, play arbitrarily. Proof: (Outline) The basic observation is that if two directions, u and ui are close, then vΓ(u) and vΓ(ui) are close. Consequently, by playing a strategy which is optimal in Γ(ui) results in a play which is almost optimal in Γ(u). Finally we can apply Blackwell’s Theorem (2.1) for the expansion of T, by noticing that a “good enough” strategy is played in every direction. Remark: It follows immediately from the last theorem that the set T itself (rather than its ϵexpansion) is approachable with a finite number of steering directions if T −ϵ, the ϵ shrinkage of T, satisfies (2). Equivalently, T is required to satisfy (2) with the 0 on the right-hand-side replaced by ϵ > 0. 3 The Multi-Criteria Reinforcement Learning Algorithm In this section we introduce and prove the convergence of the MCRL (Multi-Criteria Reinforcement Learning) algorithm. We consider the controlled Markov model of Section 2, but here we assume that P1, the learning agent, does not know the model parameters, namely the state transition probabilities and reward functions. A policy of P1 that does not rely on knowledge of these parameters will be referred to as a learning policy. P1’s task is to approach a given target set T, namely to ensure convergence of the average reward vector to this set irrespective of P2’s actions. The proposed learning algorithm relies on the construction of the previous section of approaching policies with a finite number of steering directions. The main idea is to apply a (scalar) learning algorithm for each of the projected games Γ(uj) corresponding to these directions. Recall that each such game is a standard zero-sum stochastic game with average reward. The required learning algorithm for game Γ(u) should secure an average reward that is not less than the value vΓ(u) of that game. Consider a zero-sum stochastic game, with reward function r(s, a, b), average reward ˆrn and value v. Assume for simplicity that the initial state is fixed. We say that a learning policy π of P1 is ϵ-optimal in this game if, for any policy σ of P2, the average reward satisfies lim inf n→∞ˆrn ≥v −ϵ Pπσ a.s. , where Pπσ is the probability measure induced by the algorithm π, P2’s policy σ and the game dynamics. Note that P1 may be unable to learn a min-max policy as P2 may play an inferior policy and refrain from playing certain actions, thereby keeping some parts of the game unobserved. Remark: RL for average reward zero-sum stochastic games can be devised in a similar manner to average reward Markov decision processes. For example, a Q-learning based algorithm which combines the ideas of [9] with those of [1] can be devised. An additional assumption that is needed for the analysis is that all actions of both players are used infinitely often. A different type of a scalar algorithm that overcomes this problem is [4]. The algorithm there is similar to the E3 algorithm [8] which is based on explicit exploration-exploitation tradeoffand estimation of the game reward and transition structure. We now describe the MCRL algorithm that nearly approaches any target set T that satisfies (2). The parameters of the algorithm are ϵ and M. ϵ is the approximation level and M is a known bound on the norm of the expected reward per step. The goal of the algorithm is to approach T ϵ, the ϵ expansion of T. There are J learning algorithms that are run in parallel, denoted by π1, . . . πJ. The MCRL is described in Figure 3 and is given here as a meta-algorithm (the scalar RL algorithms πi are not specified). When arriving to s∗, the decision maker checks if the average reward vector is outside the set T ϵ. In that case, he switches to an appropriate policy that is intended to “steer” the average reward vector towards the target set. The steering policy (πj) is chosen according to closest direction (uj) to the actual direction needed according to the problem geometry. Recall that each πj is actually a learning policy with respect to a scalar reward function. In general, when πj is not played, its learning pauses and the process history during that time is ignored. Note however that some “off-policy” algorithms (such as Q-learning) can learn the optimal policy even while playing a different policy. In that case a more efficient version of the MCRL is suggested, in which learning is performed by all learning policies πj continuously and concurrently. 0. Let u1, . . . uJ be an ϵ/2M cover of the unit ball. Initialize J different ϵ/2-optimal scalar algorithms, π1, . . . , πJ. 1. If s0 ̸= s∗play arbitrarily until sn = s∗. 2. (sn = s∗) If ˆmn ∈T ϵ goto step 1. Else let i = arg min1≤i≤J ||ui −u ˆmn||2. 3. While sn ̸= s∗play according to πi, the reward πi receives is ui · mn. 4. When sn = s∗goto step 2. Figure 3: The MCRL algorithm Theorem 3.1 Suppose that Assumption 1 holds and the MCRL algorithm is used with ϵ-optimal scalar learning algorithms. If the target set T satisfies (2), then T ϵ is approached using MCRL. Proof: (Outline) If a direction is played infinitely often, then eventually the learned strategy in this direction is nearly optimal. If a direction is not played infinitely often it has a negligible effect on the long term average reward vector. Since the learning algorithms are nearly optimal, then any policy πj that is played infinitely often, eventually attains a (scalar) average reward of vΓ(uj)−ϵ/2. One can apply Theorem 2.2 for the set T ϵ/2 to verify that the overall policy is an approaching policy for the target set. Note that for convex target sets the algorithm is consistent in the sense that if the set is approachable then the algorithm attains it. Remark: Multi-criteria Markov Decision Process (MDP) models may be regarded as a special case of the stochastic game model that was considered so far, with P2 eliminated from the problem. The MCRL meta-algorithm of the previous section remains the same for MDPs. Its constituent scalar learning algorithms are now learning algorithms for the optimal polices in average-reward MDPs. These are generally simpler than for the game problem. Examples of optimal or ϵ-optimal algorithms are Q-Learning with persistent exploration [2], Actor-critic schemes [2], an appropriate version of the E3 algorithm [8] and others. In the absence of an adversary, the problem of approaching a set becomes much simpler. Moreover, it can be shown that if a set is approachable then it may be approached using a stationary (possibly randomized) policy. Indeed, any point in feasible set of state-action frequencies may be achieved by such a stationary policy [5]. Thus, alternative learning schemes may be applicable to this problem. Another observation is that all steering policies learned and used within the MCRL may now be deterministic stationary policies, which simplifies the implementation of this algorithm. 4 Example Recall the humidity-temperature example from the introduction. Suppose that the system is modelled in such a way that P1 chooses a temperature-humidity curve. Then Nature (modelled as P2) chooses the exact location on the temperature-humidity curve. In Figure 4(a) we show three different temperature-humidity curves, that can be determined by P1 (each defined by a certain strategy of P1 - f0, f1, f2). We implemented MCRL algorithm with nine directions. In each direction a version of Littman’s Q-learning ([9]), adapted for average cost games, was used. A sample path of the average reward generated by the MCRL algorithm is shown in Figure 4(b). The sample path started at ’S’ and finished at ’E’. For this specific run, an even smaller number of directions would have sufficed (up and right). It can be seen that the learning algorithm pushes towards the target set so that the path is mostly on the edge of the target set. Note that in this example a small number of directions was quite enough for approaching the target set. a b 0.5 1 1.5 2 0.5 1 1.5 2 Temperature Humidity f0 f1 f2 Problem dynamics for different strategies 0.8 1 1.2 1.4 1.6 1.8 2 0.8 1 1.2 1.4 1.6 1.8 2 S E Temperature Humidity A sample path of average reward Figure 4: (a) Greenhouse problem dynamics. (b) A sample path from ’S’ to ’E’ 5 Conclusion We have presented a learning algorithm that approaches a prescribed target set in multi-dimensional performance space, provided this set satisfies a certain sufficient condition. Our approach essentially relies on the theory of approachability for stochastic games, which is based on the idea of steering the average reward vector towards the target set. We provided a key result stating that a set can be approached to a given precision using only a finite number of steering policies, which may be learned on-line. An interesting observation regarding the proposed learning algorithm is that the learned optimal polices in each direction are essentially independent of the target set T. Thus, the target set need not be fixed in advance and may be modified on-line without requiring a new learning process. This may be especially useful for constrained MDPs. Of further interest is the question of reduction of the number of steering directions used in the algorithm. In some cases, especially when the requirements embodied by the target set T are not stringent, this number may be quite small compared to the worst-case estimate used above. A possible refinement of the algorithm is to eliminate directions that are not required. The scaling of he algorithm with the dimension of the reward space is exponential. The problem is that as the dimension increases, exponentially many directions are needed to cover the unit ball. While in general this is necessary, it might happen that considerably less directions are needed. Conditions and algorithms that use much less than exponential number of directions are under current study. Acknowledgement This research was supported by the fund for the promotion of research at the Technion. References [1] J. Abounadi, D. Bertsekas, and V. Borkar. Learning algorithms for markov decision processes with average cost. LIDS-P 2434, Lab. for Info. and Decision Systems, MIT, October 1998. [2] A.G. Barto and R.S. Sutton. Reinforcement Learning. MIT Press, 1998. [3] D. Blackwell. An analog of the minimax theorem for vector payoffs. Pacific J. Math., 6(1):1–8, 1956. [4] R.I. Brafman and M. Tennenholtz. A near optimal polynomial time algorithm for learning in certain classes of stochastic games. Artificial Intelligence, 121(1-2):31–47, April 2000. [5] C. Derman. Finite state Markovian decision processes. Academic Press, 1970. [6] J. Filar and K. Vrieze. Competitive Markov Decision Processes. Springer Verlag, 1996. [7] L.P. Kaelbling, M. Littman, and A.W. Moore. Reinforcement learning - a survey. Journal of Artificial Intelligence Research, (4):237–285, May 1996. [8] M. Kearns and S. Singh. Near-optimal reinforcement learning in polynomial time. In Proc. of the 15th Int. Conf. on Machine Learning, pages 260–268. Morgan Kaufmann, 1998. [9] M.L. Littman. Markov games as a framework for multi-agent reinforcement learning. In Morgan Kaufman, editor, Eleventh International Conference on Machine Learning, pages 157–163, 1994. [10] S. Mahadevan. Average reward reinforcement learning: Foundations, algorithms, and empirical results. Machine Learning, 22(1):159–196, 1996. [11] S. Mannor and N. Shimkin. The empirical bayes envelope approach to regret minimization in stochastic games. Technical report EE- 1262, Faculty of Electrical Engineering, Technion, Israel, October 2000. [12] J.F. Mertens and A. Neyman. Stochastic games. International Journal of Game Theory, 10(2):53–66, 1981. [13] A. Schwartz. A reinforcement learning method for maximizing undiscounted rewards. In Proceedings of the Tenth International Conference on Machine Learning, pages 298–305. Morgan Kaufmann, 1993. [14] N. Shimkin and A. Shwartz. Guaranteed performance regions in markovian systems with competing decision makers. IEEE Trans. on Automatic Control, 38(1):84–95, January 1993.
|
2001
|
103
|
1,910
|
Stabilizing Value Function with the Xin Wang Department of Computer Science Oregon State University Corvallis, OR, 97331 wangxi@cs. orst.edu Thomas G Dietterich Department of Computer Science Oregon State University Corvallis, OR, 97331 tgd@cs. orst. edu Abstract We address the problem of non-convergence of online reinforcement learning algorithms (e.g., Q learning and SARSA(A)) by adopting an incremental-batch approach that separates the exploration process from the function fitting process. Our BFBP (Batch Fit to Best Paths) algorithm alternates between an exploration phase (during which trajectories are generated to try to find fragments of the optimal policy) and a function fitting phase (during which a function approximator is fit to the best known paths from start states to terminal states). An advantage of this approach is that batch value-function fitting is a global process, which allows it to address the tradeoffs in function approximation that cannot be handled by local, online algorithms. This approach was pioneered by Boyan and Moore with their GROWSUPPORT and ROUT algorithms. We show how to improve upon their work by applying a better exploration process and by enriching the function fitting procedure to incorporate Bellman error and advantage error measures into the objective function. The results show improved performance on several benchmark problems. 1 Introduction Function approximation is essential for applying value-function-based reinforcement learning (RL) algorithms to solve large Markov decision problems (MDPs). However, online RL algorithms such as SARSA(A) have been shown experimentally to have difficulty converging when applied with function approximators. Theoretical analysis has not been able to prove convergence, even in the case-of linear function approximators. (See Gordon (2001), however, for a non-divergence result.) The heart of the problem is that the approximate values of different states (e.g., 81 and 82) are coupled through the parameters of the function approximator. The optimal policy at state 81 may require increasing a parameter, while the optimal policy at state 82 may require decreasing it. As a result, algorithms based on local parameter updates tend to oscillate or even to diverge. To avoid this problem, a more global approach is called for-an approach that can consider Sl and S2 simultaneously and find a solution that works well in both states. One approach is to formulate the reinforcement learning problem as a global search through a space of parameterized policies as in the policy gradient algorithms (Williams, 1992; Sutton, McAllester, Singh, & Mansour, 2000; Konda & Tsitsiklis, 2000; Baxter & Bartlett, 2000). This avoids the oscillation problem, but the resulting algorithms are slow and only converge to local optima. We pursue an alternative approach that formulates the function approximation problem as a global supervised learning problem. This approach, pioneered by Boyan and Moore in their GROWSUPPORT (1995) and ROUT (1996) algorithms, separates the reinforcement learning problem into two subproblems: the exploration problem (finding a good partial value function) and the representation problem (representing and generalizing that value function). These algorithms alternate between two phases. During the exploration phase, a support set of points is constructed whose optimal values are known within some tolerance. In the function fitting phase, a function approximator V is fit to the support set. In this paper, we describe two ways of improving upon GROWSUPPORT and ROUT. First, we replace the support set with the set of states that lie along the best paths found during exploration. Second, we employ a combined error function that includes terms for the supervised error, the Bellman error, and the advantage error (Baird, 1995) into the function fitting process. The resulting BFBP (Batch Fit to Best Paths) method gives significantly better performance on resource-constrained scheduling problems as well as on the mountain car toy benchmark problem. 2 GrowSupport, ROUT, and BFBP Consider a deterministic, episodic MDP. Let s' == a(s) denote the state s' that results from performing a in s and r(a, s) denote the one-step reward. Both GROWSUPPORT and ROUT build a support set S == {(Si' V(Si))} of states whose optimal values V (s) are known with reasonable accuracy. Both algorithms initialize S with a set of terminal states (with V(s) == 0). In each iteration, a function approximator V is fit to S to minimize :Ei[V(Si) - V(Si)]2. Then, an exploration process attempts to identify new points to include in S. In GROWSUPPORT, a sample of points X is initially drawn from the state space. In each iteration, after fitting V, GROWSUPPORT computes a new estimate V(s) for each state sEX according to V(s) == maxa r(s, a) + V(a(s)), where V(a(s)) is computed by executing the greedy policy with respect to V starting in a(s). If V(a(s)) is within c of V(a(s)), for all actions a, then (s, V(s)) is added to S. ROUT employs a different procedure suitable for stochastic MDPs. Let P(s'ls, a) be the probability that action a in state s results in state s' and R(s'ls, a) be the expected one-step reward. During the exploration phase, ROUT generates a trajectory from the start state to a terminal state and then searches for a state s along that trajectory such that (i) V(s) is not a good approximation to the backedup value V(s) == maxa :Est P(s'ls, a)[R(s'ls, a) + V(s')], and (ii) for every state s along a set of rollout trajectories starting at s', V(s) is within c of the backed-up value maxa :Est P(s'ls, a)[R(s'ls, a) +V(s')]. If such a state is found, then (s, V(s)) is added to S. Both GROWSUPPORT and ROUT rely on the function approximator to generalize well at the boundaries of the support set. A new state s can only be added to S if V has generalized to all of s's successor states. H this occurs consistently, then eventually the support set will expand to include all of the starting states of the MDP, at which point a satisfactory policy has been found. However, if this "boundary generalization" does not occur, then no new points will be added to S, and both GROWSUPPORT and ROUT. terminate without a solution. Unfortunately, most regression methods have high bias and variance near the boundaries of their training data, so failures of boundary generalization are common. These observations led us to develop the BFBP algorithm. In BFBP, the exploration process maintains a data structure S that stores the best known path from the start state to a terminal state and a "tree" of one-step departures from this best path (Le., states that can be reached by executing an action in some state on the best path). At each state Si E S, the data structure stores the action at executed in that state (to reach the next state in the path), the one-step reward ri, and the estimated value V(Si). S also stores each action a_ that causes a departure from the best path along with the resulting state S_, reward r_ and estimated value V(s_). We will denote by B the subset of S that constitutes the best path. The estimated values V are computed as folloV1S. For states S'i E B, V(Si) is computed 'by summing the immediate rewards rj for all steps j 2: i along B. For the one-step departure states s_, V(s_) is computed from an exploration trial in which the greedy policy was followed starting in state s_. fuitially, S is empty, so a random trajectory is generated from the start state So to a terminal state, and it becomes the initial best known path. fu subsequent iterations, a state Si E B is chosen at random, and an action a' 1= at is chosen and executed to produce state s' and reward r'. Then the greedy policy (with respect to the current V) is executed until a terminal state is reached. The rewards along this new path are summed to produce V(s'). If V(s') +r' > V(Si), then the best path is revised as follows. The new best action in state Si becomes a l with estimated value V(s') +r'. This improved value is then propagated backwards to update the V estimates for all ancestor states in B. The old best action at in state Si becomes an inferior action a_ with result state s_. Finally all descendants of s_ along the old best path are deleted. This method of investigating one-step departures from the best path is inspired by Harvey and Ginsberg's (1995) limited discrepancy search (LDS) algorithm. In each exploration phase, K one-step departure paths are explored. After the exploration phase, the value function approximation V is recomputed with the goal of minimizing a combined error function: J(V) == AsL (V(s) - V(S))2 + AbL (V(s) - [r(s, a*) + V(a*(s))])2 + sES sEB Aa L L ([r(s,a-) +V(a-(s))] - [r(s,a*) + V(a*(s))]):. The three terms of this objective function are referred to as the supervised, Bellman, and advantage terms. Their relative importance is controlled by the coefficients As, Ab' and Au. The supervised term is the usual squared error between the V(s) values stored in S and the fitted values V(s). The Bellman term is the squared error between the fitted value and the backed-up value of the next state on the best path. And the advantage term penalizes any case where the backed-up value of an inferior action a_ is larger than the backed-up value of the best action a*. The notation (u)+ == u if u 2: 0 and 0 otherwise. TheoreIll 1 Let M be a deterministic MDP such that (aJ there are only a finite number of starting states, (bJ there are only· a finite set of actions executable in each state, and (c) all policies reach a terminal state. Then BFBP applied to M converges. Proof: The LDS exploration process is monotonic, since the data structure S is only updated if a new best path is found. The conditions of the theorem imply that there are only a finite number of possible paths that·can be explored from the starting states to the terminal states. Hence, the data structure S will eventually converge. Consequently, the value function V fit to S will also converge. Q.E.D. The theorem requires that the MDP contain no cycles. There are cycles in our jobshop scheduling problems, but we eliminate them by remembering all states visited along the current trajectory and barring any action that would return to a previously visited state. Note also that the theorem applies to MDPs with continuous state spaces provided the action space and the start states are finite. Unfortunately, BFBP does not necessarily converge to an optimal policy. This is because LDS exploration can get stuck in a local optimum such that all one step departures using the V-greedy policy produce trajectories that do not improve over the current best path. Hence, although BFBP resembles policy iteration, it does not have the same optimality guarantees,. because policy iteration evaluates the current greedy policy in all states in the state space. Theoretically, we could prove convergence to the optimal policy under modified conditions. If we replace LDS exploration with €-greedy exploration, then exploration will converge to the optimal paths with probability 1. When trained on those paths, if the function approximator fits a sufficiently accurate V, then BFBS will converge optimally. hI our experiments, however, we have found that €-greedy gives no improvement over LDS, whereas LDS exploration provides more complete coverage of one-step departures from the current best path, and these are used in J(V). 3 Experimental Evaluation We have studied five domains: Grid World and Puddle World (Boyan & Moore, 1995), Mountain Car (Sutton, 1996), and resource-constrained scheduling problems ART-1 and ART-2 (Zhang & Dietterich, 1995). For the first three domains, following Boyan and Moore, we compare BFBP with GROWSUPPORT. For the final domain, it is difficult to draw a sample of states X from the state space to initialize GROWSUPPORT. Hence, we compare against ROUT instead. As mentioned above, we detected and removed cycles from the scheduling domain (since ROUT requires this). We retained the cycles in the first three problems. On mountain car, we also applied SARSA(A) with the CMAC function approximator developed by Sutton (1996). We experimented with two function approximators: regression trees (RT) and locally-weighted linear regression (LWLR). Our regression trees employ linear separating planes at the internal nodes and linear surfaces at the leaf nodes. The trees are grown top-down in the usual fashion. To determine the splitting plane at a node, we choose a state Si at random from S, choose one of its inferior children S_, and construct the plane that is the perpendicular bisector of these two points. The splitting plane is evaluated by fitting the resulting child nodes to the data (as leaf nodes) and computing the value of J (V). A number C of parent-child pairs (Si' S - ) are generated and evaluated, and the best one is retained to be the splitting plane. This process is then repeated recursively until a node contains fewer than M data points~ The linear surfaces at the leaves are trained by gradient descent to minimize J(V). The gradient descent terminates after 100 steps or earlier if J becomes very small. In our experiments, we tried all combinations of the following parameters and report the best results: (a) 11 learning rates (from 0.00001 to 0.1), (b) M == 1, Table 1: Comparison of results on three toy domains. Problem Domain Algorithms Optimal Policyfj Best Policy Length Grid World GROWSUPPORT Yes 39 BFBP Yes 39 Puddle World G ROWSUPPORT Yes 39 BFBP Yes 39 Mountain Car SARSA(A) No 103 GROWSUPPORT No 93 BFBP Yes 88 Table 2: Results of ROUT and BFBP on scheduling problem ART-I-TRNOO I Performance I ROUT (RT) I ROUT (LWLR) I BFBP (RT) I Best policy explored I 1.75 I 1.55 I 1.50 I Best final learned policy I 1.8625 I 1.8125 I 1.55 10, 20, 50, 100, 1000, (c) C == 5, 10, 20, 50, 100, and (d) K == 50, 100, 150, 200. For locally-weighted linear regression, we replicated the methods of B'oyan and Moore. To compute V(s), a linear regression is performed using all points Si E S weighted by their distance to S according to the kernel exp -(Ilsi - sII 2/a2 ). We experimented with all combinations of the following parameters and report the best results: (a) 29 values (from 0.01 to 1000.0) of the tolerance E that controls the addition of new points to S, and (b) 39 values (from 0.01 to 1000.0) for a. We execute ROUT and GROWSUPPORT to termination. We execute BFBP for 100 iterations, but it converges much earlier: 36 iterations for the grid world, 3 for puddle world, 10 for mountain car, and 5 for the job-shop scheduling problems. Table 1 compares the results of the algorithms on the toy domains with parameters for each method tuned to give the best results and with As == 1 and Ab == Aa == o. In all cases, BFBP matches or beats the other methods. In Mountain Car, in particular, we were pleased that BFBP discovered the optimal policy very quickly. Table 2 compares the results of ROUT and BFBP on job-shop scheduling problem TRNOO from problem set ART-1 (again with As == 1 and Ab == Aa == 0). For ROUT, results with both LWLR and RT are shown. LWLR gives better results for ROUT. We conjecture that this is because ROUT needs a value function approximator that is conservative near the boundary of the training data, whereas BFBP does not. We report both the best policy found during the iterations and the final policy at convergence. Figure 1 plots the r,esults for ROUT (LWLR) against BFBP (RT) for eight additional scheduling problems from ART-I. The figure ofmerit is RDF, which is a normalized measure of schedule length (small values are preferred). BFBP's learned policy out-performs ROUT's in every case. The experiments above all employed only the supervised term in the error function J. These experiments demonstrate that LDS exploration gives better training sets than the support set methods of GROWSUPPORT and ROUT. Now we turn to the question of whether the Bellman and advantage terms can provide improved results. For the grid world and puddle world tasks, the supervised term already gives optimal performance, so we focus on the mountain car and job-shop scheduling problems. Table 3 summarizes the results for BFBP on the mountain car problem. All parameter settings, except for the last, succeed in finding the optimal policy. To get best policy explored + y=xbest finalleamed policy x Xx + x x 2.4 2.2 G:' Q es <l) 1.8 § § ~ 1.6 0.0... ff P=l 1.4 1.2 1 1 1.2 + + 1.4 1.6 1.8 ROUT performance (RDF) 2.2 2.4 Figure 1: Performance of Rout vs. BFBP over 8 job shop scheduling problems Table 3: Fraction of parameter settings that give optimal performance for BFBP on the mountain car problem .As .Ab .Aa # settings As Ab Aa # settings 0.0 0.0 1.0 2/1311 0.0 1.0 0.0 1/1297 1.0 0.0 0.0 52/1280 1.0 0.0 10.0 184/1291 1.0 10.0 0.0 163/1295 1.0 0.0 100.0 133/1286 1.0 100.0 0.0 4/939 1.0 1000.0 0.0 0/1299 a sense of the robustness of the method, we also report the fraction of parameter settings that gave the optimal policy. The number of parameter settings tested (the denominator) should be the same for all combinations of A values. Nonetheless, for reasons unrelated to the parameter settings, some combinations failed to be executed by our distributed process scheduler. The best settings combine As == 1 with either Ab == 10 or Aa == 10. However, if we employ either the Bellman or the advantage term alone, the results are poor. Hence, it appears that the supervised term is very important for good performance, but that the advantage and Bellman terms can improve performance substantially .and reduce the sensitivity of BFBP to the settings of the other parameters. Table 4 shows the performance of BFBP on ART-I-TRNOO. The best performance (at convergence) is obtained with As == Aa == 1 and Ab == O. As with mountain car, these experiments show that the supervised term is the most important, but that it gives even better results when combined with the advantage term. All of the above experiments compare performance on single problems. We also tested the ability of BFBP to generalize to similar problems following the formulation of (Zhang & Dietterich, 1995). Figure 2 compares the performance of neural networks and regression trees as function approximators for BFBP. Both were trained on job shop scheduling problem set ART-2. Twenty of the problems in ART-2 were used for training, 20 for cross-validation, and 50 for testing. Eleven different values for As, Ab' Aa and eight different values for the learning rate were tried, with the best values selected according to the cross-validation set. Figure 2 shows that BFBP is significantly better than the baseline performance (with RDF Table 4: Performance ofBFBP on ART-1-TENOO for different settings of the .A parameters. The ('perform;' column gives the best RDF in any iteratIon and the RDF at convergence. .A8 .Ab .Aa perform. .A8 .Ab .Aa perform. .A8 .Ab .Aa perform. 0 0 1 1.50/1.75 0 1 0 1.50/1.775 1 0 0 1.50/1.55 0 1 1 1.50/1.775 0 1 10 1.50/1.825 0 1 100 1.50/1.65 0 10 1 1.50/1.775 0 100 1 1.50/1.738 1 1 a 1.50/1.563 1 0 1 1.50/1.488 1 0 10 1.463/1.525 1 0 100 1.50/1.588 1 1 1 1.525/1.55 1 1 10 1.50/1.588 1 1 100 1.50/1.675 1.8 .------,----.-------,.---...,----,------, 1.75 BFBP neural net ----*---. BFBP regression tree --- -G---1.7 _________________________RP_E _ 1.65 LL 1.6 o0: ~ 1.55 ~ ~ 1.45 1.4'¥._?4<.,_. ,'"'::-..... ,c ._••_._._.-!n-._,._. . ...._._.,.._._._. ._,._. . ._., ._.._._. ,,_._. ._.,_T._..O_L__. . 1.35 30 25 20 15 LOS iteration 10 1.3 L..--__---l--__--'--__--L -'---__---'-__-----' o Figure 2: BFBP on ART-2 using neural nets and regression trees. "RDF" is a hand-coded heuristic, "TDL" is Zhang's TD(.A) neural network. as a search heuristic) and that its performance is comparable to TD(A) with neural networks (Zhang & Dietterich, 1995). Figure 3 shows that for ART-2, using parent/inferior-child pair splits gives better results than using axis-parallel splits. 4 Conclusions This paper has shown that the exploration strategies underlying GROWSUPPORT and ROUT can be improved by simply remembering and training on the best paths found between start and terminal states. Furthermore, the paper proved that the BFBP method converges for arbitrary function approximators, which is a result that has not yet been demonstrated for online methods such as SARSA(A). In addition, we have shown that the performance of our BFBP algorithm can be further improved (and made more robust) by incorporating a penalty for violations of the Bellman equation or a penalty for preferring inferior actions (an advantage error). Taken together, these results show that incremental-batch value function approximation can be a reliable, convergent method for solving deterministic reinforcement learning problems. The key to the success of the method is the ability to separate the exploration process from the function approximation process and to make the exploration process convergent. This insight should also be applicable to stochastic episodic MDPs. 1.9 ,-------.-----,------,----..,-------,-----, axis-parallel ....*... parent/inferior-child .. ··11 •••. 30 25 20 15 LDS iteration 10 1.3 L-__--l-'---__---'-L-__---L.__----I o Figure 3: Axis parallel splits versus parent/inferior-child pair splits on ART-2 Acknowledgments The authors gratefully acknowledge the support of AFOSR under contract F4962098-1-0375, and the NSF under grants IRl-9626584, I1S-0083292, 1TR-5710001197, and EIA-9818414. We thank Valentina Zubek for her careful reading of the paper. References Baird, L. C. (1995). Residual algorithms: Reinforcement learning with function approximation. In ICML-95, 30-37, San Francisco, CA. Morgan Kaufmann. Baxter, J., & Bartlett, P. L. (2000). Reinforcement learning in POMDP's via direct gradient ascent. In ICML-2000, 41-48. Morgan Kaufmann, San Francisco, CA. Boyan, J. A., & Moore, A. W. (1995). Generalization in reinforcement learning: Safely approximating the value function. In NIPS-7, 369-376. The MIT Press, Cambridge. Boyan, J. A., & Moore, A. W. (1996). Learning evaluation functions for large acyclic domains. In ICML-96, 63-70. Morgan Kaufmann, San Francisco, CA. Gordon, G. J. (2001). Reinforcement learning with function approximation converge to a region. In NIPS-13, 1040-1046. The MIT Press. Harvey, W. D., & Ginsberg, L. P. (1995). Limited discrepancy search. In IJCAI-95, 825-830. Morgan Kaufmann. Konda, V. R., & Tsitsiklis, J. N. (2000). Policy gradient methods for reinforcement learning with function approximation. In NIPS-12, 1008-1014 Cambridge, MA. MIT Press. Moll, R., Barto, A. G., Perkins, T. J., & Sutton, R. S. (1999). Learning instanceindependent value functions to enhance local search. In NIPS-ll, 1017-1023. Sutton, R. S., McAllester, D., Singh, S., & Mansour, Y. (2000). Policy gradient methods for reinforcement learning with function approximation. In NIPS-12, 1057-1063. Sutton, R. S. (1996). Generalization in reinforcement learning: Successful examples using sparse coarse coding. In NIPS-8, 1038-1044. The MIT Press, Cambridge. Williams, R. J. (1992). Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8,229. -.. Zhang, W., & Dietterich, T. G. (1995). A reinforcement learning approach to job-shop scheduling. In IJCAI-95, 1114-1120. Morgan Kaufmann, San Francisco, CA.
|
2001
|
104
|
1,911
|
Tree-based reparameterization for approximate inference on loopy graphs Martin J. Wainwright, Tommi Jaakkola, and Alan S. Will sky Department of Electrical Engineering and Computer Science Massachusetts Institute of Technology Cambridge, MA 02139 mjwain@mit.edu tommi@ai.mit.edu willsky@mit.edu Abstract We develop a tree-based reparameterization framework that provides a new conceptual view of a large class of iterative algorithms for computing approximate marginals in graphs with cycles. It includes belief propagation (BP), which can be reformulated as a very local form of reparameterization. More generally, we consider algorithms that perform exact computations over spanning trees of the full graph. On the practical side, we find that such tree reparameterization (TRP) algorithms have convergence properties superior to BP. The reparameterization perspective also provides a number of theoretical insights into approximate inference, including a new characterization of fixed points; and an invariance intrinsic to TRP /BP. These two properties enable us to analyze and bound the error between the TRP /BP approximations and the actual marginals. While our results arise naturally from the TRP perspective, most of them apply in an algorithm-independent manner to any local minimum of the Bethe free energy. Our results also have natural extensions to more structured approximations [e.g., 1, 2]. 1 Introduction Given a graphical model, one important problem is the computation of marginal distributions of variables at each node. Although highly efficient algorithms exist for this task on trees, exact solutions are prohibitively complex for more general graphs of any substantial size. This difficulty motivates the use of approximate inference algorithms, of which one of the best-known and most widely studied is belief propagation [3], also known as the sum-product algorithm in coding [e.g., 4]. Recent work has yielded some insight into belief propagation (BP). Several researchers [e.g., 5, 6] have analyzed the single loop case, where BP can be reformulated as a matrix powering method. For Gaussian processes on arbitrary graphs, two groups [7, 8] have shown that the means are exact when BP converges. For graphs corresponding to turbo codes, Richardson [9] established the existence of fixed points, and gave conditions for their stability. More recently, Yedidia et al. [1] showed that BP corresponds to constrained minimization of the Bethe free energy, and proposed extensions based on Kikuchi expansions [10]. Related extensions to BP were proposed in [2]. The paper [1] has inspired other researchers [e.g., 11, 12] to develop more sophisticated algorithms for minimizing the Bethe free energy. These advances notwithstanding, much remains to be understood about the behavior of BP. The framework of this paper provides a new conceptual view of various algorithms for approximate inference, including BP. The basic idea is to seek a reparameterization of the distribution that yields factors which correspond, either exactly or approximately, to the desired marginal distributions. If the graph is acyclic (i.e., a tree) , then there exists a unique reparameterization specified by exact marginal distributions over cliques. For a graph with cycles, we consider the idea of iteratively reparameterizing different parts of the distribution, each corresponding to an acyclic subgraph. As we will show, BP can be interpreted in exactly this manner, in which each reparameterization takes place over a pair of neighboring nodes. One of the consequences of this interpretation is a more storage-efficient "message-free" implementation of BP. More significantly, this interpretation leads to more general updates in which reparameterization is performed over arbitrary acyclic subgraphs, which we refer to as tree-based reparameterization (TRP) algorithms. At a low level, the more global TRP updates can be viewed as a tree-based schedule for message-passing. Indeed, a practical contribution of this paper is to demonstrate that TRP updates tend to have better convergence properties than local BP updates. At a more abstract level, the reparameterization perspective provides valuable conceptual insight, including a simple tree-consistency characterization of fixed points, as well as an invariance intrinsic to TRP /BP. These properties allow us to derive an exact expression for the error between the TRP /BP approximations and the actual marginals. Based on this exact expression, we derive computable bounds on the error. Most of these results, though they emerge very naturally in the TRP framework, apply in an algorithm-independent manner to any constrained local minimum of the Bethe free energy, whether obtained by TRP /BP or an alternative method [e.g., 11, 12]. More details of our work can be found in [13, 14]. 1.1 Basic notation An undirected graph Q = (V, £) consists of a set of nodes or vertices V = {l, ... ,N} that are joined by a set of edges £. Lying at each node s E V is a discrete random variable Xs E {a, ... ,m - I}. The underlying sample space X N is the set of all N vectors x = {xs I S E V} over m symbols, so that IXNI = m N . We focus on stochastic processes that are Markov with respect to Q, so that the Hammersley-Clifford theorem [ e.g., 3] guarantees that the distribution factorizes as p(x) ex: [lcEe 'l/Jc(xc) where 'l/Jc(xc) is a compatibility function depending only on the subvector Xc = {xs I SEC} of nodes in a particular clique C. Note that each individual node forms a singleton clique, so that some of the factors 'l/Jc may involve functions of each individual variable. As a consequence, if we have independent measurements Ys of Xs at some (or all) of the nodes, then Bayes' rule implies that the effect of including these measurements i.e., the transformation from the prior distribution p(x) to the conditional distribution p(x I y) is simply to modify the singleton factors. As a result, throughout this paper, we suppress explicit mention of measurements, since the problem of computing marginals for either p(x) or p(x I y) are of identical structure and complexity. The analysis of this paper is restricted to graphs with singleton ('l/Js) and pairwise ('l/Jst} cliques. However, it is straightforward to extend reparameterization to larger cliques, as in cluster variational methods [e.g., 10]. 1.2 Exact tree inference as reparameterization Algorithms for optimal inference on trees have appeared in the literature of various fields [e.g., 4, 3]. One important consequence of the junction tree representation [15] is that any exact algorithm for optimal inference on trees actually computes marginal distributions for pairs (s, t) of neighboring nodes. In doing so, it produces an alternative factorization p(x) = TI sEV Ps TI(s,t)E£ Pst/(PsPt) where Ps and Pst are the single-node and pairwise marginals respectively. This {Ps, Pst} representation can be deduced from a more general factorization result on junction trees [e.g. 15]. Thus, exact inference on trees can be viewed as computing a reparameterized factorization of the distribution p(x) that explicitly exposes the local marginal distributions. 2 Tree-based reparameterization for graphs with cycles The basic idea of a TRP algorithm is to perform successive reparameterization updates on trees embedded within the original graph. Although such updates are applicable to arbitrary acyclic substructures, here we focus on a set T 1 , ... , TL of embedded spanning trees. To describe TRP updates, let T be a pseudomarginal probability vector consisting of single-node marginals Ts(xs) for 8 E V; and pairwise joint distributions Tst (x s, Xt) for edges (s, t) E [. Aside from positivity and normalization (Lx Ts = 1; L x x Tst = 1) constraints, a given vecs s , t tor T is arbitraryl, and gives rises to a parameterization of the distribution as p(x; T) ex: TIsEV Ts TI(S,t)E£ Tst/ {(Lx. Tst)(L Xt Tst)}, where the dependence of Ts and Tst on x is omitted for notational simplicity. Ultimately, we shall seek vectors T that are consistent i.e., that belong to <C = {T I Lx. Tst = Tt \;/ (8, t) E [}. In the context of TRP, such consistent vectors represent approximations to the exact marginals of the distribution defined by the graph with cycles. We shall express TRP as a sequence of functional updates Tn I-t T n+1 , where superscript n denotes iteration number. We initialize at TO via T~t = Ii 'l/Js'I/Jt'I/Jst and T~ = Ii 'l/Js TItEN(S) [L X t 'l/Jst'I/Jt], where Ii denotes a normalization factor; and N(8) is the set of neighbors of node 8. At iteration n, we choose some spanning tree Ti(n) with edge set [i(n), and factor the distribution p(x; Tn) into a product of two terms ex: (la) ex: (lb) corresponding, respectively, to terms in the spanning tree; and residual terms over edges in [/ [i(n) removed to form Ti(n). We then perform a reparameterization update on pi(n) (x; Tn) explicitly: pi(n) (x'; Tn) for all (s,t) E [i(n) (2) x, s.t( x ~ ,x;)=(x. ,xtl with a similar update for the single-node marginals {Ts I s E V}. These marginal computations can be performed efficiently by any exact tree algorithm applied to Ti(n). Elements of T n+1 corresponding to terms in ri(n) (x; Tn) are left unchanged lIn general, T need not be the actual marginals for any distribution. (i.e., Ts~+l = Tst for all (8, t) E E /Ei(n)) . The only restriction placed on the spanning tree set T1, ... ,TL is that each edge (8, t) E E belong to at least one spanning tree. For practical reasons, it is desirable to choose a set of spanning trees that leads to rapid mixing throughout the graph. A natural choice for the spanning tree index i(n) is the cyclic ordering, in which i(n) == n(modL) + 1. 2.1 BP as local reparameterization Interestingly, BP can be reformulated in a "message-free" manner as a sequence of local rather than global reparameterization operations. This message-free version of BP directly updates approximate marginals, Ts and Tst, with initial values determined from the initial messages M~t and the original compatibility functions of the graphical model as T~ = Ii 'l/Js ITuEN(S) M~s for all 8 E V and T~t = Ii 'l/Jst'l/Js'l/Jt ITuEN(s)/t M~s ITuEN(t) /s M~t for all (8, t) E E, where Ii denotes a normalization factor. At iteration n, these quantities are updated according to the following recursions: (3a) T;'t (3b) The reparameterization form of BP decomposes the graph into a set of two-node trees (one for each edge (8, t)); performs exact inference on such tree via equation (3b); and merges the marginals from each tree via equation (3a). It can be shown by induction [see 13] that this simple reparameterization algorithm is equivalent to the message-passing version of BP. 2.2 Practical advantages of TRP updates Since a single TRP update suffices to transmit information globally throughout the graph, it might be expected to have better convergence properties than the purely local BP updates. Indeed, this has proven to be the case in various experiments that we have performed on two graphs (a single loop of 15 nodes, and a 7 x 7 grid). We find that TRP tends to converge 2 to 3 times faster than BP on average (rescaled for equivalent computational cost); more importantly, TRP will converge for many problems where BP fails [13]. Further research needs to address the optimal choice of trees (not necessarily spanning) in implementing TRP. 3 Theoretical results The TRP perspective leads to a number of theoretical insights into approximate inference, including a new characterization of fixed points, an invariance property, and error analysis. 3.1 Analysis of TRP updates Our analysis of TRP updates uses a cost function that is an approximation to the Kullback-Leibler divergence between p(x; T) and p(x; U) namely, the quantity Xs Given an arbitrary U E C, we show that successive iterates {Tn} of TRP updates satisfy the following "Pythagorean" identity: G(U ;T n) = G(U ;T n+l ) + G(T n+1; T n) (4) which can be used to show that TRP fixed points T * satisfy the necessary conditions to be local minima of G subject to the constraint T * E C. The cost function G, though distinct from the Bethe free energy [1], coincides with it on the constraint set C, thereby allowing us to establish the equivalence of TRP and BP fixed points. 3.2 Characterization of fixed points From the reparameterization perspective arises an intuitive characterization of any TRP /BP fixed point T *. Shown in Figure l(a) is a distribution on a graph with T1: T2~ T3~ T1: T2~ T3~ T~T; T4; T; T; T5: T; T ~ TtT; T2*T; T; T ~ (a) Fixed point on full graph (b) Tree consistency condition. Figure 1. Illustration of fixed point consistency condition. (a) Fixed point T * = {T;, T;t} on the full graph with cycles. (b) Illustration of consistency condition on an embedded tree. The quantities {T;, T;t } must be exact marginal probabilities for any tree embedded within the full graph. cycles, parameterized according to the fixed point T * = {Ts*t, T;}. The consistency condition implies that if edges are removed from the full graph to form a spanning tree, as shown in panel (b), then the quantities Ts*t and Ts* correspond to exact marginal distributions over the tree. This statement holds for any acyclic substructure embedded within the full graph with cycles not just the spanning trees Tl , ... ,TL used to implement TRP. Thus, algorithms such as TRP /BP attempt to reparameterize a distribution on a graph with cycles so that it is consistent with respect to each embedded tree. It is remarkable that the existence of such a parameterization (though obvious for trees) should hold for a positive distribution on an arbitrary graph. Also noteworthy is the parallel to the characterization of max-product2 fixed points obtained by Freeman and Weiss [16]. Finally, it can be shown [13, 14] that this characterization, though it emerged very naturally from the TRP perspective, applies more generally to any constrained local minimum of the Bethe free energy, whether obtained by TRP /BP, or an alternative technique [e.g., 11, 12]. 2Max-product is a related but different algorithm for computing approximate MAP assignments in graphs with cycles. 3.3 Invariance of the distribution A fundamental property of TRP updates is that they leave invariant the full distribution on the graph with cycles. This invariance follows from the decomposition of equation (1): in particular, the distribution pi(n) (x; Tn) is left invariant by reparameterization; and TRP does not change terms in ri(n) (x; Tn). As a consequence, the overall distribution remains invariant i.e., p(x; Tn) == p(x; TO) for all n. By continuity of the map T f-7 p(x; T), it follows that any fixed point T* of the algorithm also satisfies p(x; T*) == p(x; TO). This fixed point invariance is also an algorithmindependent result in particular, all constrained local minima of the Bethe free energy, regardless of how they are obtained, are invariant in this manner [13, 14]. This invariance has a number of important consequences. For example, it places severe restrictions on cases (other than trees) in which TRP /BP can be exact; see [14] for examples. In application to the linear-Gaussian problem, it leads to an elementary proof of a known result [7, 8] namely, the means must be exact if the BP updates converge. 3.4 Error analysis Lastly, we can analyze the error arising from any TRP /BP fixed point T* on an arbitrary graph. Of interest are the exact single-node marginals Ps of the original distribution p(x; TO) defined by the graph with cycles, which by invariance are equivalent to those of p(x; T*). Now the quantities Ts* have two distinct interpretations: (a) as the TRP /BP approximations to the actual single-node marginals on the full graph; and (b) as the exact marginals on any embedded tree (as in Figure 1). This implies that the approximations T; are related to the actual marginals Ps on the full graph by a relatively simple perturbation namely, removing edges from the full graph to reveal an embedded tree. From this observation, we can derive the following exact expression for the difference between the actual marginal PS;j and the TRP /BP approximation3 T;j: [{ ri(X;T*)} .J lEpi (x;T* ) Z(T*) - 1 J(xs = J) (5) where i E {1, ... ,L} is an arbitrary spanning tree index; pi and ri are defined in equation (1a) and (1b) respectively; Z(T*) is the partition function of p(x; T*); J(xs = j) is an indicator function for Xs to take the value j; and lEpi (x;T * ) denotes expectation using the distribution pi(x; T*). Unfortunately, while the tree distribution pi (x; T*) is tractable, the argument of the expectation includes all terms r i (x; T*) removed from the original graph to form spanning tree Ti. Moreover, computing the partition function Z (T*) is intractable. These difficulties motivate the development of bounds on the error. In [14], we use convexity arguments to derive a particular set of bounds on the approximation error. Such error bounds, in turn, can be used to compute upper and lower bounds on the actual marginals Ps;l. Figure 2 illustrates the TRP /BP approximation, as well as these bounds on the actual marginals for a binary process on a 3 x 3 grid under two conditions. Note that the tightness of the bounds is closely related to approximation accuracy. Although it is unlikely that these bounds will remain quantitatively useful for general problems on large graphs, they may still yield useful qualitative information. 3The notation T;;j denotes the /h element of the vector T; . 0.9 0.8 0.7 :;:::'0.6 " :5-"b.5 e "- o. 0.2 0.1 Bounds on single node marginals °1~~--~--~4~~5---6~~~~~~ Node number (a) Weak potentials 0.9 0.8 Bounds on single node marginals 4 5 6 Node number (b) Strong mixed potentials Figure 2. Behavior of bounds on 3 x 3 grid. Plotted are the actual marginals P s;l versus the TRP approximations T;'l> as well as upper and lower bounds on the actual marginals. (a) For weak potentials, TRP /BP approximation is excellent; bounds on exact marginals are tight. (b) For strong mixed potentials, approximation is poor. Bounds are looser, and for certain nodes, the TRP /BP approximation lies above the upper bounds on the actual marginal P8 ;1 . Much of the analysis of this paper -- including reparameterization, invariance, and error analysis -- can be extended [see 14] to more structured approximation algorithms [e.g., 1, 2]. Figure 3 illustrates the use of bounds in assessing when to use a more structured approximation. For strong attractive potentials on the 3 x 3 grid, the TRP /BP approximation in panel (a) is very poor, as reflected by relatively loose bounds on the actual marginals. In contrast, the Kikuchi approximation in (b) is excellent, as revealed by the tightness of the bounds. 4 Discussion The TRP framework of this paper provides a new view of approximate inference; and makes both practical and conceptual contributions. On the practical side, we find that more global TRP updates tend to have better convergence properties than local BP updates. The freedom in tree choice leads to open problems of a graphtheoretic nature: e.g., how to choose trees so as to guarantee convergence, or to optimize the rate of convergence? Among the conceptual insights provided by the reparameterization perspective are a new characterization of fixed points; an intrinsic invariance; and analysis of the approximation error. Importantly, most of these results apply to any constrained local minimum of the Bethe free energy, and have natural extensions [see 14] to more structured approximations [e.g., 1, 2]. Acknowledgments This work partially funded by ODDR&E MURI Grant DAAD19-00-1-0466; by ONR Grant N00014-00-1-0089; and by AFOSR Grant F49620-00-1-0362; MJW also supported by NSERC 1967 fellowship. References [1] J. Yedidia, W. T. Freeman, and Y. Weiss. Generalized belief propagation. In NIPS 13, pages 689- 695. MIT Press, 2001. Bounds on single node marginals Bounds on single node marginals - - - -0 - - - -0- - e - - - M M 0.8 _ 0 - - - - 0 - - - -0- - - - €l - - -;o::V " " :5-"b. :5-"b.5 £> £> a.. 0.4 a.. 0.4 0.3 0.2r r -+-:Ac-,-tu--:al----. 0.1 -+- TAP I BP - 0 · Bounds °1~~==~~-4~~5~~-~-~~ Node number (a) TRP /BP 0.3 :~ II =-:= ~~r~~lured approx. 1 ~ r l=-e~B=o= un=ds~==~~~~-~-~-~ °1 4 5 Node number (b) Kikuchi Figure 3. When to use a more structured approximation? (a) For strong attractive potentials on the 3 x 3 grid, BP approximation is poor, as reflected by loose bounds on the actual marginal. (b) Kikuchi approximation [1] for same problem is excellent; corresponding bounds are tight. [2] T. P. Minka. A family of algorithms for approximate Bayesian inference. PhD thesis, MIT Media Lab, 2001. [3] J. Pearl. Probabilistic reasoning in intelligent systems. Morgan Kaufman, San Mateo, 1988. [4] F. Kschischang and B. Frey. Iterative decoding of compound codes by probability propagation in graphical models. IEEE Sel. Areas Comm., 16(2):219- 230, February 1998. [5] J. B. Anderson and S. M. Hladnik. Tailbiting map decoders. IEEE Sel. Areas Comm., 16:297- 302, February 1998. [6] Y. Weiss. Correctness of local probability propagation in graphical models with loops. Neural Computation, 12:1-41, 2000. [7] Y. Weiss and W. T. Freeman. Correctness of belief propagation in Gaussian graphical models of arbitrary topology. In NIPS 12, pages 673- 679. MIT Press, 2000. [8] P. Rusmevichientong and B. Van Roy. An analysis of turbo decoding with Gaussian densities. In NIPS 12, pages 575- 581. MIT Press, 2000. [9] T. Richardson. The geometry of turbo-decoding dynamics. IEEE Trans. Info. Theory, 46(1):9- 23, January 2000. [10] R. Kikuchi. The theory of cooperative phenomena. Physical Review, 81:988- 1003, 1951. [11] M. Welling and Y. Teh. Belief optimization: A stable alternative to loopy belief propagation. In Uncertainty in Artificial Intelligence, July 2001. [12] A. Yuille. A double-loop algorithm to minimize the Bethe and Kikuchi free energies. Neural Computation, To appear, 2001. [13] M. J . Wainwright, T. Jaakkola, and A. S. Willsky. Tree-based reparameterization for approximate estimation on graphs with cycles. LIDS Tech. report P-2510: available at http://ssg.rnit.edu/group/rnjyain/rnjyain.shtrnl, May 2001. [14] M. Wainwright. Stochastic processes on graphs with cycles: geometric and variational approaches. PhD thesis, MIT, Laboratory for Information and Decision Systems, January 2002. [15] S. L. Lauritzen. Graphical models. Oxford University Press, Oxford, 1996. [16] W. Freeman and Y. Weiss. On the optimality of solutions of the max-product belief propagation algorithm in arbitrary graphs. IEEE Trans. Info. Theory, 47:736- 744, 2001.
|
2001
|
105
|
1,912
|
Cobot: A Social Reinforcement Learning Agent Charles Lee Isbell, Jr. Christian R. Shelton AT&T Labs-Research Stanford University Michael Kearns Satinder Singh Peter Stone University of Pennsylvania Syntek Capital AT&T Labs-Research Abstract We report on the use of reinforcement learning with Cobot, a software agent residing in the well-known online community LambdaMOO. Our initial work on Cobot (Isbell et al.2000) provided him with the ability to collect social statistics and report them to users. Here we describe an application of RL allowing Cobot to take proactive actions in this complex social environment, and adapt behavior from multiple sources of human reward. After 5 months of training, and 3171 reward and punishment events from 254 different LambdaMOO users, Cobot learned nontrivial preferences for a number of users, modifing his behavior based on his current state. Here we describe LambdaMOO and the state and action spaces of Cobot, and report the statistical results of the learning experiment. 1 Introduction While most applications of reinforcement learning (RL) to date have been to problems of control, game playing and optimization (Sutton and Barto1998), there has been a recent handful of applicationsto human-computer interaction. Such applications present a number of interesting challenges to RL methodology(such as data sparsity and inevitable violations of the Markov property). These previous studies focus on systems that encounter human users one at a time, such as spoken dialogue systems (Singh et al.2000). In this paper, we report on an RL-based agent for LambdaMOO, a complex, open-ended, multi-user chat environment, populated by a community of human users with rich and often enduring social relationships. Our long-term goal is to build an agent who can learn to perform useful, interesting and entertaining actions in LambdaMOO on the basis of user feedback. While this is a deliberately ambitious and underspecified goal, we describe here our implementation, the empirical experiences of our agent so far, and some of the lessons we have learned about this challenging domain. In previous work (Isbell et al.2000), we developed the software agent Cobot, who interacted in various ways with LambdaMOO users. Cobot had two primary functions. First, Cobot gathered “social statistics” (e.g. how frequently and in what ways users interacted with one another), and provided summaries of these statistics as a service. Second, Cobot had rudimentary chatting abilities based on the application of information retrieval methods to large documents. The original Cobot was entirely reactive , in that he never initiated interaction with human users, but would only respond to their actions. As we documented in our earlier paper, Cobot proved tremendously popular with LambdaMOO users, setting the stage for our current efforts. We modified Cobot to allow him to take certain actions (such as proposing conversation topics, introducing users, or engaging in common word play routines) under his own initiative. The hope is to build an agent that will eventually take unprompted actions that are meaningful, useful or amusing to users. Rather than hand-code complex rules specifying Here we mean “responding only to human-invoked interaction”, rather than “non-deliberative”. Characters in LambdaMOO have gender. Cobot’s description to users indicates that he is male. when each action is appropriate (rules that would be inaccurate and quickly become stale), we wanted Cobot to learn the individual and communal preferences of users. Thus, we provided a mechanism for users to reward or punish Cobot, and programmed Cobot to use RL algorithms to alter his behavior on the basis of this feedback. The application of RL (or any machine learning methodology) to such an environment presents a number of interesting domain-specific challenges, including: Choice of an appropriate state space. To learn how to act in a social environment such as LambdaMOO, Cobot must represent the salient features. These should include social information such as which users are present, how experienced they are in LambdaMOO, how frequently they interact with one another, and so on. Multiple reward sources. Cobot lives in an environment with multiple, often conflicting sources of reward from different human users. How to integrate these sources reasonably is a nontrivial empirical question. Inconsistency and drift of user rewards and desires. Individual users may be inconsistent in the rewards they provide (even when they implicitly have a fixed set of preferences), and their preferences may change over time (for example, due to becoming bored or irritated with an action). Even when their rewards are consistent, there can be great temporal variation in their reward pattern. Variability in user understanding. There is great variation in users’ understanding of Cobot’s functionality, and the effects of their rewards and punishments. Data sparsity. Training data is scarce for many reasons, including user fickleness, and the need to prevent Cobot from generating too much spam in the environment. Irreproducibility of experiments. As LambdaMOO is a globally distributed community of human users, it is virtually impossible to replicate experiments taking place there. We do not have any simple answers (nor do we believe that simple answers exist), but here we provide a case study of our choices and findings. Our primary findings are: Inappropriateness of average reward. We found that the average reward that Cobot received over time, the standard measure of successfor RL experiments, is an inadequate and perhaps even inappropriate metric of performance in the LambdaMOO domain. Reasons include that user preferences are not stationary, but drift as users become habituated or bored with Cobot’s behavior; and the tendency for satisfied users to stop providing Cobot with any feedback, positive or negative. Despite the inadequacy of average reward, we are still able to establish several measures by which Cobot’s RL succeeds, discussed below. A small set of dedicated “parents”. While many users provided only a moderate or small amount of RL training (rewards and punishments) to Cobot, a handful of users did invest significant time in training him. Some parents have strong opinions. While many of the users that trained Cobot did not exhibit clear preferences for any of his actions over the others, some users clearly and consistently rewarded and punished particular actions over the others. Cobot learns matching policies. For those users who exhibited clear preferences through their rewards and punishments, Cobot successfully learned corresponding policies of behavior. Cobot responds to his dedicated parents. For those users who invested the most training time in Cobot, the observed distribution of his actions is significantly altered by their presence. Some preferences depend on state. Although some users for whom we have sufficient data seem to have preferences that do not depend upon the social state features we constructed for the RL, others do in fact appear to change their preferences depending upon prevailing social conditions. The outline for the rest of the paper is as follows. In Section 2, we give brief background on LambdaMOO. In Section 3, we describe our earlier (non-RL) work on Cobot. Section 4 provides some brief background on RL. In Sections 5, 6 and 7 we describe our implementation of Cobot’s RL action space, reward mechanisms and state features, respectively. Our primary findings are presented in Section 8, and Section 9 offers conclusions. 2 LambdaMOO LambdaMOO, founded in 1990 by Pavel Curtis at Xerox PARC, is the oldest continuously operating MUD, a class of online worlds with roots in text-based multiplayer role-playing games. MUDs (multi-user dungeons) differ from most chat and gaming systems in their use of a persistent representation of a virtual world, often created by the participants, who are represented as characters of their own choosing. LambdaMOO appears as a series of interconnected rooms, populated by users and objects who may move between them. Each room provides a shared chat channel, and typically has an elaborate text description that imbues it with its own “look and feel.” In addition to speech, users express themselves via a large collection of verbs, allowing a rich set of simulated actions, and the expression of emotional states: (1) Buster is overwhelmed by all these deadlines. (2) Buster begins to slowly tear his hair out, one strand at a time. (3) HFh comforts Buster. (4) HFh [to Buster]: Remember, the mighty oak was once a nut like you. (5) Buster [to HFh]: Right, but his personal growth was assured. Thanks anyway, though. (6) Buster feels better now. Lines (1) and (2) are initiated by verb commands by user Buster, expressing his emotional state, while lines (3) and (4) are examples of verbs and speech acts, respectively, by HFh. Lines (5) and (6) are speech and verb acts by Buster. Though there are many standard verbs, such as the use of the verb comfort in line (3) above, the variety is essentially unlimited, as players have the ability to create their own verbs. The rooms and objects in LambdaMOO are created by users themselves, who devise descriptions, and control access by other users. Users can also create objects with verbs that can be invoked by other players. As last count, the database contains 118,154 objects, including 4836 active user accounts. LambdaMOO’s long existence and its user-created nature combine to give it one of the strongest senses of virtual community in the on-line world. Many users have interacted extensively with each other over many years, and users are widely acknowledged for their contribution of interesting objects. LambdaMOO is an attractive environment for experiments in AI (Foner1997; Mauldin1994), including learning. The population is generally curious and technically savvy, and users are interested in automated objects meant to display some form of intelligence. 3 Cobot Cobot is a software agent residing in LambdaMOO. Like a human user, he connects via telnet, and from the point of view of the LambdaMOO server, is a user with all the rights and responsibilities implied. Once actually connected, Cobot wanders into the Living Room, where he spends most of his time. The Living Room is a central public place, frequented both by many regulars, and by users new to LambdaMOO. There are several permanent objects in the Living Room, including a couch with various features and a cuckoo clock. The Living Room usually has between five and twenty users, and is perpetually busy. Over a year, Cobot noted over 2.5 million separate events (about one event every eleven seconds) Previously, we implemented a variety of functionality on Cobot centering around gathering and reporting social statistics. Cobot notes who takes what actions, and on whom. Cobot can answer queries about these statistics, and describe the similarities and differences between users. He also has a rudimentary chatting ability based on the application of information retrieval methods to large documents. He can also search the web to answer specific questions posed to him. A more complete description of Cobot’s abilities, and his early experiences as a social agent in LambdaMOO, can be found in (Isbell et al.2000). Our focus here is to make Cobot proactive—i.e., let him take actions under his own initiative—in a way that is useful, interesting, or pleasing to LambdaMOO users. It is impossible to program rules anticipating when any given action is appropriate in such a complex and dynamic environment, so we applied reinforcement learning to adapt directly from user feedback. We emphasize that Cobot’s original reactive functionality remained on during the RL experiment. Cobot’s persona is largely due to this original functionality, and we felt it was most interesting, and even necessary, to add RL work in this context. Null Action Choose to remain silent for this time period. Topic Change (4) Introduce a conversationaltopic. Cobot declares that he wants to discuss sports or politics, or he utters a sentence from either the sports section or political section of the Boston Globe. Roll Call (2) Initiate a “roll call,” a common word play routine in LambdaMOO. For example, someone who is tired of Monica Lewinsky may emote “TIRED OF LEWINSKY ROLL CALL.” Sympathetic users agree with the roll call. Cobot takes a recent utterance, and extracts either a single noun, or a verb phrase. Social Commentary Make a comment describing the current social state of the Living Room, such as “It sure is quiet” or “Everyone here is friendly.” These statements are based on Cobot’s statistics from recent activity. Several different utterances possible, but they are treated as a single action for RL purposes. Introductions Introduce two users who have not yet interacted in front of Cobot. Table 1: The 9 RL actions available to Cobot. 4 RL Background In RL, problems of decision-making by agents interacting with uncertain environments are usually modeled as Markov decision processes (MDPs). In the MDP framework, at each time step the agent senses the state of the environment, and chooses and executes an action from the set of actions available to it in that state. The agent’s action (and perhaps other uncontrolled external events) cause a stochastic change in the state of the environment. The agent receives a (possibly zero) scalar reward from the environment. The agent’s goal is to choose actions so as to maximize the expected sum of rewards over some time horizon. An optimal policy is a mapping from states to actions that achieves the agent’s goal. Many RL algorithms have been developed for learning good approximations to an optimal policy from the agent’s experience in its environment. At a high level, most algorithms use this experience to learn value functions (or -values) that map state-action pairs to the maximal expected sum of reward that can be achieved starting from that state-action pair. The learned value function is used to choose actions stochastically, so that in each state, actions with higher value are chosen with higher probability. In addition, many RL algorithms use some form of function approximation (parametric representations of complex value functions) both to map state-action features to their values and to map states to distributions over actions (i.e., the policy). See (Sutton and Barto1998) for an extensive introduction to RL. In the next sections, we describe the Cobot’s actions, our choice of state features, and how we dealt with multiple sources of reward. The particular RL algorithm we use is a variant of (Sutton et al.1999)’s policy gradient algorithm. Its details are beyond the scope of this paper; however, see (Shelton2000) for details. One aspect of our RL algorithm that is relevant to understanding our results is that we use a linear function approximator to store our policy. In other words, for each state feature, we maintain a vector of real-valued weights indexed by the possible actions. A positive weight for some action means that the feature increases the probability of taking that action, while a negative weight decreases the probability. The weight’s magnitude determines the strength of this contribution. 5 Cobot’s RL Actions To have any hope of learning to behave in a way interesting to LambdaMOO users, Cobot’s actions must “make sense” to them, fit in with the social chat-based environment, and minimize the risk of causing irritation. Conversation, word play, and emoting routines are among the most common activity in LambdaMOO, so we designed a set of actions along these lines, as detailed in Table 1. Many of these actions extract an utterance from the recent conversations, or from a continually changing external source, such as the online Boston Globe. Thus a single action may cause an infinite variety of behavior by Cobot. At set time intervals (only every few minutes on average, to minimize spam), Cobot selects an action to perform from this set according to a distribution determined by the Q-values in his current state. Any rewards or punishments received before the next RL action are attributed to the current action, and used to update Cobot’s value functions. It is worth remembering that Cobot has two different categories of action: those actions taken proactively as a result of the RL, and those actions taken in response to a user’s action towards Cobot. Some users are certainly aware of the distinction and can easily determine which actions fall into which category, but other users may occasionally reward or punish Cobot in response to a reactive action. Such “erroneous” rewards and punishments act as a source of noise in the training process. 6 The RL Reward Function Cobot learns to behave directly from the feedback of LambdaMOO users, any of whom can reward or punish him. There are both explicit and implicit feedback mechanisms. We implemented explicit reward and punish verbs on Cobot that LambdaMOO users can invoke at any time. These verbs give a numerical (positive and negative, respectively) training signal to Cobot that is the basis of the RL. The signal is attributed as immediate feedback for the current state and RL action, and “backed up” to previous states and actions in accordance with the standard RL algorithms. There are several standard LambdaMOO verbs that are commonly used to express, sometimes playfully, approval or disapproval. Examples of the former include the verb hug, and of the latter the verb spank. In the interest of allowing the RL process to integrate naturally with the LambdaMOO environment, we chose to accept a number of such verbs as implicit reward and punishment signals for Cobot; however, such implicit feedback is numerically weaker than the feedback generated by the explicit mechanisms. One fundamental design choice is whether to learn a single value function for the entire community, or to learn separate value functions for each user based on individual feedback, combining the value functions of those present to determine how to act at each moment. We opted for the latter for three primary reasons. First, it was clear that for learning to have any hope of success, ths system must represent who is present at any given moment—different users simply have different personalities and preferences. We felt that representing which users are present as additional state features would throw away valuable domain information, as the RL would have to discover on its own the primacy of user identity. Having separate reward functions for each user is thus a way of asserting the importance of identity to the learning process. Second, despite the extremely limited number of training examples available in this domain ( per month), learning must be quick and significant. Without a clear sense that their training has some impact on Cobot’s behavior, users will quickly lose interest in providing feedback. A known challenge for RL is the “curse of dimensionality,” (i.e. the size of the state space increases exponentially with the number of state features). By avoiding the need to represent the presence or absence of roughly 250 users, we are able to maintain a fairly small state space and so speed up learning. Third, we (correctly) anticipated the fact that certain users would provide an inordinate amount of training to Cobot, and we did not want the overall policy followed by Cobot to be dominated by the preferences of these individuals. By learning separate policies for each user, and then combining these policies among those users present, we can limit the impact any single user can have on Cobot’s actions. 7 Cobot’s RL State Features The decision to maintain and learn separate value functions for each user means that we can maintain separate state spaces as well, in the hopes of simplifying states and speeding learning. Cobot can be viewed as running a large number of separate RL processes in parallel, with each process having a different state space. The state space for a user contains a number of features containing statistics about that particular user. LambdaMOO is a social environment, and Cobot is learning to take social actions, so we felt that his state features should contain information allowing him to gauge social activity and relationships. Table 2 provides a description of the state features used for RL by Cobot for each user. Even though we have simplified the state space by partitioning by user, the state space for a single user remains sufficiently complex to preclude standard table-based representation of value functions (also, each user’s state space is effectively infinite, as there are real-valued state features). Thus, linear function approximation is used for each user’s policy. Cobot’s RL actions are then chosen according to a mixture of the policies of the users present. We refer the reader to (Shelton2000) for more details on the method by which policies are learned and combined. Social Summary Vector A vector of four numbers: the rate at which the user is producing events; the rate at which events are being produced that are directed at the user; the percentage of the other users present who are among this user’s ten most frequently interacted-with users (“playmates”); and the percentage of the other users present for whom this user is among their top ten playmates. Mood Vector A vector measuring the recent use of eight groups of common verbs (e.g., one group includes verbs grin and smile). Verbs were grouped according to how well their usage was correlated. Rates Vector A vector measuring the rate at which events are produced by those present. Current Room The room where Cobot currently resides. Roll Call Vector Indicates if Cobot’s currently saved roll call text has been used before, if someone has done a roll call since the last time Cobot did, and if there has been a roll call since the last time Cobot grabbed new text. Bias Each user has one feature that is always “on”; that is, this bias is always set to a value of 1. Intuitively, it is the feature indicating the user’s “presence.” Table 2: State space of Cobot. Each user has his own state space and value function; the table thus describes the state space maintained for a generic user. 8 Experimental Procedure and Findings Cobot has been present in LambdaMOO more or less continuously since September, 1999. The RL version of Cobot debuted May 10, 2000. Again, Cobot’s various reactive functionality was left intact for the duration of the RL experiment. Cobot is a working system with real human users, and we wanted to perform the RL experiment in this context. Upon launching the RL functionality publicly in the Living Room, Cobot logged all RL-related data (states visited, actions taken, rewards received from each user, parameters of the value functions, etc.) from May 10 until October 10, 2000. During this time, 63123 RL actions were taken (in addition, of course, to many more reactive non-RL actions), and 3171 reward and punishment events were received from 254 different users. The findings we now summarize are based on these extensive logs: Inappropriatenessof average reward. The most standard and obvious sign of successfulRL would be an increase in the average reward over time. Instead, as shown in Figure 1a, the average cumulative reward received by Cobot actually goes down. However, rather than indicating that users are becoming more dissatisfied as Cobot learns, the decay in reward reveals some peculiarities of human feedback in such an open-ended environment. There are at least two difficulties with average cumulative reward in an environment of human users. The first is that humans are fickle, and their tastes and preferences may drift over time. Indeed, our experiences as users, and with the original reactive functionality of Cobot, suggest that novelty is highly valued in LambdaMOO. Thus a feature that is popular and exciting to users when it is introduced may eventually become an irritant (there are many examples of this phenomenon). In RL terminology, we do not have a fixed, consistent reward function, and thus we are always learning a moving target. While difficult to quantify in such a complex environment, this phenomenon is sufficiently prevalent in LambdaMOO to cast serious doubts on the use of average cumulative reward as the primary measure of performance. The second and related difficulty is that even when users do maintain relatively fixed preferences, they tend to give Cobot less feedback of either type (reward or punishment) as he manages to learn their preferences accurately. Simply put, once Cobot seems to be behaving as they wish, users feel no need to continually provide reward for his “correct” actions or to punish him for the occasional “mistake.” This reward pattern is in contrast to typical RL applications, where there is an automated and indefatigable reward source. Strong empirical evidence for this second phenomenon is provided by User M and User S. These two users were among Cobot’s most dedicated trainers, each had strong preferences for certain actions, and Cobot learned to strongly modify his behavior in their presence to match their preferences. Nevertheless, both users tended to provide less frequent feedback to Cobot as the experiment progressed, as shown in Figure 1a. We conclude that there are serious conceptual difficulties with the use of average cumulative reward in such a human-centric application of RL, and that alternative measures must be investigated, which we do below. A small set of dedicated “parents.” Among the 254 users who gave at least one reward or punishment event to Cobot, 218 gave 20 or fewer, while 15 gave 50 or more. Thus, we found that while many users exhibited a passing interest in training Cobot, there was a small group that was willing to invest nontrivial time and effort in teaching Cobot their preferences. In particular, User M and User S, generated 594 and 69 rewards and punishments events, respectively. By “event”, we simply mean an RL action that received some feedback. The actual absolute User O Roll Call. User O appears to especially dislike roll call actions when there have been repeated roll calls and/or Cobot is repeating the same roll calls. Rates. The overall rate of events being generating has slightly more relevance than that of the rate of events being generated just by User O. User B Social Summary. User B is effected by the presence of his friends. Not shown here are other Social Summary features (deviating about 6 degrees). It appears that User B is more likely to ignore Cobot when he is with many friends. User C Roll Call. User C appears to have strong preferences about Cobot’s behavior when a “roll call party” is in progress (i.e., everyone is generating roll calls). User P Room. User P would follow Cobot to his home, where he is generally alone, and has trained him there. He appears to have different preferences for Cobot under those circumstances. Table 3: Relevant features for users with non-uniform policies. Several of our top users had some features that deviated from their bias feature. The second column indicates the number of degrees between the weight vectors for those features and the weight vectors for the bias feature. We have only included features that deviated by more than 10 degrees. For the users above the double line, we have included only features whose weights had a length greater than 0.2. Each of these users had bias weights of length greater than 1. For those below the line, we have included only features with a length greater than 0.1 (these all had bias weights of length much less than 1). Some parents have strong opinions. For the vast majority of users who participated in the RL training of Cobot, the policy learned was quite close to the uniform distribution. Quantification of this statement is somewhat complex, since policies are dependent on state. However, we observed that for most users the learned policy’s dependence on state was weak, and the resulting distribution near uniform (though there are interesting and notable exceptions, as we shall see below). This result is perhaps to be expected: most users provided too little feedback for Cobot to detect strong preferences, and may not have been exhibiting strong and consistent preferences in the feedback they did provide. However, there was again a small group of users for whom a highly non-uniform policy was learned. In particular, for Users M and S mentioned above, the resulting policies were relatively independent of state and their entropies were 0.03 and 1.93, respectively. (The entropy of the uniform distribution over the actions is 2.2.) Several other users also exhibited less dramatic but still non-uniform distributions. User M seemed to have a strong preference for roll call actions, with the learned policy selecting these with probability 0.99, while User S preferred social commentary actions, with the learned policy giving them probability 0.38. (Each action in the uniform distribution is given weight 1/9 = 0.11.) Cobot learns matching policies. In Figure 1b, we demonstrate that the policy learned by Cobot for User M does in fact reflect the empirical pattern of rewards received over time. Similar results obtain for User S, not shown here. Thus, repeated feedback given to Cobot for a non-uniform set of preferences clearly pays off in a corresponding policy. Cobot responds to his dedicated parents. The policies learned by Cobot for users can have strong impact on the empirical distribution of actions he actually ends up taking in LambdaMOO. For User M, we find that his presence causes a significant shift towards his preferences. In other words, Cobot does his best to “please” these dedicated trainers whenever they arrive in the Living Room, and returns to a more uniform policy upon their departure. Some preferences depend on state. Finally, we show that the policies learned by Cobot sometimes depend upon the features Cobot maintains in his state. We use two facts about the RL weights (described in Section 4) maintained by Cobot to determine which features are relevant for a given user. First, we note that by construction, the RL weights learned for the bias feature described in Table 2 representthe user’s preferences independentof state (since this feature is always on whenever the user is present). Second, we note that because we initialized all weights to 0, only features with non-zero weights will contribute to the policy that Cobot uses. Thus, we can determine that a feature is relevant for a user if that feature’s weight vector is far from that user’s bias feature weight vector, and from the all-zero vector. For our purposes, we have used (1) the normalized inner product (the cosine of the angle between two vectors) as a measure of a feature’s distance from the bias feature, and (2) a feature’s weight vector length to determine if it is away from zero. Thesemeasures show that for most users, Cobot learned a state-independent policy (e.g., User M prefers roll calls); however, as we can see in Table 3, Cobot has learned a policy for some users that depends upon state. numerical reward received may be larger or smaller than 1 at any time, as implicit rewards provide fractional reward, and the user may repeatedly reward or punish an action, with the feedback being summed. For example, the total absolute value of rewards and punishments provided by User M was 607.63 over 594 feedback events. 0 1 2 3 4 5 6 x 10 4 0 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 0.2 Average Cumulative Reward per Timestep Time Reward reward all users abs reward all users reward user M abs reward user M reward user S abs reward user S 1 2 3 4 5 6 7 8 9 −0.4 −0.2 0 0.2 0.4 0.6 0.8 1 Rewards / Policy / Empirical Distribution Comparision for User M action change rewards policy empirical Figure 1: a) Average Cumulative Reward Over Time. b) Rewards received, policy learned, and effect on actions for User M. Figure a) shows that average cumulative reward decreases over time, for both total and absolute reward; however, Figure b shows that proper learning is taking place. For each of the RL actions, three quantities are shown. The blue bars (left) show the average reward given by User M for each action (the average reward given by User M across all actions has been subtracted off to indicate relative preferences). The yellow bars (middle) show the policy learned by Cobot for User M (the probability assigned to each action in the uniform distribution (1/9) has been subtracted off). The red bars (right) show the empirical frequency with which each action was taken in the presence of User M (minus the empirical frequency with which that action was taken by Cobot over all time steps). These bars indicate the extent to which the presence of User M biases Cobot’s behavior towards M’s preferences. We see that the policy learned by Cobot for User M aligns nicely with the preferences expressed by M and that Cobot’s behavior shifts strongly towards the learned policy for User M wheneverM is present. To go beyond a qualitative visual analysis, we have defined a metric that measures the extent to which two rankings of actions agree, while taking into account that some actions are extremely close in the each ranking. The details are beyond the scope of the paper, but the agreement between the action rankings shown here are in near-perfect agreement by this measure. Similar results obtain for User S. 9 Conclusions We have reported on our efforts to apply reinforcement learning in a complex human online social environment where many of the standard assumptions (stationary rewards, Markovian behavior, appropriateness of average reward) are clearly violated. We feel that the results obtained with Cobot so far are compelling, and offer promise for the application of RL in such open-ended social settings. Cobot continues to take RL actions and receive rewards and punishments from LambdaMOO users, and we plan to continue and embellish this work as part of our overall efforts on Cobot. References Foner, L. (1997). Entertaining Agents: a Sociological Case Study. In Proceedings of the First International Conference on Autonomous Agents. Isbell, C. L., Kearns, M., Kormann, D., Singh, S., and Stone, P. (2000). Cobot in LambdaMOO: A Social Statistics Agent. To appear in Proceedings of AAAI-2000. Mauldin, M. (1994). Chatterbots, TinyMUDs, and the Turing Test: Entering the Loebner Prize Competition. In Proceedings of the Twelfth National Conference on Artificial Intelligence. Shelton, C. R. (2000). Balancing Multiple Sources of Reward in Reinforcement Learning. Submitted for publication in Neural Information Processing Systems-2000. Singh, S., Kearns, M., Littman, D., and Walker, M. (2000). Empirical Evaluation of a Reinforcement Learning Dialogue System. To appear in Proceedings of AAAI-2000. Sutton, R. S. and Barto, A. G. (1998). Reinforcement Learning: An Introduction. MIT Press, Cambridge, MA. Sutton, R. S., McAllester, D., Singh, S., and Mansour, Y. (1999). Policy gradient methods for reinforcement learning with function approximation. In Neural Information Processing Systems1999.
|
2001
|
106
|
1,913
|
Semi-Supervised MarginBoost F. d'Alche-Buc LIP6,UMR CNRS 7606, Universite P. et M. Curie 75252 Paris Cedex, France florence. dAlche@lip6.fr Yves Grandvalet Heudiasyc, UMR CNRS 6599, Universite de Technologie de Compiegne, BP 20.529, 60205 Compiegne cedex, France Yves. Grandvalet@hds.utc.fr Christophe Ambroise Heudiasyc, UMR CNRS 6599, Universite de Technologie de Compiegne, BP 20.529, 60205 Compiegne cedex, France Christophe A mbroise@hds.utc.fr Abstract In many discrimination problems a large amount of data is available but only a few of them are labeled. This provides a strong motivation to improve or develop methods for semi-supervised learning. In this paper, boosting is generalized to this task within the optimization framework of MarginBoost . We extend the margin definition to unlabeled data and develop the gradient descent algorithm that corresponds to the resulting margin cost function. This meta-learning scheme can be applied to any base classifier able to benefit from unlabeled data. We propose here to apply it to mixture models trained with an Expectation-Maximization algorithm. Promising results are presented on benchmarks with different rates of labeled data. 1 Introduction In semi-supervised classification tasks, a concept is to be learnt using both labeled and unlabeled examples. Such problems arise frequently in data-mining where the cost of the labeling process can be prohibitive because it requires human help as in video-indexing, text-categorization [12] and medical diagnosis. While some works proposed different methods [16] to learn mixture models [12], [1], SVM [3], cotrained machines [5] to solve this task, no extension has been developed so far for ensemble methods such as boosting [7, 6]. Boosting consists in building sequentially a linear combination of base classifiers that focus on the difficult examples. For AdaBoost and extensions such as MarginBoost [10], this stage-wise procedure corresponds to a gradient descent of a cost functional based on a decreasing function of the margin, in the space of linear combinations of base classifiers. We propose to generalize boosting to semi-supervised learning within the framework of optimization. We extend the margin notion to unlabeled data, derive the corresponding criterion to be maximized, and propose the resulting algorithm called Semi-Supervised MarginBoost (SSMBoost). This new method enhances our previous work [9] based on a direct plug-in extension of AdaBoost in the sense that all the ingredients of the gradient algorithm such as the gradient direction and the stopping rule are defined from the expression of the new cost function. Moreover, while the algorithm has been tested using the mixtures of models [1], 55MBoost is designed to combine any base classifiers that deals with both labeled and unlabeled data. The paper begins with a brief presentation of MarginBoost (section 2). Then, in section 3, the 55MBoost algorithm is presented. Experimental results are discussed in section 5 and we conclude in section 6. 2 Boosting with MarginBoost Boosting [7, 6, 15] aims at improving the performance of any weak "base classifier" by linear combination. We focus here on normalized ensemble classifiers gt E LinCH) whose normalized1 coefficients are noted aT = I ~: I and each base classifier with outputs in [-1, 1] is hT E 1{: t gt(x) = L aThT(x) (1) T=l Different contributions [13, 14],[8], [10] have described boosting within an optimization scheme, considering that it carries out a gradient descent in the space of linear combinations of base functions. We have chosen the MarginBoost algorithm, a variant of a more general algorithm called Any Boost [10], that generalizes AdaBoost and formally justifies the interpretation in terms of margin. If S is the training sample {(Xi,Yi) ,i = l..l}, MarginBoost, described in Fig. 1, minimizes the cost functional C defined for any scalar decreasing function c of the margin p : I C(gt) = L c(p(gt(Xi), Yi))) (2) i=l Instead of taking exactly ht+l = - \1C(gt) which does not ensure that the resulting function gt+! belongs to Lin(1{), ht+! is chosen such as the inner product2 - < \1C(gt), ht+l > is maximal. The equivalent weighted cost function to be maximized can thus be expressed as : JF = L Wt(i)Yiht+! (Xi) iES 3 Generalizing MarginBoost to semi-supervised classification 3.1 Margin Extension (3) For labeled data, the margin measures the quality of the classifier output. When no label is observed, the usual margin cannot be calculated and has to be estimated. A first estimation could be derived from the expected margin EypL(gt(X) , y). We can use the output of the classifier (gt(x) + 1)/2 as an estimate of the posterior probability P(Y = +llx). This leads to the following margin pi; which depends on the input and is linked with the response of the classifier: lOr> 0 and L1 norm is used for normalization: IOrl = L~=l Or 2< f, 9 >= LiE S f(X;)g(Xi) Let wo(i) = l/l, i = 1, ... ,l. Let go(x) = 0 For t = 1 ... T (do the gradient descent): 1. Learn a gradient direction htH E 1i with a high value of J{ = L,iEswt(i)YihtH(Xi) 2. Apply the stopping rule: if J{ ::::: L,iES Wt(i)Yigt(Xi) then return gt else go on. 3. Choose a step-length for the obtained direction by a line-search or by fixing it as a constant f 4 Add the new direction to obtain 9 = (l a t I9t+a ttlhtt') . HI lattl l 5. Fix the weight distribution: Wt 1 = c'(p(9ttl(Xi),Yi)) + 2: jE S c'(p(9ttl(Xj),Yj)) Figure 1: MarginBoost algorithm (with L1 normalization of the combination coefficients) Another way of defining the extended margin is to use directly the maximum a posteriori estimate of the true margin. This MAP estimate depends on the sign of the classifier output and provides the following margin definition pC; : (5) 3.2 Semi-Supervised MarginBoost : generalization of marginBoost to deal with unlabeled data The generalization of the margin can be used to define an appropriate cost functional for the semi-supervised learning task. Considering that the training sample S is now divided into two disjoint subsets L for labeled data and U for unlabeled data, the cost falls into two parts involving PL = P and PU: (6) iEL iEU The maximization of - < \lC(gt), htH > is equivalent to optimize the new quantity JtS that falls now into two terms J{ = Jf + J? The first term one can be directly obtained from equation (3) : Jf = LWt(i).YihtH(Xi) (7) iEL The second term, J?, can be expressed as following: (8) with the weight distribution Wt now defined as : { c'(pL(9t(Xi),Yi)) ( .) _ IWt l Wt z c'(PU(9t(Xi))) IWt l if i E L .. with IWt I = 2:= Wt (i) If z E U iES (9) This expression of JP comes directly from differential calculus and the chosen inner product: ( )() { YiC'(Pd9t(Xi),Yi)) if x = Xi and i E L 'VC gt Xi = c'(p (g (x.))) apU(9t(Xi)) if x = x, and i E U U t t a9t(Xi) 0 (10) Implementation of 55MBoost with margins pI[; and Pu requires their derivatives. Let us notice that the "signed margin", pus, is not derivable at point O. However, according to the results of convex analysis (see for instance [2]), it is possible to define the "sub derivative' of Pus since it is a continuous and convex function. The value of the sub derivative corresponds here to the average value of the right and left derivatives. apUS(gt(Xi)) = {sign(g(Xi)) agt (Xi) 0 if X :f": 0 if x = 0 And, for the "squared margin" Pu9 , we have: apu 9 (gt(Xi)) = 2g(Xi) agt(Xi) (11) (12) This completes the set of ingredients that must be incorporated into the algorithm of Fig. 1 to obtain 55MBoost. 4 Base Classifier The base classifier should be able to make use of the unlabeled data provided by the boosting algorithm. Mixture models are well suited for this purpose, as shown by their extensive use in clustering. Hierarchical mixtures provide flexible discrimination tools, where each conditional distribution f(xlY = k) is modelled by a mixture of components [4]. At the high level, the distribution is described by K f(x; if» = 2:= Pk!k (x; Ok) , (13) k=l where K is the number of classes, Pk are the mixing proportions, Ok the conditional distribution parameters, and if> denotes all parameters {Pk; 0df=l. The high-level description can also be expressed as a low-level mixture of components, as shown here for binary classification: Kl K2 f(x;if» = 2:= PkJkl(X;Okl) + 2:= Pk2!k2(X;Ok2) (14) With this setting, the EM algorithm is used to maximize the log-likelihood with respect to if> considering the incomplete data is {Xi, Yi}~= l and the missing data is the component label Cik, k = 1, ... , K 1 + K2 [11]. An original implementation of EM based on the concept of possible labels [1] is considered here. It is well adapted to hierarchical mixtures, where the class label Y provides a subset of possible components. When Y = 1 the first Kl modes are possible, when Y = -1 the last K2 modes are possible, and when an example is unlabeled, all modes are possible. A binary vector Zi E {0,1}(Kl+K2) indicates the components from which feature vector Xi may have been generated, in agreement with the assumed mixture model and the (absence of) label Yi. Assuming that the training sample {Xi, Zi }i=l is i.i.d, the weighted log-likelihood is given by I L(<I>;{Xi,zdi=l = LWt(i) log (j(Xi,zi;<I») , (15) i=l where Wt(i) are provided by boosting at step t. L is maximized using the following EM algorithm: E-Step Compute the expectation of L( <I>; {Xi, zdi=l) conditionally to {Xi, zdi=l and the current value of <I> (denoted <I>q): with Uik I Kl+K2 L L Wt(i)Uik log (Pk!k(Xi; Ok)) i=l k=l ZikPk!k(Xi; Ok) L£ ZUP£!£(Xi; O£) M-Step Maximize Q(<I>I<I>q) with respect to <I>. (16) Assuming that each mode k follows a normal distribution with mean ILk' and covariance ~k ' <I>q+l = {ILk+! ; ~k+!;Pk+l}f~iK2 is given by: (17) (18) 5 Experimental results Tests of the algorithm are performed on three benchmarks of the boosting literature: twonorm and ringnorm [6] and banana [13]. Information about these datasets and the results obtained in discrimination are available at www.first.gmd.de/-raetsch/ 10 different samples were used for each experiment. We first study the behavior of 55MBoost according the evolution of the test error with increasing rates of unlabeled data (table 1). We consider five different settings where 0%, 50%, 75%, 90% and 95% of labels are missing. 55MB is tested for the margins P~ and Pu with c(x) = exp( -x). It is compared to mixture models and AdaBoost. 55MBoost and AdaBoost are trained identically, the only difference being that AdaBoost is not provided with missing labels. Both algorithms are run for T = 100 boosting steps, without special care of overfitting. The base classifier (called here base(EM)) is a hierarchical mixture model with an arbitrary choice of 4 modes per class but the algorithm (which may be stalled in local minima) is restarted 100 times from different initial solutions, and the best final solution (regarding training error rate) is selected. We report mean error rates together with the lower and upper quartiles in table 1. For sake of space, we did not display the results obtained without missing labels: in this case, AdaBoost and 55MBoost behave nearly identically and better than EM only for Banana. For rates of unlabeled data inferior to 95%, 55MBoost beats slightly AdaBoost for Ringnorm and Twonorm (except for 75%) but is not able to do as well as Table 1: Mean error rates (in %) and interquartiles obtained with 4 different percentages of unlabeled data for mixture models base(EM), AdaBoost and 55MBoost. Ringnorm 50% 75% 90% 95% base(EM) 2.1 [ 1.7, 2.1] 4.3[ 1.9, 5.7] 9.5 [ 2.7,12.0] 23.7[14.5,27.0] AdaBoost 1.8[ 1.6, 2.0] 3.1[ 1.9, 4.1] 11.5[ 4.2,12.1] 28.7[11.5,37.6] 55MBoost pS 1. 7[ 1.5, 1.8] 2.0 [ 1.5, 2.4] 3.7[ 2.1, 4.8] 6.9[ 5.6,10.7] 55MBoost pg 1. 7[ 1.6, 1.8] 2.O[ 1.4, 2.5] 4.5 [ 2.2, 3.6] 8.1 [ 4.2, 9.0] Twonorm 50% 75% 90% 95% base(EM) 3.2[ 2.7, 3.1] 6.5[ 3.0, 9.0] 20.6[10.3,22.5] 24.8[18.3,31.9] AdaBoost 3.2[ 2.9, 3.2] 3.2[ 3.0, 3.5] 11.0[ 5.2,14.2] 38.9[29.4,50.0] 55MBoost pS 2.7[ 2.5, 2.9] 3.4[ 2.8, 4.3] 10.1 [ 5.8,13.6] 20.4[11.9,32.3] 55MBoost pg 2.7[ 2.5, 2.8] 3.4[ 2.8, 4.2] 11.0[ 5.6,16.2] 21.1 [12.5,30.8] Banana 50% 75% 90% 95% base(EM) 18.2[16.7,18.6] 21.8[18.0,25.0] 26.1[20.7,29.8] 31.7[23.8,35.8] AdaBoost 12.6[11.7,13.1] 15.2 [13.0,16.8] 22.1 [18.0,24.3] 37.5 [32.2,42.2] 55MBoost pS 13.3 [12.7,14.3] 17.0[15.3,17.8] 22.2[18.0,28.0] 28.3 [20.2,35.2] 55MBoost pg 13.3[12.8,14.2] 16.9[15.6,17.8] 22.8[18.3,29.3] 28.6[21.5,34.2] AdaBoost on Banana data. One possible explanation is that the discrimination frontiers involved in the banana problem are so complex that the labels really bring crucial informations and thus adding unlabeled data does not help in such a case. Nevertheless, at rate 95% which is the most realistic situation, the margin Pu obtains the minimal error rate for each of the three problems. It shows that it is worth boosting and using unlabeled data. As there is no great difference between the two proposed margins, we conducted further experiments using only the Pu' Second, in order to study the relation between the presence of noise in the dataset and the ability of 55MBoost to enhance generalization performance, we draw in Fig. 2, the test errors obtained for problems with different values of Bayes error when varying the rate of labeled examples. We see that even for difficult tasks (very noisy problems), the degradation in performance for large subsets of unlabeled data is still low. This reflects some consistency in the behavior of our algorithm. Third, we test the sensibility of 55MBoost to overfitting. Overfitting can usually be avoided by techniques such as early stopping, softenizing of the margin ([13], [14]) or using an adequate margin function such as 1 - tanh(p) instead of exp( -p) [10]. Here we keep using c = exp and ran 55MBoost with a maximal number of step T = 1000 with 95% of unlabeled data. Of course, this does not correspond to a realistic use of boosting in practice but it allows to check if the algorithm behaves consistently in terms of gradient steps number. It is remarkable that no overfitting is observed and in the Twonorm case (see Fig. 3), the test error still decreases ! We also observe that the standard error deviation is reduced at the end of the process. For the banana problem (see Fig. 3 b.), we observe a stabilization near the step t = 100. A massive presence of unlabeled data implies thus a regularizing effect. 50 40 20 10 Bayes error:;;; 2.3% Bayes error:;;; 15.7% Bayes error:;;; 3 1.2% °0L---~,7 0 --~207---~~7----4~0--~5~0--~6=0--~7=0----8=0----9=0~~ '00 Rate of missing labels (%) Figure 2: Consistency of the 55MBoost behavior: evolution of test error versus the missing labels rate with respect to various Bayes error (twonorm ). 70 60' , , I \ "' \ " Mean (Error Test) +/- 1 std Mean (Error test) -'-",---.- - --/----oL-~ __ ~ __ ~ __ _L __ _L __ ~ __ ~ __ ~ __ L_~ o ~ ~ _ ~ ~ ~ ~ ~ ~ _ Steptofgradientdescent(boosting process} 70 60 ~ 50 i \!) 0: 40" , \ 10 I ~ Mean of Error Test +/- std Mean of Error test ~_~~ __ -r~/_ ~ ~",~~,.~. '-. -.I~" ~~-I °OL-~'OO~-2~OO~~3~OO--~400~~500~-=~~~7~OO~~8=OO--~~~~'~ Step t of gradient descent Figure 3: Evolution of Test error with respect to maximal number T of iterations with 95% of missing labels (Two norm and Banana). 6 Conclusion MarginBoost algorithm has been extended to deal with both labeled and unlabeled data. Results obtained on three classical benchmarks of boosting litterature show that it is worth using additional information conveyed by the patterns alone. No overfitting was observed during processing 55MBoost on the benchmarks when 95% of the labels are missing: this should mean that the unlabeled data should playa regularizing role in the ensemble classifier during the boosting process. After applying this method to a large real dataset such as those of text-categorization, our future works on this theme will concern the use of the extended margin cost function on the base classifiers itself like multilayered perceptrons or decision trees. Another approach could also be conducted from the more general framework of AnyBoost that optimize any differential cost function. References [1] C. Ambroise and G. Govaert. EM algorithm for partially known labels. In IFCS 2000, july 2000. [2] J.-P. Aubin. L 'analyse non lineaire et ses applications d l'economie. Masson, 1984. [3] K P. Bennett and A. Demiriz. Semi-supervised support vector machines. In D. Cohn, M. Kearns, and S. Solla, editors, Advances in Neural Information Processing Systems, pages 368-374. MIT Press, 1999. [4] C.M. Bishop and M.E. Tipping. A hierarchical latent variable model for data vizualization. IEEE PAMI, 20:281- 293, 1998. [5] A. Blum and Tom Mitchell. Combining labeled and unlabeled data with co-training. In Proceedings of the 1998 Conference on Computational Learning Theory, July 1998. [6] L. Breiman. Prediction games and arcing algorithms. Technical Report 504, Statistics Department, University of California at Berkeley, 1997. [7] Y. Freund and R. E. Schapire. Experiments with a new boosting algorithm. In Machine Learning: Proceedings of the Thirteenth International Conference, pages 148- 156. Morgan Kauffman, 1996. [8] J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: a statistical view of boosting. The Annals of Statistics, 28(2):337- 407, 2000. [9] Y. Grandvalet, F. d'Alche Buc, and C. Ambroise. Boosting mixture models for semisupervised learning. In ICANN 2001, august 200l. [10] L. Mason, J. Baxter, P. L. Bartlett, and M. Frean. Functional gradient techniques for combining hypotheses. In Advances in Large Margin Classifiers. MIT, 2000. [11] G.J. McLachlan and T. Krishnan. The EM algorithm and extensions. Wiley, 1997. [12] K Nigam, A. K McCallum, S. Thrun, and T. Mitchell. Text classification from labeled and unlabeled documents using EM. Machine learning, 39(2/3):135- 167, 2000. [13] G. Riitsch, T. Onoda, and K-R. Muller. Soft margins for AdaBoost. Technical report, Department of Computer Science, Royal Holloway, London, 1998. [14] G. Riitsch, T. Onoda, and K-R. Muller. Soft margins for AdaBoost. Machine Learning, 42(3):287- 320, 200l. [15] R. E. Schapire, Y. Freund, P. Bartlett, and W. S. Lee. Boosting the margin: A new explanation for the effectiveness of voting methods. Th e Annals of Statistics, 26(5):1651- 1686, 1998. [16] Matthias Seeger. Learning with labeled and unlabeled data,www.citeseer.nj.nec.com/seegerOllearning.html.
|
2001
|
107
|
1,914
|
ALGONQUIN - Learning dynamic noise models from noisy speech for robust speech recognition Brendan J. Freyl, Trausti T. Kristjanssonl , Li Deng2 , Alex Acero2 1 Probabilistic and Statistical Inference Group, University of Toronto http://www.psi.toronto.edu 2 Speech Technology Group, Microsoft Research Abstract A challenging, unsolved problem in the speech recognition community is recognizing speech signals that are corrupted by loud, highly nonstationary noise. One approach to noisy speech recognition is to automatically remove the noise from the cepstrum sequence before feeding it in to a clean speech recognizer. In previous work published in Eurospeech, we showed how a probability model trained on clean speech and a separate probability model trained on noise could be combined for the purpose of estimating the noisefree speech from the noisy speech. We showed how an iterative 2nd order vector Taylor series approximation could be used for probabilistic inference in this model. In many circumstances, it is not possible to obtain examples of noise without speech. Noise statistics may change significantly during an utterance, so that speechfree frames are not sufficient for estimating the noise model. In this paper, we show how the noise model can be learned even when the data contains speech. In particular, the noise model can be learned from the test utterance and then used to de noise the test utterance. The approximate inference technique is used as an approximate E step in a generalized EM algorithm that learns the parameters of the noise model from a test utterance. For both Wall Street Journal data with added noise samples and the Aurora benchmark, we show that the new noise adaptive technique performs as well as or significantly better than the non-adaptive algorithm, without the need for a separate training set of noise examples. 1 Introduction Two main approaches to robust speech recognition include "recognizer domain approaches" (Varga and Moore 1990; Gales and Young 1996), where the acoustic recognition model is modified or retrained to recognize noisy, distorted speech, and "feature domain approaches" (Boll 1979; Deng et al. 2000; Attias et al. 2001; Frey et al. 2001), where the features of noisy, distorted speech are first denoised and then fed into a speech recognition system whose acoustic recognition model is trained on clean speech. One advantage of the feature domain approach over the recognizer domain approach is that the speech modeling part of the denoising model can have much lower complexity than the full acoustic recognition model. This can lead to a much faster overall system, since the denoising process uses probabilistic inference in a much smaller model. Also, since the complexity of the denoising model is much lower than the complexity of the recognizer, the denoising model can be adapted to new environments more easily, or a variety of denoising models can be stored and applied as needed. We model the log-spectra of clean speech, noise, and channel impulse response function using mixtures of Gaussians. (In contrast, Attias et al. (2001) model autoregressive coefficients.) The relationship between these log-spectra and the log-spectrum of the noisy speech is nonlinear, leading to a posterior distribution over the clean speech that is a mixture of non-Gaussian distributions. We show how a variational technique that makes use of an iterative 2nd order vector Taylor series approximation can be used to infer the clean speech and compute sufficient statistics for a generalized EM algorithm that can learn the noise model from noisy speech. Our method, called ALGONQUIN, improves on previous work using the vector Taylor series approximation (Moreno 1996) by modeling the variance of the noise and channel instead of using point estimates, by modeling the noise and channel as a mixture mixture model instead of a single component model, by iterating Laplace's method to track the clean speech instead of applying it once at the model centers, by accounting for the error in the nonlinear relationship between the log-spectra, and by learning the noise model from noisy speech. 2 ALGONQUIN's Probability Model For clarity, we present a version of ALGONQUIN that treats frames of log-spectra independently. The extension of the version presented here to HMM models of speech, noise and channel distortion is analogous to the extension of a mixture of Gaussians to an HMM with Gaussian outputs. Following (Moreno 1996), we derive an approximate relationship between the log spectra of the clean speech, noise, channel and noisy speech. Assuming additive noise and linear channel distortion, the windowed FFT Y(j) for a particular frame (25 ms duration, spaced at 10 ms intervals) of noisy speech is related to the FFTs of the channel H(j), clean speech 5(j) and additive noise N(j) by Y(j) = H(j)5(j) + N(j). (1) We use a mel-frequency scale, in which case this relationship is only approximate. However, it is quite accurate if the channel frequency response is roughly constant across each mel-frequency filter band. For brevity, we will assume H(j) = 1 in the remainder of this paper. Assuming there is no channel distortion simplifies the description of the algorithm. To see how channel distortion can be accounted for in a nonadaptive way, see (Frey et al. 2001). The technique described in this paper for adapting the noise model can be extended to adapting the channel model. Assuming H(j) = 1, the energy spectrum is obtained as follows: IY(j)12 = Y(j)*Y(j) = 5(j)* 5(j) + N(j)* N(j) + 2Re(N(j)* 5(j)) = 15(j)12 + IN(j)12 + 2Re(N(j)* 5(j)) , where "*,, denotes complex conjugate. If the phase of the noise and the speech are uncorrelated, the last term in the above expression is small and we can approximate the energy spectrum as follows: IYUW ~ ISUW + INUW· (2) Although we could model these spectra directly, they are constrained to be nonnegative. To make density modeling easier, we model the log-spectrum instead. An additional benefit to this approach is that channel distortion is an additive effect in the log-spectrum domain. Letting y be the vector containing the log-spectrum log IY(:W, and similarly for s and n , we can rewrite (2) as exp(y) ~ exp(s) + exp(n) = exp(s) 0 (1 + exp(n - s)), where the expO function operates in an element-wise fashion on its vector argument and the "0" symbol indicates element-wise product. Taking the logarithm, we obtain a function gO that is an approximate mapping of sand n to y (see (Moreno 1996) for more details): y ~ g([~]) = s + In(l + exp(n - s)). (4) "T" indicates matrix transpose and InO and expO operate on the individual elements of their vector arguments. Assuming the errors in the above approximation are Gaussian, the observation likelihood is p(yls,n) =N(y;g([~]),W), (5) where W is the diagonal covariance matrix of the errors. A more precise approximation to the observation likelihood can be obtained by writing W as a function of s and n , but we assume W is constant for clarity. Using a prior p(s, n), the goal of de noising is to infer the log-spectrum of the clean speech s, given the log-spectrum ofthe noisy speech y. The minimum squared error estimate of sis s = Is sp(sly), where p(sly) ex InP(yls, n)p(s, n). This inference is made difficult by the fact that the nonlinearity g([s n]T) in (5) makes the posterior non-Gaussian even if the prior is Gaussian. In the next section, we show how an iterative variational method that uses a 2nd order vector Taylor series approximation can be used for approximate inference and learning. We assume that a priori the speech and noise are independent p(s, n) = p(s)p(n) and we model each using a separate mixture of Gaussians. cS = 1, ... , NS is the class index for the clean speech and en = 1, ... ,Nn is the class index for the noise. The mixing proportions and Gaussian components are parameterized as follows: p(s) = LP(cS)p(slcS), p(CS) =7r~s , p(slcS) =N(s;JL~ s ,~~ s ), C S We assume the covariance matrices ~~ s and ~~n are diagonal. Combining (5) and (6), the joint distribution over the noisy speech, clean speech class, clean speech vector, noise class and noise vector is p(y, s, cs, n , en) = N(y; g([~]), w)7r~s N(s; JL~ s , ~~s )7r~N(n; JL~n , ~~n). (7) Under this joint distribution, the posterior p(s, nly) is a mixture of non-Gaussian distributions. In fact, for a given speech class and noise class, the posterior p(s, nics, en, y) may have multiple modes. So, exact computation of s is intractable and we use an approximation. 3 Approximating the Posterior For the current frame of noisy speech y, ALGONQUIN approximates the posterior using a simpler, parameterized distribution, q: p(s,cS, n,cnly) ~ q(s,cS,n,cn). The "variational parameters" of q are adjusted to make this approximation accurate, and then q is used as a surrogate for the true posterior when computing § and learning the noise model (c.f. (Jordan et al. 1998)). For each cS and en, we approximate p(s, nics, en, y) by a Gaussian, (9) where 1J~'en and 1J~'en are the approximate posterior means of the speech and noise for classes cS and en, and <P ~~en, <P~.r;,n and <P~::'en specify the covariance matrix for the speech and noise for classes cS and en. Since rows of vectors in (4) do not interact and since the likelihood covariance matrix q, and the prior covariance matrices ~ ~. and ~~n are diagonal, the matrices <P~~ en, <P~.r;,n and <P~::'en are diagonal. The posterior mixing proportions for classes cS and en are q( cS , en) = Pc' en. The approximate posterior is given by q(s,n,cs,cn) = q(s,nlcs,cn)q(cS, en). The goal of variational inference is to minimize the relative entropy (KullbackLeibler divergence) between q and p: K "''''11 ( S n) q(s,n ,cS,cn) = ~ ~ q s, n , c ,c In ( S n I ). c' en s n P s, c , n , c y This is a particularly good choice for a cost function, because, since lnp(y) doesn't depend on the variational parameters, minimizing K is equivalent to maximizing () K "''''11 ( S n) p(s,cS,n,cn,y) F = lnp y = ~ ~ q s, n , c ,c In ( S n) , e' en s n q s, n, c ,c which is a lower bound on the log-probability of the data. So, variational inference can be used as a generalized E step (Neal and Hinton 1998) in an algorithm that alternatively maximizes a lower bound on lnp(y) with respect to the variational parameters and the noise model parameters, as described in the next section. Variational inference begins by optimizing the means and variances in (9) for each CS and en. Initially, we set the posterior means and variances to the prior means and variances. F does not have a simple form in these variational parameters. So, at each iteration, we make a 2nd order vector Taylor series approximation of the likelihood, centered at the current variational parameters, and maximize the resulting approximation to F. The updates are where g' 0 is a matrix of derivatives whose rows correspond to the noisy speech y and whose columns correspond to the clean speech and noise [s n]. The inverse posterior covariance matrix is the sum of the inverse prior covariance matrix and the inverse likelihood covariance matrix, modified by the Jacobian g' 0 for the mapping from s and n to y The posterior means are moved towards the prior means and toward values that match the observation y. These two effects are weighted by the inverse prior covariance matrix and the inverse likelihood covariance matrix. After iterating the above updates (in our experiments, 3 to 5 times) for each eS and en, the posterior mixing proportions that maximize :F are computed: where A is a normalizing constant that is computed so that L e.en Pe'en = 1. The minimum squared error estimate of the clean speech, s, is We apply this algorithm on a frame-by-frame basis, until all frames in the test utterance have been denoised. 4 Speed Since elements of s, nand y that are in different rows do not interact in (4), the above matrix algebra reduces to efficient scalar algebra. For 256 speech components, 4 noise components and 3 iterations of inference, our unoptimized C code takes 60 ms to denoise each frame. We are confident that this time can be reduced by an order of magnitude using standard implementation tricks. 5 Adapting the Noise Model Using Noisy Speech The version of ALGONQUIN described above requires that a mixture model of the noise be trained on noise samples, before the log-spectrum of the noisy speech can be denoised. Here, we describe how the iterative inference technique can be used as the E step in a generalized EM algorithm for learning the noise model from noisy speech. For a set of frames y(1), . .. , yeT) in a noisy test utterance, we construct a total bound :F = L:F(t) :::; Llnp(y(t)). t t The generalized EM algorithm alternates between updating one set of variational (t) n(t) t £ h f tIT d ... T· h parameters Pe.en, 11 e'en, e c. or eac rame = , ... , ,an maximizIng.r WIt respect to the noise model parameters 7r~n, J.t~n and ~~n. Since:F:::; Ltlnp(y(t)), this procedure maximizes a lower bound on the log-probability of the data. The use of the vector Taylor series approximations leads to an algorithm that maximizes an approximation to a lower bound on the log-probability of the data. Restaurant Street Airport Station Average 20 dB 2.12 2.96 1.82 1.73 2.16 15 dB 3.87 4.78 2.27 3.24 3.54 10 dB 9.18 10.73 5.49 6.48 7.97 5 dB 20.51 13.52 14.97 15.18 18.54 o dB 47.04 45.68 36.00 37.24 41.49 -5dB 78.69 72.34 69.04 67.26 71.83 Average 16.54 17.53 12.11 12.77 14.74 Table 1: Word error rates (in percent) on set B of the Aurora test set, for the adaptive version of ALGONQUIN with 4 noise componentsset. Setting the derivatives of :F with respect to the noise model parameters to zero, we obtain the following M step updates: ~n ('"' '"' (t) (opnn(t) +d· (( n (t) n)( n (t) n )T))) / ('"' '"' (t) ) en +--~ ~ Pe. en e' en lag 11e' en -#-ten 11e' en -#-ten ~ ~ Pe. en . t c B t c B The variational parameters can be updated multiple times before updating the model parameters, or the variational parameters can updated only once before updating the model parameters. The latter approach may converge more quickly in some situations. 6 Experimental Results After training a 256-component speech model on clean speech, we used the adaptive version of ALGONQUIN to denoise noisy test utterances on two tasks: the publically available Aurora limited vocabulary speech recognition task (http://www.etsi.org/technicalactiv/dsr.htm); the Wall Street Journal (WSJ) large vocabulary speech recognition task, with Microsoft's Whisper speech recognition system. We obtained results on all 48 test sets from partitions A and B of the Aurora database. Each set contains 24,000 sentences that have been corrupted from one of 4 different noise types and one of 6 different signal to noise ratios. Table 1 gives the error rates for the adaptive version of ALGONQUIN, with 4 noise components. These error rates are superior to error rates obtained by our spectral subtraction technique for (Deng et al. 2000), and highly competitive with other results on the Aurora task. Table 2 compares the performances of the adaptive version of ALGONQUIN and the non-adaptive version. For the non-adaptive version, 20 non-speech frames are used to estimate the noise model. For the adaptive version, the parameters are initialized using 20 non-speech frames and then 3 iterations of generalized EM are used to learn the noise model. The average error rate over all noise types and SNRs for set B of Aurora drops from 17.65% to 15.19% when the noise adaptive algorithm is used to update the noise model. This is a relative gain of 13.94%. When 4 components are used there is a further gain of 2.5%. The Wall Street Journal test set consists of 167 sentences spoken by female speakers. The Microsoft Whisper recognizer with a 5,000 word vocabulary was used to recognize these sentences. Table 2 shows that the adaptive version of algonquin WER WER Reduction WER Reduction 20 frames 1 comp in WER 4 comps in WER Aurora, Set A 18.10% 15.91% 12.10% 15.62% 13.70% Aurora, Set B 17.65% 15.19% 13.94% 14.74% 16.49% WSJ, XD14, 10dB 30.00% 21.8% 27.33% 21.50% 28.33% WSJ, XD10, 10dB 21.80% 20.6% 5.50'70 20.6% 5.50 '70 Table 2: Word error rates (WER) and percentage reduction in WER for the Aurora test data and the Wall Street Journal test data, without scaling. performs better than the non-adaptive version, especially on noise type "XD14", which consists of the highly-nonstationary sound of a jet engine shutting down. For noise type "XD1O", which is stationary noise, we observe a gain, but we do not see any further gain for multiple noise components. 7 Conclusions A far as variational methods go, ALGONQUIN is a fast technique for denoising logspectrum or cepstrum speech feature vectors. ALGONQUIN improves on previous work using the vector Taylor series approximation, by using multiple component speech and noise models, and it uses an iterative variational method to produce accurate posterior distributions for speech and noise. By employing a generalized EM method, ALGONQUIN can estimate a noise model from noisy speech data. Our results show that the noise adaptive ALGONQUIN algorithm can obtain better results than the non-adaptive version. This is especially important for nonstationary noise, where the non-adaptive algorithm relies on an estimate of the noise based on a subset of the frames, but the adaptive algorithm uses all the frames of the utterance, even those that contain speech. A different approach to denoising speech features is to learn time-domain models. Attias et al. (2001) report results on a non-adaptive time-domain technique. Our results cannot be directly compared with theirs, since our results are for unscaled data. Eventually, the two approaches should be thoroughly compared. References Attias, H., Platt, J . C., Acero, A., and Deng, L. 2001. Speech denoising and dereverberation using probabilistic models. In Advances in Neural Information Processing Systems 13. MIT Press, Cambridge MA. Boll, S. 1979. Suppression of acoustic noise in speech using spectral subtraction. IEEE Transactions on Acoustics, Speech and Signal Processing, 27:114- 120. Deng, L., Acero, A., Plumpe, M., and Huang, X. D. 2000. Large-vocabulary speech recognition under adverse acoustic environments. In Proceedings of the International Conference on Spoken Language Processing, pages 806- 809. Frey, B. J., Deng, L. , Acero, A., and Krist jansson, T. 2001. ALGONQUIN: Iterating Laplace's method to remove multiple types of acoustic distortion for robust speech recognition. In Proceedings of Eurospeech 2001. Gales, M. J. F. and Young, S. J . 1996. Robust continuous speech recognition using parallel model combination. IEEE Speech and Audio Processing, 4(5):352- 359. Jordan, M. 1., Ghahramani, Z., Jaakkola, T. S., and Saul, L. K. 1998. An introduction to variational methods for graphical models. In Jordan, M. 1., editor, Learning in Graphical Models. Kluwer Academic Publishers, Norwell MA. Moreno, P. 1996. Speech Recognition in Noisy Environments. Carnegie Mellon University, Pittsburgh PA. Doctoral dissertation. Neal, R. M. and Hinton, G. E. 1998. A view of the EM algorithm that justifies incremental, sparse, and other variants. In Jordan, M. 1., editor, Learning in Graphical Models, pages 355- 368. Kluwer Academic Publishers, Norwell MA. Varga, A. P. and Moore, R. K. 1990. Hidden Markov model decomposition of speech and noise. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing, pages 845- 848. IEEE Press.
|
2001
|
108
|
1,915
|
Adaptive N earest Neighbor Classification using Support Vector Machines Carlotta Domeniconi, Dimitrios Gunopulos Dept. of Computer Science, University of California, Riverside, CA 92521 { carlotta, dg} @cs.ucr.edu Abstract The nearest neighbor technique is a simple and appealing method to address classification problems. It relies on the assumption of locally constant class conditional probabilities. This assumption becomes invalid in high dimensions with a finite number of examples due to the curse of dimensionality. We propose a technique that computes a locally flexible metric by means of Support Vector Machines (SVMs). The maximum margin boundary found by the SVM is used to determine the most discriminant direction over the query's neighborhood. Such direction provides a local weighting scheme for input features. We present experimental evidence of classification performance improvement over the SVM algorithm alone and over a variety of adaptive learning schemes, by using both simulated and real data sets. 1 Introduction In a classification problem, we are given J classes and l training observations. The training observations consist of n feature measurements x = (Xl,'" ,Xn)T E ~n and the known class labels j = 1, ... , J. The goal is to predict the class label of a given query q. The K nearest neighbor classification method [4, 13, 16] is a simple and appealing approach to this problem: it finds the K nearest neighbors of q in the training set, and then predicts the class label of q as the most frequent one occurring in the K neighbors. It has been shown [5, 8] that the one nearest neighbor rule has asymptotic error rate that is at most twice the Bayes error rate, independent of the distance metric used. The nearest neighbor rule becomes less appealing with finite training samples, however. This is due to the curse of dimensionality [2]. Severe bias can be introduced in the nearest neighbor rule in a high dimensional input feature space with finite samples. As such, the choice of a distance measure becomes crucial in determining the outcome of nearest neighbor classification. The commonly used Euclidean distance implies that the input space is isotropic, which is often invalid and generally undesirable in many practical applications. Several techniques [9, 10, 7] have been proposed to try to minimize bias in high dimensions by using locally adaptive mechanisms. The "lazy learning" approach used by these methods, while appealing in many ways, requires a considerable amount of on-line computation, which makes it difficult for such techniques to scale up to large data sets. The feature weighting scheme they introduce, in fact, is query based and is applied on-line when the test point is presented to the "lazy learner" . In this paper we propose a locally adaptive metric classification method which, although still founded on a query based weighting mechanism, computes off-line the information relevant to define local weights. Our technique uses support vector machines (SVMs) as a guidance for the process of defining a local flexible metric. SVMs have been successfully used as a classification tool in a variety of areas [11, 3, 14], and the maximum margin boundary they provide has been proved to be optimal in a structural risk minimization sense. The solid theoretical foundations that have inspired SVMs convey desirable computational and learning theoretic properties to the SVM's learning algorithm, and therefore SVMs are a natural choice for seeking local discriminant directions between classes. The solution provided by SVMs allows to determine locations in input space where class conditional probabilities are likely to be not constant, and guides the extraction of local information in such areas. This process produces highly stretched neighborhoods along boundary directions when the query is close to the boundary. As a result, the class conditional probabilities tend to be constant in the modified neighborhoods, whereby better classification performance can be achieved. The amount of elongation-constriction decays as the query moves further from the boundary vicinity. 2 Feature Weighting SVMs classify patterns according to the sign(f(x)), where f(x) L:~=l (XiYiK(Xi, x) - b, K(x,y) = cpT(x). cp(y) (kernel junction), and cp: 3(n -+ 3(N is a mapping of the input vectors into a higher dimensional feature space. Here we assume Xi E 3(n, i = I, . . . ,l, and Yi E {-I,I}. Clearly, in the general case of a non-linear feature mapping cp, the SVM classifier gives a non-linear boundary f(x) = 0 in input space. The gradient vector lld = "Vdj, computed at any point d of the level curve f(x) = 0, gives the perpendicular direction to the decision boundary in input space at d. As such, the vector lld identifies the orientation in input space on which the projected training data are well separated, locally over d's neighborhood. Therefore, the orientation given by lld, and any orientation close to it, is highly informative for the classification task at hand, and we can use such information to define a local measure of feature relevance. Let q be a query point whose class label we want to predict. Suppose q is close to the boundary, which is where class conditional probabilities become locally non uniform, and therefore estimation of local feature relevance becomes crucial. Let d be the closest point to q on the boundary f(x) = 0: d = argminp Ilq - pll, subject to the constraint f(p) = O. Then we know that the gradient lld identifies a direction along which data points between classes are well separated. As a consequence, the subspace spanned by the orientation lld, locally at q, is likely to contain points having the same class label as q . Therefore, when applying a nearest neighbor rule at q, we desire to stay close to q along the lld direction, because that is where it is likely to find points similar to q in terms of class posterior probabilities. Distances should be constricted (large weight) along lld and along directions close to it. The farther we move from the lld direction, the less discriminant the correspondent orientation becomes. This means that class labels are likely not to change along those orientations, and distances should be elongated (small weight), thus including in q's neighborhood points which are likely to be similar to q in terms of the class conditional probabilities. Formally, we can measure how close a direction t is to lld by considering the dot product lla ·t. In particular, by denoting with Uj the unit vector along input feature j, for j = 1, . .. , n, we can define a measure of relevance for feature j, locally at q (and therefore at d), as Rj(q) == Iu] . lldl = Ind,j l, where lld = (nd,l,'" ,nd,n)T. The measure of feature relevance, as a weighting scheme, can then be given by the following exponential weighting scheme: Wj(q) = exp(ARj(q))1 2::7=1 exp(ARi(q)), where A is a parameter that can be chosen to maximize (minimize) the influence of R j on Wj' When A = 0 we have Wj = lin, thereby ignoring any difference between the Rj's. On the other hand, when A is large a change in Rj will be exponentially reflected in Wj' The exponential weighting scheme conveys stability to the method by preventing neighborhoods to extend infinitely in any direction. This is achieved by avoiding zero weights, which would instead be allowed by linear or quadratic weightings. Thus, the exponential weighting scheme can be used as weights associated with features for weighted distance computation D(x,y) = )2::7=1 Wi(Xi - Yi)2. These weights enable the neighborhood to elongate less important feature dimensions, and, at the same time, to constrict the most influential ones. Note that the technique is query-based because weightings depend on the query. 3 Local Flexible Metric Classification based on SVMs To estimate the orientation of local boundaries, we move from the query point along the input axes at distances proportional to a given small step (whose initial value can be arbitrarily small, and doubled at each iteration till the boundary is crossed). We stop as soon as the boundary is crossed along an input axis i, i.e. when a point Pi is reached that satisfies the condition sign(f(q)) x sign(f(pi)) = -1. Given Pi, we can get arbitrarily close to the boundary by moving at (arbitrarily) small steps along the segment that joins Pi to q. Let us denote with di the intercepted point on the boundary along direction i. We then approximate lld with the gradient vector lldi = \7 di f, computed at di. We desire that the parameter A in the exponential weighting scheme increases as the distance of q from the boundary decreases. By using the knowledge that support vectors are mostly located around the boundary surface, we can estimate how close a query point q is to the boundary by computing its distance from the closest non bounded support vector: Bq = minsi Ilq - si ll, where the minimum is taken over the non bounded (0 < D:i < C) support vectors Si. Following the same principle, in [1] the spatial resolution around the boundary is increased by enlarging volume elements locally in neighborhoods of support vectors. Then, we can achieve our goal by setting A = D - B q , where D is a constant input parameter of the algorithm. In our experiments we set D equal to the approximated average distance between the training points Xk and the boundary: D = t 2::xk {minsi Ilxk - sill}. If A becomes negative it is set to zero. By doing so the value of A nicely adapts to each query point according to its location with respect to the boundary. The closer q is to the decision boundary, the higher the effect of the Rj's values will be on distances computation. We observe that this principled guideline for setting the parameters of our technique takes advantage of the sparseness representation of the solution provided by the SVM. In fact, for each query point q, in order to compute Bq we only need to consider the support vectors, whose number is typically small compared to the Input: Decision boundary f(x) = a produced by a SVM; query point q and parameter K. 1. Compute the approximated closest point d i to q on the boundary; 2. Compute the gradient vector ndi = \l dJ; 3. Set feature relevance values Rj(q) = Indi,jl for j = 1, . . . ,n; 4. Estimate the distance of q from the boundary as: Bq = minsi Ilq - sill; 5. Set A = D - B q , where D = t EXk {minsi Ilxk - sill}; 6. Set Wj(q) = exp(ARj(q))/ E~=l exp(ARi(q)), for j 1, ... ,n; 7. Use the resulting w for K-nearest neighbor classification at the query point q. Figure 1: The LFM-SVM algorithm total number of training examples. Furthermore, the computation of D's value is carried out once and off-line. The resulting local flexible metric technique based on SVMs (LFM-SVM) is summarized in Figure 1. The algorithm has only one adjustable tuning parameter, namely the number K of neighbors in the final nearest neighbor rule. This parameter is common to all nearest neighbor classification techniques. 4 Experimental Results In the following we compare several classification methods using both simulated and real data. We compare the following classification approaches: (1) LFM-SVM algorithm described in Figure 1. SV Mlight [12] with radial basis kernels is used to build the SVM classifier; (2) RBF-SVM classifier with radial basis kernels. We used SV Mlight [12], and set the value of"( in K(Xi' x) = e-rllxi-xI12 equal to the optimal one determined via cross-validation. Also the value of C for the soft-margin classifier is optimized via cross-validation. The output of this classifier is the input of LFM-SVM; (3) ADAMENN-adaptive metric nearest neighbor technique [7]. It uses the Chi-squared distance in order to estimate to which extent each dimension can be relied on to predict class posterior probabilities; (4) Machete [9]. It is a recursive partitioning procedure, in which the input variable used for splitting at each step is the one that maximizes the estimated local relevance. Such relevance is measured in terms of the improvement in squared prediction error each feature is capable to provide; (5) Scythe [9]. It is a generalization of the machete algorithm, in which the input variables influence each split in proportion to their estimated local relevance; (6) DANN-discriminant adaptive nearest neighbor classification [10]. It is an adaptive nearest neighbor classification method based on linear discriminant analysis. It computes a distance metric as a product of properly weighted within and between sum of squares matrices; (7) Simple K-NN method using the Euclidean distance measure; (8) C4.5 decision tree method [15]. In all the experiments, the features are first normalized over the training data to have zero mean and unit variance, and the test data features are normalized using the corresponding training mean and variance. Procedural parameters (including K) for each method were determined empirically through cross-validation. 4.1 Experiments on Simulated Data For all simulated data, 10 independent training samples of size 200 were generated. For each of these, an additional independent test sample consisting of 200 observations was generated. These test data were classified by each competing method using the respective training data set. Error rates computed over all 2,000 such classifications are reported in Table 1. The Problems. (1) Multi-Gaussians. The data set consists of n = 2 input features, l = 200 training data, and J = 2 classes. Each class contains two spherical bivariate normal subclasses, having standard deviation 1. The mean vectors for one class are (-3/4, -3) and (3/4,3); whereas for the other class are (3, -3) and (-3,3). For each class, data are evenly drawn from each of the two normal subclasses. The first column of Table 1 shows the results for this problem. The standard deviations are: 0.17, 0.01, 0.01, 0.01, 0.01 0.01, 0.01 and 1.50, respectively. (2) Noisy-Gaussians. The data for this problem are generated as in the previous example, but augmented with four predictors having independent standard Gaussian distributions. They serve as noise. Results are shown in the second column of Table 1. The standard deviations are: 0.18, 0.01, 0.02, 0.01, 0.01, 0.01, 0.01 and 1.60, respectively. Results. Table 1 shows that all methods have similar performances for the MultiGaussians problem, with C4.5 being the worst performer. When the noisy predictors are added to the problem (NoisyGaussians), we observe different levels of deterioration in performance among the eight methods. LFM-SVM shows the most robust behavior in presence of noise. K-NN is instead the worst performer. In Figure 2 we plot the performances of LFM-SVM and RBF-SVM as a function of an increasing number of noisy features (for the same MultiGaussians problem). The standard deviations for RBF -SVM (in order of increasing number of noisy features) are: 0.01, 0.01, 0.03, 0.03, 0.03 and 0.03. The standard deviations for LFM-SVM are: 0.17,0.18,0.2,0.3,0.3 and 0.3. The LFM-SVM technique shows a considerable improvement over RBF -SVM as the amount of noise increases. Table 1: Average classification error rates for simulated and real data. MultiGauss NoisyGauss Iris Sonar Liver Vote Breast OQ Pima LFM-SVM 3.3 3.4 4.0 11.0 28.1 2.6 3.0 3.5 19.3 RBF-SVM 3.3 4.1 4.0 12.0 26.1 3.0 3.1 3.4 21.3 ADAMENN 3.4 4.1 3.0 9.1 30.7 3.0 3.2 3.1 20.4 Machete 3.4 4.3 5.0 21.2 27.5 3.4 3.5 7.4 20.4 Scythe 3.4 4.8 4.0 16.3 27.5 3.4 2.7 5.0 20.0 DANN 3.7 4.7 6.0 1.1 30.1 3.0 2.2 4.0 22.2 K-NN 3.3 7.0 6.0 12.5 32.5 7.8 2.7 5.4 24.2 C4.5 5.0 5.1 8.0 23.1 38.3 3.4 4.1 9.2 23.8 4.2 Experiments on Real Data In our experiments we used seven different real data sets. They are all taken from DCI Machine Learning Repository at http://www.cs.uci.edu/,,,-,mlearn/ MLRepository.html. For a description of the data sets see [6]. For the Iris, Sonar, Liver and Vote data we perform leave-one-out cross-validation to measure performance, since the number of available data is limited for these data sets. For the 36'--'--'---r--'--~--'--'--~--.--'--~ 34 32 30 28 26 24 22 20 18 16 14 12 10 8 6 ~ ~~=='P'LFM-SVM --+-RBF-SVM ---)(--O L-~--~--~~--~--~~--~--~~--~ o 10 12 14 16 18 20 22 Number of Noisy Variables Figure 2: Average Error Rates of LFM-SVM and RBF-SVM as a function of an increasing number of noisy predictors. i J. T I I I - 1~ • • -~ ""'!"" :E :E z " i z z 3 z 1j z z ~ > OJ "" ;l :z :E ~ Q '" "' "" "" " Q ..J "" Figure 3: Performance distributions for real data. Breast, OQ-Ietter and Pima data we randomly generated five independent training sets of size 200. For each of these, an additional independent test sample consisting of 200 observations was generated. Table 1 (columns 3-9) shows the cross-validated error rates for the eight methods under consideration on the seven real data. The standard deviation values are as follows. Breast data: 0.2, 0.2, 0.2, 0.2, 0.2, 0.9, 0.9 and 0.9, respectively. OQ data: 0.2, 0.2, 0.2, 0.3, 0.2, 1.1, 1.5 and 2.1, respectively. Pima data: 0.4, 0.4, 0.4, 0.4, 0.4, 2.4, 2.1 and 0.7, respectively. Results. Table 1 shows that LFM-SVM achieves the best performance in 2/7 of the real data sets; in one case it shows the second best performance, and in the remaining four its error rate is still quite close to the best one. Following Friedman [9], we capture robustness by computing the ratio bm of the error rate em of method m and the smallest error rate over all methods being compared in a particular example: bm = emf minl~k~8 ek· Figure 3 plots the distribution of bm for each method over the seven real data sets. The dark area represents the lower and upper quartiles of the distribution that are separated by the median. The outer vertical lines show the entire range of values for the distribution. The spread of the error distribution for LFM-SVM is narrow and close to one. The results clearly demonstrate that LFM-SVM (and ADAMENN) obtained the most robust performance over the data sets. The poor performance of the machete and C4.5 methods might be due to the greedy strategy they employ. Such recursive peeling strategy removes at each step a subset of data points permanently from further consideration. As a result, changes in an early split, due to any variability in parameter estimates, can have a significant impact on later splits, thereby producing different terminal regions. This makes predictions highly sensitive to the sampling fluctuations associated with the random nature of the process that produces the traning data, thus leading to high variance predictions. The scythe algorithm, by relaxing the winner-take-all splitting strategy of the machete algorithm, mitigates the greedy nature of the approach, and thereby achieves better performance. In [10], the authors show that the metric employed by the DANN algorithm approximates the weighted Chi-squared distance, given that class densities are Gaussian and have the same covariance matrix. As a consequence, we may expect a degradation in performance when the data do not follow Gaussian distributions and are corrupted by noise, which is likely the case in real scenarios like the ones tested here. We observe that the sparse solution given by SVMs provides LFM-SVM with principled guidelines to efficiently set the input parameters. This is an important advantage over ADAMENN, which has six tunable input parameters. Furthermore, LFM-SVM speeds up the classification process since it applies the nearest neighbor rule only once, whereas ADAMENN applies it at each point within a region centered at the query. We also observe that the construction of the SVM for LFM-SVM is carried out off-line only once, and there exist algorithmic and computational results which make SVM training practical also for large-scale problems [12]. The LFM-SVM offers performance improvements over the RBF-SVM algorithm alone, for both the (noisy) simulated and real data sets. The reason for such performance gain may rely on the effect of our local weighting scheme on the separability of classes, and therefore on the margin, as shown in [6]. Assigning large weights to input features close to the gradient direction, locally in neighborhoods of support vectors, corresponds to increase the spatial resolution along those orientations, and therefore to improve the separability of classes. As a consequence, better classification results can be achieved as demonstrated in our experiments. 5 Related Work In [1], Amari and Wu improve support vector machine classifiers by modifying kernel functions. A primary kernel is first used to obtain support vectors. The kernel is then modified in a data dependent way by using the support vectors: the factor that drives the transformation has larger values at positions close to support vectors. The modified kernel enlarges the spatial resolution around the boundary so that the separability of classes is increased. The resulting transformation depends on the distance of data points from the support vectors, and it is therefore a local transformation, but is independent of the boundary's orientation in input space. Likewise, our transformation metric depends, through the factor A, on the distance of the query point from the support vectors. Moreover, since we weight features, our metric is directional, and depends on the orientation of local boundaries in input space. This dependence is driven by our measure of feature relevance, which has the effect of increasing the spatial resolution along discriminant directions around the boundary. 6 Conclusions We have described a locally adaptive metric classification method and demonstrated its efficacy through experimental results. The proposed technique offers performance improvements over the SVM alone, and has the potential of scaling up to large data sets. It speeds up, in fact, the classification process by computing offline the information relevant to define local weights, and by applying the nearest neighbor rule only once. Acknowledgments This research has been supported by the National Science Foundation under grants NSF CAREER Award 9984729 and NSF IIS-9907477, by the US Department of Defense, and a research award from AT&T. References [1] S. Amari and S. Wu, "Improving support vector machine classifiers by modifying kernel functions", Neural Networks, 12, pp. 783-789, 1999. [2] R.E. Bellman, Adaptive Control Processes. Princeton Univ. Press, 1961. [3] M. Brown, W. Grundy, D. Lin, N. Cristianini, C. Sugnet, T. Furey, M. Ares, and D. Haussler, "Knowledge-based analysis of microarray gene expressions data using support vector machines", Tech. Report, University of California in Santa Cruz, 1999. [4] W.S. Cleveland and S.J. Devlin, "Locally Weighted Regression: An Approach to Regression Analysis by Local Fitting", J. Amer. Statist. Assoc. 83, 596-610, 1988 [5] T.M. Cover and P.E. Hart, "Nearest Neighbor Pattern Classification", IEEE Trans. on Information Theory, pp. 21-27, 1967. [6] C. Domeniconi and D. Gunopulos, "Adaptive Nearest Neighbor Classification using Support Vector Machines", Tech. Report UCR-CSE-01-04, Dept. of Computer Science, University of California, Riverside, June 200l. [7] C. Domeniconi, J. Peng, and D. Gunopulos, "An Adaptive Metric Machine for Pattern Classification", Advances in Neural Information Processing Systems, 2000. [8] R.O. Duda and P.E. Hart, Pattern Classification and Scene Analysis. John Wiley & Sons, Inc., 1973. [9] J.H. Friedman "Flexible Metric Nearest Neighbor Classification", Tech. Report, Dept. of Statistics, Stanford University, 1994. [10] T. Hastie and R. Tibshirani, "Discriminant Adaptive Nearest Neighbor Classification", IEEE Trans. on Pattern Analysis and Machine Intelligence, Vol. 18, No.6, pp. 607-615, 1996. [11] T. Joachims, "Text categorization with support vector machines", Pmc. of European Conference on Machine Learning, 1998. [12] T. Joachims, "Making large-scale SVM learning practical" Advances in Kernel Methods - Support Vector Learning, B. Sch6lkopf and C. Burger and A. Smola (ed.), MITPress, 1999. http://www-ai.cs.uni-dortmund.de/thorsten/svm_light.html [13] D.G. Lowe, "Similarity Metric Learning for a Variable-Kernel Classifier", Neural Computation 7(1):72-85, 1995. [14] E. Osuna, R. Freund, and F. Girosi, "Training support vector machines: An application to face detection", Pmc. of Computer Vision and Pattern Recognition, 1997. [15] J.R. Quinlan, C4.5: Programs for Machine Learning. Morgan-Kaufmann Publishers, Inc., 1993. [16] C.J. Stone, Nonparametric regression and its applications (with discussion). Ann. Statist. 5, 595, 1977.
|
2001
|
109
|
1,916
|
Probabilistic Inference of Hand Motion from Neural Activity in Motor Cortex Y. Gao M. J. Black E. Bienenstock S. Shoham J. P. Donoghue Division of Applied Mathematics, Brown University, Providence, RI 02912 Dept. of Computer Science, Brown University, Box 1910, Providence, RI 02912 Princeton University, Dept. of Molecular Biology Princeton, NJ, 08544 Dept. of Neuroscience, Brown University, Providence, RI 02912 gao@cfm.brown.edu, black@cs.brown.edu, elie@dam.brown.edu, sshoham@princeton.com, john donoghue@brown.edu Abstract Statistical learning and probabilistic inference techniques are used to infer the hand position of a subject from multi-electrode recordings of neural activity in motor cortex. First, an array of electrodes provides training data of neural firing conditioned on hand kinematics. We learn a nonparametric representation of this firing activity using a Bayesian model and rigorously compare it with previous models using cross-validation. Second, we infer a posterior probability distribution over hand motion conditioned on a sequence of neural test data using Bayesian inference. The learned firing models of multiple cells are used to define a nonGaussian likelihood term which is combined with a prior probability for the kinematics. A particle filtering method is used to represent, update, and propagate the posterior distribution over time. The approach is compared with traditional linear filtering methods; the results suggest that it may be appropriate for neural prosthetic applications. 1 Introduction This paper explores the use of statistical learning methods and probabilistic inference techniques for modeling the relationship between the motion of a monkey’s arm and neural activity in motor cortex. Our goals are threefold: (i) to investigate the nature of encoding in motor cortex, (ii) to characterize the probabilistic relationship between arm kinematics (hand position or velocity) and activity of a simultaneously recorded neural population, and (iii) to optimally reconstruct (decode) hand trajectory from population activity to smoothly control a prosthetic robot arm (cf [14]). A multi-electrode array (Figure 1) is used to simultaneously record the activity of 24 neurons in the arm area of primary motor cortex (MI) in awake, behaving, macaque monkeys. This activity is recorded while the monkeys manually track a smoothly and randomly mov
C. ! !" # $ % & ' ( ) ) * + ,.-0/1 2 3 4 4 4 4 5 564 White Matter Connector Acrylic Bone Silicone Figure 1: Multi-electrode array. A. 10X10 matrix of electrodes. Separation 400 7 m (size 4X4 mm). B. Location of array in the MI arm area. C. Illustration of implanted array (courtesy N. Hatsopoulos). ing visual target on a computer monitor [12]. Statistical learning methods are used to derive Bayesian estimates of the conditional probability of firing for each cell given the kinematic variables (we consider only hand velocity here). Specifically, we use non-parametric models of the conditional firing, learned using regularization (smoothing) techniques with cross-validation. Our results suggest that the cells encode information about the position and velocity of the hand in space. Moreover, the non-parametric models provide a better explanation of the data than previous parametric models [6, 10] and provide new insight into neural coding in MI. Decoding involves the inference of the hand motion from the firing rate of the cells. In particular, we represent the posterior probability of the entire hand trajectory conditioned on the observed sequence of neural activity (spike trains). The nature of this activity results in ambiguities and a non-Gaussian posterior probability distribution. Consequently, we represent the posterior non-parametrically using a discrete set of samples [8]. This distribution is predicted and updated in non-overlapping 50 ms time intervals using a Bayesian estimation method called particle filtering [8]. Experiments with real and synthetic data suggest that this approach provides probabilistically sound estimates of kinematics and allows the probabilistic combination of information from multiple neurons, the use of priors, and the rigorous evaluation of models and results. 2 Methods: Neural Recording The design of the experiment and data collection is described in detail in [12]. Summarizing, a ten-by-ten array of electrodes is implanted in the primary motor cortex (MI) of a Macaque monkey (Figure 1) [7, 9, 12]. Neural activity in motor cortex has been shown to be related to the movement kinematics of the animal’s arm and, in particular, to the direction of hand motion [3, 6]. Previous behavioral tasks have involved reaching in one of a fixed number of directions [3, 6, 14]. To model the relationship between continuous, smooth, hand motion and neural activity, we use a more complex scenario where the monkey performs a continuous tracking task in which the hand is moved on a 2D tablet while holding a low-friction manipulandum that controls the motion of a feedback dot viewed on a computer monitor (Figure 2a) [12]. The monkey receives a reward upon completion of a successful trial in which the manipulandum is moved to keep the feedback dot within a pre-specified distance of the target. The path of the target is chosen to be a smooth random walk that effectively samples the space of hand positions and velocities: measured hand positions and velocities have a roughly Gaussian distribution (Figure 2b and c) [12]. Neural activity is amplified, waveforms are thresholded, and spike sorting is performed off-line to isolate the activity of individual cells [9]. Recordings from 24 motor cortical cells are measured simultaneously with hand kinematics. Monitor Tablet Manipulandum Trajectory Target 0 5 10 15 20 25 0 2 4 6 8 10 12 14 16 a b c Figure 2: Smooth tracking task. (a) The target moves with a smooth random walk. Distribution of the position (b) and velocity (c) of the hand. Color coding indicates the frequency with which different parts of the space are visited. (b) Position: horizontal and vertical axes represent and position of the hand. (c) Velocity: the horizontal axis represents direction, , and the vertical axis represents speed, . 0 0.5 1 1.5 2 2.5 3
cell 3 cell 16 cell 19 Figure 3: Observed mean conditional firing rates in 50 ms intervals for three cells given hand velocity. The horizontal axis represents the direction of movement, , in radians (“wrapping” around from to ). The vertical axis represents speed, , and ranges from 0 cm/s to 12 cm/s. Color ranges from dark blue (no measurement) to red (approximately 3 spikes). 3 Modeling Neural Activity Figure 3 shows the measured mean firing rate within 50 ms time intervals for three cells conditioned on the subject’s hand velocity. We view the neural firing activity in Figure 3 as a stochastic and sparse realization of some underlying model that relates neural firing to hand motion. Similar plots are obtained as a function of hand position. Each plot can be thought of as a type of “tuning function” [12] that characterizes the response of the cell conditioned on hand velocity. In previous work, authors have considered a variety of models of this data including a cosine tuning function [6] and a modified cosine function [10]. Here we explore a non-parametric model of the underling activity and, adopting a Bayesian formulation, seek a maximum a posterior (MAP) estimate of a cell’s conditional firing. Adopting a Markov Random Field (MRF) assumption [4], let the velocity space, ! "!#%$ , be discretized on a &(')' *+&(')' grid. Let g be the array of true (unobserved) conditional neural firing and , be the corresponding observed mean firing. We seek the posterior probability -/. g 01,1235476 .98:-/.9; 6<0>=?6@2A4CBDFE/G -/. =?6<0H=?6?IJ2K2 (1) −3 −2 −1 0 1 2 3 0 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 −3 −2 −1 0 1 2 3 −12 −10 −8 −6 −4 −2 0 a b Figure 4: Prior probability of firing variation ( = ). (a) Probability of firing variation computed from training data (blue). Proposed robust prior model (red) plotted for ' . (b) Logarithm of the distributions shown to provide detail. where 8 is a normalizing constant independent of g, ; 6 and = 6 are the observed and true mean firing at velocity respectively, = 6 I represents the firing rate for the th neighboring velocity of , and the neighbors are taken to be the four nearest velocities ( ). The first term on the right hand side represents the likelihood of observing a particular firing rate ; 6 given that the true rate is =6 . Here we compare two generative models of the neural spiking process within 50 ms; a Poisson model, , and a Gaussian model,
: . ; 0>= 2 & ; =
.9; 0H= 2 & ? . ; =A2 ! The second term is a spatial prior probability that encodes our expectations about = , the variation of neural activity in velocity space. The MRF prior states that the firing, =?6 , at velocity depends only on the firing at neighboring velocities. We consider two possible prior models for the distribution of = : Gaussian and “robust”. A Gaussian prior corresponds to an assumption that the firing rate varies smoothly. A robust prior assumes a heavy-tailed distribution of the spatial variation (see Figure 4), = , (derivatives of the firing rate in the and directions) and implies piecewise smooth data. The two spatial priors are -#" . = 2 %$ . '& = 2
. = 2 & ? . = 2( )! The various models (cosine, a modified cosine (Moran and Schwartz [10]), Gaussian+Gaussian, and Poisson+Robust) are fit to the training data as shown in Figure 5. G In the case of the Gaussian+Gaussian and Poisson+Robust models, the optimal value of the parameter is computed for each cell using cross validation. During cross-validation, each time 10 trials out of 180 are left out for testing and the models are fit with the remaining training data. We then compute the log likelihood of the test data given the model. This provides a measure of how well the model captures the statistical variation in the training set and is used for quantitative comparison. The whole procedure is repeated 18 times for different test data sets. The solution to the Gaussian+Gaussian model can be computed in closed form but for the Poisson+Robust model no closed form solution for g exists and an optimal Bayesian estimate could be achieved with simulated annealing [4]. Instead, we derive an approximate * By “Gaussian+Gaussian” we mean both the likelihood and prior terms are Gaussian whereas “Poisson+Robust”implies a Poisson likelihood and robust spatial prior. 0.2 0.4 0.6 0.8 1 1.2 0.5
1
1.5
0.5
1
0.2
0.4
0.6
0.8
1
1.2
0.5
1
1.5
0.2
0.4
0.6
0.8
0.4
0.6
0.8
0.7
0.8
0.9
0.65
0.7
0.75
0.3 0.5 0.7 0.6 0.8 1 0.5 0.6 0.7 0.8 Cosine Moran&Schwartz (M&S) Gaussian+Gaussian Poisson+Robust cell 3 cell 16 cell 19 Figure 5: Estimated firing rate for cells in Figure 3 using different models. Method: Log Likelihood Ratio p-value G+G over Cosine 24.9181 7.6294e-06 G+G over M&S 15.8333 0.0047 P+R over Cosine 50.0685 7.6294e-06 P+R over M&S 32.2218 7.6294e-06 Table 1: Numerical comparison; log likelihood ratio of pairs of models and the significance level given by Wilcoxon signed rank test (Splus, MathSoft Inc., WA). solution for g in (1) by minimizing the negative logarithm of the distribution using standard regularization techniques [1, 13] with missing data, the learned prior model, and a Poisson likelihood term [11]. Simple gradient descent [1] with deterministic annealing provides a reasonable solution. Note that the negative logarithm of the prior term can be approximated by the robust statistical error function . =A2 = . & . =A2 12 which has been used extensively in machine vision and image processing for smoothing data with discontinuities [1, 5]. Figure 5 shows the various estimates of the receptive fields. Observe that the pattern of firing is not Gaussian. Moreover, some cells appear to be tuned to motion direction, , and not to speed, , resulting in vertically elongated patterns of firing. Other cells (e.g. cell 19) appear to be tuned to particular directions and speeds; this type of activity is not well fit by the parametric models. Table 1 shows a quantitative comparison using cross-validation. The log likelihood ratio (LLR) is used to compare each pair of models: LLR(model 1, model 2) = log( (observed firing 0 model 1)/Pr(observed firing 0 model 2)). The positive values in Table 1 indicate that the non-parametric models do a better job of explaining new data than the parametric models with the Poisson+Robust fit providing the best description of the data. This P+R model implies that the conditional firing rate is well described by regions of smooth activity with relatively sharp discontinuities between them. The non-parametric models reduce the strong bias of the parametric models with a slight increase in variance and hence achieve a lower total error. 4 Temporal Inference Given neural measurements our goal is to infer the motion of the hand over time. Related approaches have exploited simple linear filtering methods which do not provide a probabilistic interpretation of the data that can facilitate analysis and support the principled combination of multiple sources of information. Related probabilistic approaches have exploited Kalman filtering [2]. We note here however, that the learned models of neural activity are not-Gaussian and the dynamics of the hand motion may be non-linear. Furthermore with a small number of cells, our interpretation of the neural data may be ambiguous and the posterior probability of the kinematic variables, given the neural activity, may be best modeled by a non-Gaussian, multi-modal, distribution. To cope with these issues in a sound probabilistic framework we exploit a non-parametric approach that uses factored sampling to discretely approximate the posterior distribution, and particle filtering to propagate and update this distribution over time [8]. Let the state of the system be s 7 ! ?# at time . Let D be the mean firing rate of cell at time where the mean firing rate is estimated within non-overlapping 50 ms temporal windows. Also, let c G # represent the firing rate of all cells at time . Similarly let D represent the sequence of these firing rates for cell up to time and let C G # represent the firing of all cells up to time . We assume that the temporal dynamics of the states, s , form a Markov chain for which the state at time depends only on the state at the previous time instant: -/. s 0 S G 2 . s 0 s G 2 where S s F s
(# denotes the state history. We also assume that given s , the current observation c and the previous observations C G are independent. Using Bayes rule and the above assumptions, the probability of observing the state at time given the history of firing can be written as . s H0 C J2 8 . c 0 s "2 -/. s H0 C G 2 (2) where 8 is a normalizing term that insures that the distribution integrates to one. The likelihood term -/. c 0 s 2 DFE/G . D 0 s 2 assumes conditional independence of the individual cells where the likelihood for the firing rate of an individual cell is taken to be a Poisson distribution with the mean firing rate for the speed and velocity given by s determined by the conditional firing models learned in the previous section. Plotting this likelihood term for a range of states reveals that its structure is highly non-Gaussian with multiple peaks. The temporal prior term, . s 0 C G 2 can be written as -/. s H0 C G 2 -/. s H0 s G 2 -/. s G 0 C G 2 s G (3) where . s H0 s G 2 embodies the temporal dynamics of the hand velocity which are assumed to be constant with Gaussian noise; that is, a diffusion process. Note, . s G 0 C G 2 is the posterior distribution over the state space at time & . The posterior, . s 0 C J2 , is represented with a discrete, weighted set, of ')'' random samples which are propagated in time using a standard particle filter (see [8] for details). Unlike previous applications of particle filtering, the likelihood of firing for an individual cell in 125 126 127 128 129 130 131 132 133 134 135 -10 -5 0 5 10 trial No. 8, Vx in cm/s, blue:true, red:reconstruction time in second 125 126 127 128 129 130 131 132 133 134 -10 -5 0 5 10 Vy in cm/s 125 126 127 128 129 130 131 132 133 134 135 -10 -5 0 5 10 trial No. 8, Vx in cm/s, blue:true, red:reconstruction time in second 125 126 127 128 129 130 131 132 133 134 135 -10 -5 0 5 10 Vy in cm/s a b Figure 6: Tracking results using 1008 synthetic cells showing horizontal velocity, , (top) and vertical velocity, , (bottom). Blue indicates true velocity of hand. (a) Bayesian estimate using particle filtering. Red curve shows expected value of the posterior. The error is ' for and ' ' for . (b) Linear filtering method shown in red; ' for and ' ' for . 50 ms provides very little information. For the posterior to be meaningful we must combine evidence from multiple cells. Our experiments indicate that the responses from our 24 cells are insufficient for this task. To demonstrate the feasibility of the particle filtering method, we synthesized approximately 1000 cells by taking the learned models of the 24 cells and translating them along the axis to generate a more complete covering of the velocity space. Note that the assumption of such a set of cells in MI is quite reasonable give the sampling of cells we have observed in multiple monkeys. From the set of synthetic cells we then generate a synthetic spike train by taking a known sequence of hand velocities and stochastically generating spikes using the learned conditional firing models with a Poisson generative model. Particle filtering is used to estimate the posterior distribution over hand velocities given the synthetic neural data. The expected value of the horizontal and vertical velocity is displayed in Figure 6a. For comparison, a standard linear filtering method [6, 14] was trained on the synthetic data from 50 ms intervals. The resulting prediction is shown in Figure 6b. Linear filtering works well over longer time windows which introduce lag. The Bayesian analysis provides a probabilistic framework for sound causal estimates over short time intervals. We are currently experimenting with modified particle filtering schemes in which linear filtering methods provide a proposal distribution and importance sampling is used to construct a valid posterior distribution. We are also comparing these results with those of various Kalman filters. 5 Conclusions We have described a Bayesian model for neural activity in MI that relates this activity to actions in the world. Quantitative comparison with previous models of MI activity indicate that the non-parametric models computed using regularization more accurately describe the neural activity. In particular, the robust spatial prior term suggests that neural firing in MI is not a smooth function of velocity but rather exhibits discontinuities between regions of high and low activity. We have also described the Bayesian decoding of hand motion from firing activity using a particle filter. Initial results suggest that measurements from several hundred cells may be required for accurate estimates of hand velocity. The application of particle filtering to this problem has many advantages as it allows complex, non-Gaussian, likelihood models that may incorporate non-linear temporal properties of neural firing (e.g. refractory period). Unlike previous linear filtering methods this Bayesian approach provides probabilistically sound, causal, estimates in short time windows of 50ms. Current work is exploring correlations between cells [7] and the relationship between the neural activity and other kinematic variables [12]. Acknowledgments. This work was supported by the Keck Foundation and by the National Institutes of Health under grants #R01 NS25074 and #N01-NS-9-2322 and by the National Science Foundation ITR Program award #0113679. We are very grateful to M. Serruya, M. Fellows, L. Paninski, and N. Hatsopoulos who provided the neural data and valuable insight. References [1] M. Black and A. Rangarajan. On the unification of line processes, outlier rejection, and robust statistics with applications in early vision. IJCV, 19(1):57–92, 1996. [2] E. Brown, L. Frank, D. Tang, M. Quirk, and M. Wilson. A statistical paradigm for neural spike train decoding applied to position prediction from ensemble firing patterns of rat hippocampal place cells. J. Neuroscience, 18(18):7411–7425, 1998. [3] Q-G. Fu, D. Flament, J. Coltz, and T. Ebner. Temporal encoding of movement kinematics in the discharge of primate primary motor and premotor neurons. J. of Neurophysiology, 73(2):836– 854, 1995. [4] S. Geman and D. Geman. Stochastic relaxation, Gibbs distributions and Bayesian restoration of images. PAMI, 6(6):721–741, November 1984. [5] S. Geman and D. McClure. Statistical methods for tomographic image reconstruction. Bulletin of the Int. Stat. Inst., LII-4:5–21, 1987. [6] A. Georgopoulos, A. Schwartz, and R. Kettner. Neuronal population coding of movement direction. Science, 233:1416–1419, 1986. [7] N. Hatsopoulos, C. Ojakangas, L. Paninski, and J. Donoghue. Information about movement direction obtained from synchronous activity of motor cortical neurons. Proc. Nat. Academy of Sciences, 95:15706–15711, 1998. [8] M. Isard and A. Blake. Condensation – conditional density propagation for visual tracking. IJCV, 29(1): 5–28, 1998. [9] E. Maynard, N. Hatsopoulos, C. Ojakangas, B. Acuna, J. Sanes, R. Normann, and J. Donoghue. Neuronal interaction improve cortical population coding of movement direction. J. of Neuroscience, 19(18):8083–8093, 1999. [10] D. Moran and A. Schwartz. Motor cortical representation of speed and direction during reaching. J. Neurophysiol, 82:2676-2692, 1999. [11] R. Nowak and E. Kolaczyk. A statistical multiscale framework for Poisson inverse problems. IEEE Inf. Theory, 46(5):1811–1825, 2000. [12] L. Paninski, M. Fellows, N. Hatsopoulos, and J. Donoghue. Temporal tuning properties for hand position and velocity in motor cortical neurons. submitted, J. Neurophysiology, 2001. [13] D. Terzopoulos. Regularization of inverse visual problems involving discontinuities. PAMI, 8(4):413–424, 1986. [14] J. Wessberg, C. Stambaugh, J. Kralik, P. Beck, M. Laubach, J. Chapin, J. Kim, S. Biggs, M. Srinivasan, and M. Nicolelis. Real-time prediction of hand trajectory by ensembles of cortical neurons in primates. Nature, 408:361–365, 2000.
|
2001
|
11
|
1,917
|
Fast, large-scale transformation-invariant clustering Brendan J. Frey Machine Learning Group University of Toronto www.psi.toronto.edu/∼frey Nebojsa Jojic Vision Technology Group Microsoft Research www.ifp.uiuc.edu/∼jojic Abstract In previous work on “transformed mixtures of Gaussians” and “transformed hidden Markov models”, we showed how the EM algorithm in a discrete latent variable model can be used to jointly normalize data (e.g., center images, pitch-normalize spectrograms) and learn a mixture model of the normalized data. The only input to the algorithm is the data, a list of possible transformations, and the number of clusters to find. The main criticism of this work was that the exhaustive computation of the posterior probabilities over transformations would make scaling up to large feature vectors and large sets of transformations intractable. Here, we describe how a tremendous speed-up is acheived through the use of a variational technique for decoupling transformations, and a fast Fourier transform method for computing posterior probabilities. For N ×N images, learning C clusters under N rotations, N scales, N x-translations and N y-translations takes only (C + 2 log N)N 2 scalar operations per iteration. In contrast, the original algorithm takes CN 6 operations to account for these transformations. We give results on learning a 4-component mixture model from a video sequence with frames of size 320×240. The model accounts for 360 rotations and 76,800 translations. Each iteration of EM takes only 10 seconds per frame in MATLAB, which is over 5 million times faster than the original algorithm. 1 Introduction The task of clustering raw data such as video frames and speech spectrograms is often obfuscated by the presence of random, but well-understood transformations in the data. Examples of these transformations include object motion and camera motion in video sequences and pitch modulation in spectrograms. The machine learning community has proposed a variety of sophisticated techniques for pattern analysis and pattern classification, but these techniques have mostly assumed the data is already normalized (e.g., the patterns are centered in the images) or nearly normalized. Linear approximations to the transformation manifold have been used to significantly improve the performance of feedforward discriminative classifiers such as nearest neighbors and multilayer perceptrons (Simard, LeCun and Denker 1993). Linear generative models (factor analyzers, mixtures of factor analyzers) have also been modified using linear approximations to the transformation manifold to build in some degree of transformation invariance (Hinton, Dayan and Revow 1997). A multi-resolution approach can be used to extend the usefulness of linear approximations (Vasconcelos and Lippman 1998), but this approach is susceptable to local minima – e.g. a pie may be confused for a face at low resolution. For significant levels of transformation, linear approximations are far from exact and better results can be obtained by explicitly considering transformed versions of the input. This approach has been used to design “convolutional neural networks” that are invariant to translations of parts of the input (LeCun et al. 1998). In previous work on “transformed mixtures of Gaussians” (Frey and Jojic 2001) and “transformed hidden Markov models” (Jojic et al. 2000), we showed how the EM algorithm in a discrete latent variable model can be used to jointly normalize data (e.g., center video frames, pitch-normalize spectrograms) and learn a mixture model of the normalized data. We found “that the algorithm is reasonably fast (it learns in minutes or hours) and very effective at transformation-invariant density modeling.” Those results were for 44 × 28 images, but realistic applications such as home video summarization require near-real-time processing of medium-quality video at resolutions near 320 × 240. In this paper, we show how a variational technique and a fast Fourier method for computing posterior probabilities can be used to achieve this goal. 2 Background In (Frey and Jojic 2001), we introduced a single discrete variable that enumerates a discrete set of possible transformations that can occur in the input. Here, we break the transformation into a sequence of transformations. Tk is the random variable for the transformation matrix at step k. So, if Tk is the set of possible transformation matrices corresponding to the type of transformation at step k (e.g., image rotation), Tk ∈Tk. The generative model is shown in Fig. 1a and consists of picking a class c, drawing a vector of image pixel intensities z0 from a Gaussian, picking the first transformation matrix T1 from Tk, applying this transformation to z0 and adding Gaussian noise to obtain z1, and repeating this process until the last transformation matrix TK is drawn from TK and is applied to zK−1 to obtain the observed data zK. The joint distribution is p(c, z0, T1, z1, . . . , TK, zK) = p(c)p(z0|c) K Y k=1 p(Tk)p(zk|zk−1, Tk). (1) The probability of class c ∈{1, . . . , C} is parameterized by p(c) = πc and the untransformed latent image has conditional density p(z0|c) = N(z0; µc, Φc), (2) where N() is the normal distribution, µc is the mean image for class c and Φc is the diagonal noise covariance matrix for class c. Notice that the noise modeled by Φc gets transformed, so Φc can model noise sources that depend on the transformations, such as background clutter and object deformations in images. (c) TK z0 z1 T1 c (b) z z0 z1 T1 TK c (a) K Figure 1: (a) The Bayesian network for a generative model that draws an image z0 from class c, applies a randomly drawn transformation matrix T1 of type 1 (e.g., image rotation) to obtain z1, and so on, until a randomly drawn transformation matrix TK of type K (e.g., image translation) is applied to obtain the observed image zK. (b) The Bayesian network for a factorized variational approximation to the posterior distribution, given zK. (c) When an image is measured on a discrete, radial 2-D grid, a scale and rotation correspond to a shift in the radial and angular coordinates. The probability of transformation matrix Tk at step k is p(Tk) = λk,Tk. (In our experiments, we often fix this to be uniform.) At each step, we assume a small amount of noise with diagonal covariance matrix Ψ is added to the image, so p(zk|zK−1, Tk) = N(zk; Tkzk−1, Ψ). (3) Tk operates on zk−1 to produce a transformed image. In fact, Tk can be viewed as a permutation matrix that rearranges the pixels in zk−1. Usually, we assume Ψ = ψI and in our experiments we often set ψ to a constant, small value, such as 0.01. In (2001), an exact EM algorithm for learning this model is described. The sufficient statistics for πc, µc and Φc are computed by averaging the derivatives of ln(πcN(z0; µc, Φc)) over the posterior distribution, p(c, z0|zK) = X T1 · · · X TK p(z0|c, T1, . . . , TK, zK)p(c, T1, . . . , TK|zK). (4) Since z0, . . . , zK are jointly Gaussian given c and T1, . . . , TK, p(z0|c, T1, . . . , TK, zK) is Gaussian and its mean and covariance are computed using linear algebra. Also, p(c, T1, . . . , TK|zK) is computed using linear algebra. The problem with this direct approach is that the number of scalar operations in (4) is very large for large feature vectors and large sets of transformations. For N × N images, learning C clusters under N rotations, N scales, N x-translations and N y-translations leads to N 4 terms in the summation. Since there are N 2 pixels, each term is computed using N 2 scalar operations. So, each iteration of EM takes CN 6 scalar operations per training case. For 10 classes and images of size 256 × 256, the direct approach takes 2.8 × 1015 scalar operations per image for each iteration of EM. We now describe how a variational technique for decoupling transformations, and a fast Fourier transform method for computing posterior probabilities can reduce the above number to (C + 2 log N)N 2 scalar operations. For 10 classes and images of size 256 × 256, the new method takes 2, 752, 512 scalar operations per image for each iteration of EM. 3 Factorized variational technique To simplify the computation of the required posterior in (4), we use a variational approximation (Jordan et al. 1998). As shown in Fig. 1b, our variational approximation is a completely factorized approximation to the true posterior: p(c, z0, T1, z1, . . . , TK|zK) ≈q(c, z0, T1, z1, . . . , TK) = q(c)q(z0) K−1 Y k=1 q(Tk)q(zk) q(TK). (5) The q-distributions are parameterized and these variational parameters are varied to make the approximation a good one. p(c, z0|zK) ≈q(c)q(zK), so the sufficient statistics can be readily determined from q(c) and q(zK). The variational parameters are q(c) = ρc, q(Tk) = γk,Tk, q(zk) = N(zk; ηk, Ωk). The generalized EM algorithm (Neal and Hinton 1998) maximizes a lower bound on the log-likelihood of the observed image zK: B = X Z q(c, z0, T1, z1, . . . , TK) ln p(c, z0, T1, z1, . . . , TK, zK) q(c, z0, T1, z1, . . . , TK) ≤ln p(zK). (6) In the E step, the variational parameters are adjusted to maximize B and in the M step, the model parameters are adjusted to maximize B. Assuming constant noise, Ψ = ψI, the derivatives of B with respect to the variational parameters produce the following E-step updates: Ω0 ← X c ρcΦ−1 c + ψ−1I −1 η0 ←Ω0 X c ρcΦ−1 c µc + ψ−1 X T1 γ1,T1T−1 1 η1 (7) ρc ←πc exp −1 2tr(Ω0Φ−1 c ) −1 2(η0 −µc)′Φ−1 c (η0 −µc) Ωk ←1 2ψI ηk ←1 2 X Tk γk,TkTkηk−1 + X Tk+1 γk+1,Tk+1T−1 k+1ηk+1 (8) γk,Tk ←λk,Tk exp −1 2tr(Ωkψ−1) −1 2ψ−1(ηk −Tkηk−1)′(ηk −Tkηk−1) . (9) Each time the ρc’s are updated, they should be normalized and similarly for the γk,Tk’s. One or more iterations of the above updates are applied for each training case and the variational parameters are stored for use in the M-step, and as the initial conditions for the next E-step. The derivatives of B with respect to the model parameters produce the following M-step updates: πc ←⟨ρc⟩ µc ←⟨ρcη0⟩ Φc ←⟨ρc(Ω0 + diag((η0 −µc)(η0 −µc)′)⟩, (10) where ⟨⟩indicates an average over the training set. This factorized variational inference technique is quite greedy, since at each step, the method approximates the posterior with one Gaussian. So, the method works best for a small number of steps (2 in our experiments). 4 Inference using fast Fourier transforms The M-step updates described above take very few computations, but the E-step updates can be computationally burdensome. The dominant culprits are the computation of the distance of the form dT = (g −Th)′(g −Th) (11) in (9), for all possible transformations T, and the computation of the form X T γTTh (12) in (7) and (8). Since the variational approximation is more accurate when the transformations are broken into fewer steps, it is a good idea to pack as many transformations into each step as possible. In our experiments, x-y translations are applied in one step, and rotations are applied in another step. However, the number of possible x-y translations in a 320 × 240 image is 76,800. So, 76,800 dT’s must be computed and the computation of each dT uses a vector norm of size 76,800. It turns out that if the data is defined on a coordinate system where the effect of a transformation is a shift, the above quantities can be computed very quickly using fast Fourier transforms (FFTs). For images measured on rectangular grids, an x-y translation corresponds to a shift in the coordinate system. For images measured on a radial grid, such as the one shown in Fig. 1c, a scale and rotation corresponds to a shift in the coordinate system (Wolberg and Zokai 2000). When updating the variational parameters, it is straightforward to convert them to the appropriate coordinate system, apply the FFT method and convert them back. We now use a very different notation to describe the FFT method. The image is measured on a discrete grid and x is the x-y coordinate of a pixel in the image (x is a 2-vector). The images g and h in (11) and (12) are written as functions of x: g(x), h(x). In this representation, T is an integer 2-vector, corresponding to a shift in x. So, (11) becomes d(T) = X x (g(x) −h(x + T))2 = X x (g(x)2 −2g(x)h(x + T) + h(x + T)2) (13) and (12) becomes X T γ(T)h(x + T). (14) The common form is the correlation f(T) = X x g(x)h(x + T), (15) For an N × N grid, computing the correlation directly for all T takes N 4 scalar operations. The FFT can be used to compute the correlation in N 2 log N time. The FFTs G(ω) and H(ω) of g and h are computed in N 2 log N time. Then, the FFT F(ω) of f is computed in N 2 as follows, F(ω) = G(ω)∗H(ω), (16) where “∗” denotes complex conjugate. Then the inverse FFT f(T) of F(ω) is computed in N 2 log N time. Using this method, the posterior and sufficient statistics for all N 2 shifts in an N × N grid can be computed in N 2 log N time. Using this method along with the variational technique, C classes, N scales, N rotations, N x-translations and N y-translations can be accounted for using (C + 2 log N)N 2 scalar operations. 5 Results In order to compare our new learning algorithm with the previously published result, we repeated the experiment on clustering head poses in 200 44x28 frames. We achieved essentially the same result, but in only 10 seconds as opposed to 40 minutes that the original algorithm needed to compete the task. Both algorithms were implemented in Matlab. It should be noted that the original algorithm tested only for 9 vertical and 9 horizontal shifts (81 combinations), while the new algorithm dealt with all 1232 possible discrete shifts. This makes the new algorithm 600 times faster on low resolution data. The speed-up is even more drastic at higher resolutions, and when rotations and scales are added, since the complexity of the original algorithm is CN 6, where C is the number of classes and N is the number of pixels. The speed-up promised in the abstract is based on our computations, but obviously we were not able to run the original algorithm on full 320x240 resolution data. To illustrate that the fast variational technique presented here can be efficiently used to learn data means in the presence of scale change, significant rotations and translations in the data, we captured 10 seconds of a video at 320x240 resolution and trained a two-stage transformation-invariant where the first stage modeled rotations and scales as shifts in the log-polar coordinate system and the second stage modeled all possible shifts as described above. In Fig. 2 we show the results of training an ordinary Gaussian model, shift-invariant model and finally the scale, rotation and shift invariant model on the sequence. We also show three frames from the sequence stabilized using the variational inference. 6 Conclusions We describes how a tremendous speed-up in training transformation-invariant generative model can be achieved through the use of a variational technique for decoupling transformations, and a fast Fourier transform method for computing posterior probabilities. For N × N images, learning C clusters under N rotations, N scales, N x-translations and N y-translations takes only (C +2 log N)N 2 scalar operations per iteration. In contrast, the original algorithm takes CN 6 operations to account for these transformations. In this way we were able to reduce the computation to only seconds per frame for the images of 320x240 resolution using a simple Matlab implementation. This opens the door for generative models of pixel intensities in video to be efficiently used for transformation-invariant video summary and search. As opposed to most techniques used in computer vision today, the generative modeling approach provides the likelihood model useful for search or retrieval, automatic clustering of the data and the extensibility through adding new hidden variables. The model described here could potentially be useful for other high-dimensional data, such as audio. References Dempster, A. P., Laird, N. M., and Rubin, D. B. 1977. Maximum likelihood from incomplete data via the EM algorithm. Proceedings of the Royal Statistical Society, B-39:1–38. Frey, B. J. and Jojic, N. 2001. Transformation invariant clustering and dimensionality reduction. IEEE Transactions on Pattern Analysis and Machine Intelligence. To appear. Available at http://www.cs.utoronto.ca/∼frey. Figure 2: Learning a rotation, scale and translation invariant model on 320x240 video Hinton, G. E., Dayan, P., and Revow, M. 1997. Modeling the manifolds of images of handwritten digits. IEEE Transactions on Pattern Analysis and Machine Intelligence, 8:65–74. Jojic, N., Petrovic, N., Frey, B. J., and Huang, T. S. 2000. Transformed hidden markov models: Estimating mixture models of images and inferring spatial transformations in video sequences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Jordan, M. I., Ghahramani, Z., Jaakkola, T. S., and Saul, L. K. 1998. An introduction to variational methods for graphical models. In Jordan, M. I., editor, Learning in Graphical Models. Kluwer Academic Publishers, Norwell MA. LeCun, Y., Bottou, L., Bengio, Y., and Haffner, P. 1998. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324. Neal, R. M. and Hinton, G. E. 1998. A view of the EM algorithm that justifies incremental, sparse, and other variants. In Jordan, M. I., editor, Learning in Graphical Models, pages 355–368. Kluwer Academic Publishers, Norwell MA. Simard, P. Y., LeCun, Y., and Denker, J. 1993. Efficient pattern recognition using a new transformation distance. In Hanson, S. J., Cowan, J. D., and Giles, C. L., editors, Advances in Neural Information Processing Systems 5. Morgan Kaufmann, San Mateo CA. Vasconcelos, N. and Lippman, A. 1998. Multiresolution tangent distance for affineinvariant classification. In Jordan, M. I., Kearns, M. I., and Solla, S. A., editors, Advances in Neural Information Processing Systems 10. MIT Press, Cambridge MA. Wolberg, G. and Zokai, S. 2000. Robust image registration using log-polar transform. In Proceedings IEEE Intl. Conference on Image Processing, Vancouver, Canada.
|
2001
|
110
|
1,918
|
A Rotation and Translation Invariant Discrete Saliency Network Lance R. Williams Dept. of Computer Science Univ. of New Mexico Albuquerque, NM 87131 John W. Zweck Dept. of CS and EE Univ. of Maryland Baltimore County Baltimore, MD 21250 Abstract We describe a neural network which enhances and completes salient closed contours. Our work is different from all previous work in three important ways. First, like the input provided to V1 by LGN, the input to our computation is isotropic. That is, the input is composed of spots not edges. Second, our network computes a well defined function of the input based on a distribution of closed contours characterized by a random process. Third, even though our computation is implemented in a discrete network, its output is invariant to continuous rotations and translations of the input pattern. 1 Introduction There is a long history of research on neural networks inspired by the structure of visual cortex whose functions have been described as contour completion, saliency enhancement, orientation sharpening, or segmentation[6, 7, 8, 9, 12]. A similiar network has been proposed as a model of visual hallucinations[1]. In this paper, we describe a neural network which enhances and completes salient closed contours. Our work is different from all previous work in three important ways. First, like the input provided to V1 by LGN, the input to our computation is isotropic. That is, the input is composed of spots not edges. Second, our network computes a well defined function of the input based on a distribution of closed contours characterized by a random process. Third, even though our computation is implemented in a discrete network, its output is invariant to continuous rotations and translations of the input pattern. There are two important properties which a computation must possess if it is to be invariant to rotations and translations, i.e., Euclidean invariant. First, the input, the output, and all intermediate representations must be Euclidean invariant. Second, all transformations of these representations must also be Euclidean invariant. The models described in [6, 7, 8, 9, 12] are not Euclidean invariant, first and foremost, because their input representations are not Euclidean invariant. That is, not all rotations and translations of the input can be represented equally well. This problem is often skirted by researchers by choosing input patterns which match particular choices of sampling rate and phase. For example, Li [7] used only six samples in orientation (including ) and Heitger and von der Heydt[5] only twelve (including , and ). Li’s first test pattern was a dashed line of orientation, , while Heitger and von der Heydt used a Kanizsa Triangle with sides of , , and orientation. There is no reason to believe that the experimental results they showed would be similiar if the input patterns were rotated by as little as . To our knowledge, no researcher in this area has ever commented on this problem before. 2 A continuum formulation of the saliency problem The following section reviews the continuum formulation of the contour completion and saliency problem as described in Williams and Thornber[11]. 2.1 Shape distribution Mumford[3] observed that the probability distribution of object boundary shapes could be modeled by a Fokker-Planck equation of the following form:
"! (1) where $#&% $' )( +* is the probability that a particle is located at position, % #,"'-* , and is moving in direction, , at time, . This partial differential equation can be viewed as a set of independent advection equations in and (the first and second terms) coupled in the dimension by the diffusion equation (the third term). The advection equations translate probability mass in direction, , with unit speed, while the diffusion term models the Brownian motion in direction, with diffusion parameter, . The combined effect of these three terms is that particles tend to travel in straight lines, but over time they drift to the left or right by an amount proportional to . Finally, the effect of the fourth term is that particles decay over time, with a half-life given by the decay constant, . 2.2 The propagators The Green’s function, . #&% /' )( 10324% 5 '76 ( 98:* , gives the probability that a particle observed at position, % 5 , and direction, 6 , at time, 8 , will later be observed at position, % , and direction, , at time, 0 . It is the solution, $#&% /' )( 0 * , of the Fokker-Planck initial value problem with initial value, $#&% "' )( 98;* =< #&% % 5 * < # 6>* where < is the Dirac delta function. The Green’s function is used to define two propagators. The long-time propagator: ? 8 #&% /' 2@% 5 '76>* BADC 8FE /GH#I+* . #&% /' )( @24% 5 '16 ( * (2) gives the probability that #&% /' * and #&% 5 '16* are distinct edges from the boundary of a single object. 1 The short-time propagator: ? 0 #:% $' 2@% 5 '16* BA C 8 E KJ GL#,+*NM . #&% /' )( @24% 5 '76 ( * (3) gives the probability that #&% $' * and #:% 5 '16>* are from the boundary of a single object but are really the same edge. In both of these propagators, GH#O!P* is a cut-off function with GL# * and Q RTSIU C GL#,+* : GL#,+* 0 WV XY[Z7Y W\I] V S ^ `_"a&bca ! (4) The cut-off function is characterized by three parameters, _ , ] , and d . The parameter, _ , specifies where the cut-off is and ] specifies how hard hard it is. The parameter, d , is the scale of the edge detection process. 1We assume that the probability that two edges are the same depends only on the distance between them, and that e/fOg,h ikj h l g m@ne/fpo9m for particles travelling at unit speed. 2.3 Eigenfunctions The integral linear operator, #+! * , combines three sources of information: 1) the probability that two edges belong to the same object; 2) the probability that the two edges are distinct; and 3) the probability that the two edges exist. It is defined as follows: #&% /' 24% 5 '76>* #&% >* ? 8#&% /' 2@% 5 '76>* #&% 5 * (5) where the input bias function, #&% >* , gives the probability that an edge exists at % . As described in Williams and Thornber[11], the right and left eigenfunctions, #+! * and #O!P* , of #O!P* with largest positive real eigenvalue, , play a central role in the computation of saliency: #&% /' * AWATA
E % 5 E 6 #&% $' 24% 5 '76>* #:% 5 '16>* (6) #&% $' * AWATA E % 5 E 6 #:% 5 '16>* #&% 5 '16 24% "' * ! (7) Because #+! * is invariant under a transformation which reverses the order and direction of its arguments: #&% /' 24% 5 '76>* #&% 5 '76 24% "' * (8) the right and left eigenfunctions are related as follows: #:% $' * #:% $' * ! (9) 2.4 Stochastic completion field The magnitude of the stochastic completion field, #&% 5 '16* , equals the probability that a closed contour satisfying a subset of the constraints exists at #&% 5 '76>* . It is the sum of three terms: #&% 5 '76>* 8#&% 5 '16>* 8 #&% 5 '76>* 8#&% 5 '76>* 0 #:% 5 '16* 0 #&% 5 '76>* 8 #&% 5 '76>* AWATA E % E #&% $' * #&% "' * (10) where #&% 5 '76>* is a source field, and #&% 5 '76>* is a sink field: #:% 5 '16>* ATATA E % E ? #&% 5 '16 2@% "' * #&% >* #&% $' * (11) #:% 5 '16>* ATATA E % E #&% /' * #&% >* ? #:% "' 2@% 5 '16*
! (12) The purpose of writing #&% 5 '16* in this way is to remove the contribution, 0 #&% 5 '76>* 0 #&% 5 '16* , of closed contours at scales smaller than d which would otherwise dominate the completion field. Given the above expression for the completion field, it is clear that the key problem is computing the eigenfunction, #O!P* , of #O!P* with largest positive real eigenvalue. To accomplish this, we can use the well known power method (see [4]). In this case, the power method involves repeated application of the linear operator, #O!P* , to the function, #O!P* , followed by normalization: 0 #&% "' * ATATA E % 5 E 6 #&% $' 2 % 5 '76>* #&% 5 '76>* AWA A AWATA E % E E % 5 E 6 #&% $' 2 % 5 '76>* #&% 5 '76>* ! (13) In the limit, as gets very large, 0 #&% /' * converges to the eigenfunction of #+! * , with largest positive real eigenvalue. We observe that the above computation can be considered a continuous state, discrete time, recurrent neural network. 3 A discrete implementation of the continuum formulation The continuous functions comprising the state of the computation are represented as weighted sums of a finite set of shiftable-twistable basis functions. The weights form the coefficient vectors for the functions. The computation we describe is biologically plausible in the sense that all transformations of state are effected by linear transformations (or other vector parallel operations) on the coefficient vectors. 3.1 Shiftable-twistable bases The input and output of the above computation are functions defined on the continuous space, 0 , of positions in the plane, , and directions in the circle, 0 . For such computations, the important symmetry is determined by those transformations,
, of 0 , which perform a shift in by % 8 , followed by a twist in 0 through an angle, 8 . A twist through an angle, 8 , consists of two parts: (1) a rotation,
, of and (2) a translation in 0 , both by 8 . The symmetry,
, which is called a shift-twist transformation, is given by the formula,
#&% $' * #
#&% % 8:*>' k 8:* ! (14) A visual computation, , on 0 is called shift-twist invariant if, for all #&% 8' 8:* 0 , a shift-twist of the input by #&% 8 ' 8 * produces an identical shift-twist of the output. This property can be depicted in the following commutative diagram: #&% "' * #&% /' *
#
#&% % 8 *c' k 8 * #
#&% % 8 *c' k 8 * where #O!P* is the input, #+! * , is the output, is the computation, and ! "# is the shifttwist transformation. Correspondingly, we define a shiftable-twistable basis2 of functions on 0 to be a set of functions on 0 with the property that whenever a function, $ #&% $' * , is in their span, then so is $ #
#&% /' *+* , for every choice of #&% 8 ' 8 * in 0 . As such, the notion of a shiftable-twistable basis on %& 0 generalizes that of a shiftablesteerable basis on [2, 10]. Shiftable-twistable bases can be constructed as follows. Let ' #&% $' * be a function on 0 which is periodic (with period ( ) in both spatial variables, % . In analogy with the definition of a shiftable-steerable function on , we say that ' is shiftable-twistable on ) 0 if there are integers, * and + , and interpolation functions, ,- . #&% 8 ' 8 * , such that for each #:% 8 ' 8 */ 0 , the shift-twist of ' by #:% 8 ' 8 * is a linear combination of a finite number of basic shift-twists of ' by amounts # % 0 d ' d
* , i.e., if ' # 1 2
#&% /' *+* 43 . ,5 . #&% 8' 8;* ' # 6 . ^ ^ " #&% "' *+*
! (15) Here d (879* is the basic shift amount and d
7:+ is the basic twist amount. The sum in equation (15) is taken over all pairs of integers, % 0 # 0 ' 0<; * , in the range, >= 0 ' 0 ;? * , and all integers, , in the range, >= ? + . The Gaussian-Fourier basis is the product of a shiftable-steerable basis of Gaussians in % and a Fourier series basis in . For the experiments in this paper, the standard deviation of the Gaussian basis function, @ #&% * 0 ^&A<BDC C FE ^ , equals the basic shift amount, d . We regard @ #&% * as a periodic function of period, ( , which is chosen to be much larger than d , so that @ # (87 ' (87 * and its derivatives are essentially zero. For each frequency, G , and shift amount, d (where * (H74d is an integer), we define the Gaussian-Fourier basis functions, ' . I , by ' . I #&% $' * @ #&% % 0 d * AKJ IL
! (16) Zweck and Williams[13] showed that the Gaussian-Fourier basis is shiftable-twistable. 2We use this terminology even though the basis functions need not be linearly independent. 3.2 Power method update formula Suppose that #&% $' * can be represented in the Gaussian-Fourier basis as #&% /' * 3 . I . I ' . I #&% /' *c! (17) The vector, , with components, . I , will be called the coefficient vector of #&% /' * . In the next two sections, we demonstrate how the following integral linear transform: 0 #:% $' * A AWA
E % 5 E 6 ? 8 #:% $' 2@% 5 '16>* #&% 5 * #:% 5 '16>* (18) (i.e., the basic step in the power method) can be implemented as a discrete linear transform in a Gaussian-Fourier shiftable-twistable basis: 0 ! (19) 3.3 The propagation operator P In practice, we do not explicitly represent the matrix, . Instead we compute the necessary matrix-vector product using the advection-diffusion-decayoperator in the Gaussian-Fourier shiftable-twistable basis, , described in detail in Zweck and Williams[13]: 0 Q R U C (20) where 8
8 and where: 0 GH# d +*
0 (21)
0 # *
! (22) In the shiftable-twistable basis, the advection operator, , is a discrete convolution: 3 . I , B . B I # d +* . I (23) with the following kernel: , . # d +* 0 X A X 8 , . # d J - ' ) M * "# * E (24) where the , . are sinc functions. Let ! be the number of Fourier series frequencies, G , used in the shiftable-twistable basis, and let d 7"! . The diffusion-decay operator, , is a diagonal matrix: 0 . I A B ^ S E$# # A B J I ^
# * A J I ^
:* . I (25) where % ^ S ^
. 3.4 The bias operator In the continuum, the bias operator effects a multiplication of the function, #:% >* , by the input bias function, #:% * . Our aim is to identify an equivalent linear operator in the shiftabletwistable basis. Suppose that both and are represented in a Gaussian basis, @ . #:% >* . Their product is: #&% >* #&% * 3 . . @ . #&% >*'& 3 @ #&% * 3 . 3 . @ . #&% >* @ #&% >* ! (26) Now, the product of two Gaussian basis functions, @ . and @ , is a Gaussian of smaller variance which cannot be represented in the Gaussian basis, @ . . Because #&% >* #:% >* is a linear combination of the products of pairs of Gaussian basis functions, it cannot be represented in the Gaussian basis either. However, we observe that the convolution of #:% * #:% >* and a Gaussian, #&% * J #:% >* #&% >*9M , where #:% >* 0 ^ X A B E ^ , can be represented in the Gaussian basis. It follows that there exists a matrix, , such that: #&% * J #:% * #&% *NM 3 . J M . @ . #&% *c! (27) The formula for the matrix, , is derived by first completing the square in the exponent of the product of two Gaussians to obtain: @ #&% d % 0 * @ #&% d % * @ # #:% ^ # % 0 % *+** @ # ^ # % 0 % **c! (28) This product is then convolved with to obtain a function, $ #&% >* , which is a shift of the Gaussian basis function, @ #:% * . Finally we use the shiftability formula: @ #&% % 8;* 3 . ,5 . #&% 8[* @ . #&% * (29) where , . are the interpolation functions, @ . #&% * equals @ #:% d % 0 * , and d (879* is the shift amount, to express $ #:% * in the Gaussian basis. The result is: . 3 J J $# 2 2 % $ % 2 2 7 * , . # d # % @ % * 7 *c! (30) 4 Experimental results In our experiments the Gaussian-Fourier basis consisted of *
translates (in each spatial dimension) of a Gaussian (of period, ( ! ), and !
harmonic signals in the orientation dimension. The standard deviation of the Gaussian was set equal to the shift amount, d (879* . For illustration purposes, all functions were rendered at a resolution of 4 4 . The diffusion parameter, , equaled ! , and the decay constant, , equaled ! . The time step, d , used to solve the Fokker-Planck equation in the basis equaled d 7 . The parameters for the cut-off function used to eliminate self-loops were _` and ] ; . In the first experiment, the input bias function, #&% >* , consisted of twenty randomly positioned spots and twenty spots on the boundary of an avocado. The positions of the spots are real valued, i.e., they do not lie on the grid of basis functions. See Fig. 1 (left). The stochastic completion field computed using 32 iterations of the power method is shown in Fig. 1 (right). In the second experiment, the input bias function from the first experiment was rotated by and translated by half the distance between the centers of adjacent basis functions, # #&% J ^ ' ^ M *+* . See Fig. 2 (left). The stochastic completion field is identical (up to rotation and translation) to the one computed in the first experiment. This demonstrates the Euclidean invariance of the computation. See Fig. 2 (right). The estimate of the largest positive real eigenvalue, , as a function of , the power method iteration is shown in Fig. 3. 5 Conclusion We described a neural network which enhances and completes salient closed contours. Even though the computation is implemented in a discrete network, its output is invariant under continuous rotations and translations of the input pattern. References [1] Cowan, J.D., Neurodynamics and Brain Mechanisms, Cognition, Computation and Consciousness, Ito, M., Miyashita, Y. and Rolls, E., (Eds.), Oxford UP, 1997. Figure 1: Left: The input bias function, #:% * . Twenty randomly positioned spots were added to twenty spots on the boundary of an avocado. The positions are real valued, i.e., they do not lie on the grid of basis functions. Right: The stochastic completion field, A #&% 5 '76>* E 6 , computed using
basis functions. Figure 2: Left: The input bias function from Fig. 1, rotated by and translated by half the distance between the centers of adjacent basis functions, # #&% J ^ ' ^ M /*+* . Right: The stochastic completion field, is identical (up to rotation and translation) to the one shown in Fig. 1. This demonstrates the Euclidean invariance of the computation. 0 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0 5 10 15 20 25 30 35 Figure 3: The estimate of the largest positive real eigenvalue, , as a function of , the power method iteration. Both the final value and all intermediate values are identical in the rotated and non-rotated cases. [2] Freeman, W., and Adelson, E., The Design and Use of Steerable Filters, IEEE Transactions on Pattern Analysis and Machine Intelligence 13 (9), pp.891-906, 1991. [3] Mumford, D., Elastica and Computer Vision, Algebraic Geometry and Its Applications, Chandrajit Bajaj (ed.), Springer-Verlag, New York, 1994. [4] Golub, G.H. and C.F. Van Loan, Matrix Computations, Baltimore, MD, Johns Hopkins Univ. Press, 1996. [5] Heitger, R. and von der Heydt, R., A Computational Model of Neural Contour Processing, Figure-ground and Illusory Contours, Proc. of 4th Intl. Conf. on Computer Vision, Berlin, Germany, 1993. [6] Iverson, L., Toward Discrete Geometric Models for Early Vision, Ph.D. dissertation, McGill University, 1993. [7] Li, Z., A Neural Model of Contour Integration in Primary Visual Cortex, Neural Computation 10(4), pp. 903-940, 1998. [8] Parent, P., and Zucker, S.W., Trace Inference, Curvature Consistency and Curve Detection, IEEE Transactions on Pattern Analysis and Machine Intelligence 11, pp. 823-889, 1989. [9] Shashua, A. and Ullman, S., Structural Saliency: The Detection of Globally Salient Structures Using a Locally Connected Network, 2nd Intl. Conf. on Computer Vision, Clearwater, FL, pp. 321-327, 1988. [10] Simoncelli, E., Freeman, W., Adelson E. and Heeger, D., Shiftable Multiscale Transforms, IEEE Trans. Information Theory 38(2), pp. 587-607, 1992. [11] Williams, L.R., and Thornber, K.K., Orientation, Scale, and Discontinuity as Emergent Properties of Illusory Contour Shape, Neural Computation 13(8), pp. 16831711, 2001. [12] Yen, S. and Finkel, L., Salient Contour Extraction by Temporal Binding in a Cortically-Based Network, Neural Information Processing Systems 9, Denver, CO, 1996. [13] Zweck, J., and Williams, L., Euclidean Group Invariant Computation of Stochastic Completion Fields Using Shiftable-Twistable Functions, Proc. European Conf. on Computer Vision (ECCV ’00), Dublin, Ireland, 2000.
|
2001
|
111
|
1,919
|
A theory of neural integration in the head-direction system Richard H.R. Hahnloser , Xiaohui Xie and H. Sebastian Seung Howard Hughes Medical Institute Dept. of Brain and Cognitive Sciences Massachusetts Institute of Technology Cambridge, MA 02139 rhahnloser|xhxie|seung @mit.edu Abstract Integration in the head-direction system is a computation by which horizontal angular head velocity signals from the vestibular nuclei are integrated to yield a neural representation of head direction. In the thalamus, the postsubiculum and the mammillary nuclei, the head-direction representation has the form of a place code: neurons have a preferred head direction in which their firing is maximal [Blair and Sharp, 1995, Blair et al., 1998, ?]. Integration is a difficult computation, given that head-velocities can vary over a large range. Previous models of the head-direction system relied on the assumption that the integration is achieved in a firing-rate-based attractor network with a ring structure. In order to correctly integrate head-velocity signals during high-speed head rotations, very fast synaptic dynamics had to be assumed. Here we address the question whether integration in the head-direction system is possible with slow synapses, for example excitatory NMDA and inhibitory GABA(B) type synapses. For neural networks with such slow synapses, rate-based dynamics are a good approximation of spiking neurons [Ermentrout, 1994]. We find that correct integration during high-speed head rotations imposes strong constraints on possible network architectures. 1 Introduction Several network models have been designed to emulate the properties of head-direction neurons (HDNs) [Zhang, 1996, Redish et al., 1996, Goodridge and Touretzky, 2000]. The model by Zhang reproduces persistent activity during stationary head positions. Persistent neural activity is generated in a ring-attractor network with symmetric excitatory and inhibitory synaptic connections. Independently, he and Redish et al. showed that integration is possible by adding asymmetrical connections to the attractor network. They assumed that the strength of these asymmetrical connections is modulated by head-velocity. When the rat moves its head to the right, the asymmetrical connections induce a rightward shift of the activity in the attractor network. A more plausible model without multiplicative modulation of connections has been studied recently by Goodridge and Touretzky. There, the head-velocity input has a modulatory influence on firing rates of intermittent neurons rather than on connection strengths. The intermittent neurons are divided into two groups that make spatially offset connections, one group to the right, the other to the left. The different types of neurons in the Goodridge and Touretzky model have firing properties that are comparable to neurons in the various nuclei of the head-direction system. What all these previous models have in common is that the integration is performed in an inherent double-ring network with very fast synapses (less than ms for [Goodridge and Touretzky, 2000]). The connections made by one ring are responsible for rightward turns and the connections made by the other ring are responsible for leftward turns. In order to derive a network theory of integration valid for fast and slow synapses, here we solve a simple double-ring network in the linear and in the saturated regimes. An important property of the head-direction system is that the integration be linear over a large range of head-velocities. We are interested in finding those type of synaptic connections that yield a large linear range and pose our findings as predictions on optimal network architectures. Although our network is conceptually simpler than previous models, we show that using two simple read-out methods, averaging and extracting the maximum, it is possible to approximate head-velocity independent tuning curves as observed in the Postsubiculum (PoS) and anticipatory responses in the anterior dorsal thalamus (ADN). 2 Definition of the model We assume that the number of neurons in the double-ring network is large and write its dynamics as a continuous neural field
(1)
(2) where ! #" $&% (') '+* $), ('& .-/ 102.354 '76 398: ! " $ , ('& ' -/ ; $ % ('& * 102.3 4 6 398:=< " > 0 : @?BADC E
> denotes a rectification nonlinearity. and are the firing rates of neurons in the left and right ring, respectively. The quantities D and F represent synaptic activations (amount of neurotransmitter release caused by the firing rates and ). is a synaptic time constant. The vestibular inputs 3G4 '6 3 and 354H 6 3 are purely excitatory, ' 354JI 6 3IK354 . For simplicity, we assume that 6 3 is proportional to angular headvelocity. The synaptic connection profiles $ % between neurons on the same ring and $ , between neurons on different rings are given by: $)% LNM4OJM QPSRT $), UWVB4LV XPGRT S< (3) M4 , M , VB4 and V define the intra and inter-ring connection strengths. * is the intra-ring connection offset and the inter-ring offset. 3 Integration When the animal is not moving, the vestibular inputs to the two rings are equal, 6 3O E . In this case, within a certain range of synaptic connections, steady bumps of activities appear on the two rings. When the head of the animal rotates, the activity bumps travel at a velocity determined by 6 3 . For perfect integration, should be proportional to 6 3 over the full range of possible head-velocities. This is a difficult computational problem, in particular for slow synapses. 4 Small head-velocity approximation When the head is not rotating ( 6 3 E ), the two stationary bumps of synaptic activation are of the form " PGRT (') 4 ' 0 : and L " PSR T (') 4 ' 0 : (4) where 4 is the current head direction and is the offset between the two bumps. How to calculate , and is shown in the Appendix. The half width of these bumps is given by WA PPSRT
S< (5) When the angular head velocity is small ( 6 3
3S4 ), we linearize the dynamics around the stationary solution Eq. (4), see Appendix. We find that G (' " ' PGRT ('& 4 ' ' ' 0 : (6) F (' " PGRT ('& 4 ' ; 'J 10 : (7) where the velocity is given by 6 3 M XT 1* " M4 ' VB4F ' 'J M4 ' VB4 T 0 (8) and M PGRT *' V ' M T * (9) T PGRT ' ! T < (10) Equation (8) is the desired result, relating the velocity of the two bumps to the differential vestibular input 6 3 . In Fig. 1 we show simulation results using slow synapses ( " E ms). The integration is linear over almost the entire range of head-velocities (up to more than # EE$%
!'&%( ) when V M , i.e., when the amplitudes of inter-ring and intra-ring connections are equal. We point out that the condition V NM cannot directly be deduced from the above formulas, some empirical tuning (for example V 4 E ) was necessary to achieve this large range of linearity (large both in 6 3 and ). When the bumps move, their amplitudes tend to decrease. Fig. 1d shows the peak firing rates of neurons in the two rings as a function of vestibular input. As can be seen, the firing rates are a linear function of vestibular input, in agreement with equations 17 and 18 of the Appendix. However, a linear firing-rate modulation by head velocity is not universal, for some parameters we have seen asymmetrically head-velocity tuning, with a preference for small head velocities (not shown). a. b. −1 −0.5 0 0.5 1 −800 −600 −400 −200 0 200 400 600 800 ∆ b/b0 v (degrees/sec) Simulation Theory −1 −0.5 0 0.5 1 −800 −600 −400 −200 0 200 400 600 800 ∆ b/b0 v (degrees/sec) Simulation Theory c. d. −1 −0.5 0 0.5 1 −600 −400 −200 0 200 400 600 ∆ b/b0 v (degrees/sec) Simulation Theory −1 −0.5 0 0.5 1 50 55 60 65 70 75 ∆ b/b0 Firing rate (Hz) Left Right Figure 1: Velocity of activity bumps as a function of vestibular input 6 3
3G4 . a. Sublinear integration. V M , * E$ , VB4 E . b. Supralinear integration. V M , * $ , V 4 E . c. Linear (perfect) integration. V M , * $ , V 4 E . d. Head-velocity dependent modulation of firing rates (on the right and on the left ring). Same parameters as in c. " E ms. * " $ , and " $ . 5 Saturating velocity When 6 3 is very large, at some point, the left ring becomes inactive. Because inactivating the left ring means that the push-pull competition between the two rings is minimized, we are able to determine the saturating velocity of the double-ring network. The saturating velocity is given by the on-ring connections $ % . Define $ M4 @M XPGRT ('+* M4 @M XPGRT * PSR T 9A * T $&% ' 9A * $ % where $)% M4 M XPGRT * PGRT . Now, let be the steady solution of a ring network with symmetric connections $&% . By differentiating, it follows that '
is the solution of a ring network with connections $ . Hence, the saturating velocity is given by 9A * < (11) Notice that a traveling solution may not always exist if one ring is inactive (this is the case when there are no intra-ring excitatory connections). However, even without a traveling solution, equation (11) remains valid. In Figs. 1a and b, the saturating velocity is indicated by the horizontal dotted lines, in Fig. 1a we find " E $%
!'&%( and in Fig. 1b E $'
'&'( . 6 ADN and POs neurons Goodridge and Touretzky’s integrator model was designed to emulate details of neuronal tuning as observed in the different areas of the head-direction system. Wondering whether the simple double ring studied here can also reproduce multiple tuning curves, we analyze simple read-out methods of the firing rates and . What we find is that two readout methods can indeed approximate response behavior resembling that of ADN and POs neurons. ADN neurons: By reading out firing rates using a maximum operation, ?BADC ! , anticipatory head-direction tuning arises due to the fact that there is an activity offset between the two rings, equation (13). When the head turns to the right, the activity on the right ring is larger than on the left ring and so the tuning of is biased to the right. Similarly, for left turns, is biased to the left. Thus, the activity offset between the two rings leads to an anticipation time for ADN neurons, see Figure 2. Because, by assumption is head-velocity independent, it follows that is inversely proportional to head-velocity (assuming perfect integration),
. In other words, the anticipation time tends to be smaller for fast head rotations and larger for slow head rotations. POs neurons: By reading out the double ring activity as an average,
D , neurons in POs do not have any anticipation time: because averaging is a symmetric operation, all information about the direction of head rotations is lost. 0 90 180 270 360 Head−direction (degs) Firing Rate Left ring Right ring Max Average Left turn Right turn Figure 2: Snapshots of the activities on the two rings (top). Reading out the activities by averaging and by a maximum operation (bottom). 7 Discussion Here we discuss how the various connection parameters contribute to the double-ring network to function as an integrator. In particular we discuss how parameters have to be tuned in order to yield an integration that is large in 6 3 and in . : By assumption the synaptic time constant is large. has the simplest effect of all parameters on the integrator properties. According to equation (8), scales the range of . Notice that if were small, a large range of could be trivially achieved. The art here is to achieve this with large . * : The connection offset * between neurons receiving similar vestibular input is the sole parameter besides determing the saturating head-velocity, beyond which integration is impossible. According to equation (11), the saturating velocity is large if * is close to E$ (we want the saturating velocity to be large). In other words, for good integration, excitatory connections should be strongest (or inhibitory connections weakest) for neuron pairs with preferred head-directions differing by close to E $ . - : The connection offset between neurons receiving different vestibular input determines the anticipation time of thalamic neurons. If is large, then , the activity offset in equation (13) is large. And, because is proportional to (assuming perfect integration), we conclude that should preferentially be large (close to E$ ) if is to be large. Notice that by equation (8), the range of is not affected by - . VB4 and V : The inter-ring connections should be mainly excitatory, which implies that VB4 should not be too negative ( V=4 E was found to be optimal). The intuitive reason is the following. We want the integration to be as linear in 6 3 as possible, which means that we want our linear expansions (6) and (7) to deviate as little as possible from (4). Hence, the differential gain between the two rings should be small, which is the case when the two rings excite each other. The interring excitation makes sure, even for large values of 6 3 , that there are comparable activity levels on the two rings. This is one of the main points of this study. M4 and M : The intra-ring connections should be mainly inhibitory, which implies that M4 should be strongly negative. The reason for this is that inhibition is necessary to result in proper and stable integration. Since inhibition cannot come from the inter-ring connections, it has to come from M 4 . Notice also that according to equation (15), M cannot be much larger than V . If this were the case, the persistent activity in the no head-movement case would become unstable. For linear integration we have found that the condition V M is necessary; small deviations from this condition cause the integrator to become sub- or supralinear. 8 Conclusion We have presented a theory for integration in the head-direction system with slow synapses. We have found that in order to achieve a large range of linear integration, there should be strong excitatory connections between neurons with dissimilar head-velocity tuning and inhibitory connections between neurons with similar head-velocity tuning (see the discussion). Similar to models of the occulomotor integrator [Seung, 1996], we have found that linear integration can only be achieved by precise tuning of synaptic weights (for example V NM ). Appendix To study the traveling pulse solution with velocity , it is convenient to go into a moving coordinate frame by the change of variables B' . The stationary solution in the moving frame reads ' and ' W (12) Set E . In order to find the fixed points of equation (12), we use the ansatz (4) and equate the coefficients of the 3 Fourier modes T , PGRT and the -independent mode. This leads to A PT M QT 1*
V ' (13) 3 4 ' M4L.VB4F 4 ' PGRT (14) M XPGRT * V ' M T 1* 8 (15) where the functions D4 and are given by 4 U D " T '& PGRT 10 D " ' T 0 < The above set of equations fully characterize the solution for E . Eq. (13) determines the offset between the two rings, eq. (15) determines the threshold , eq. (14) the amplitude and eq. (5) the bias . When the vestibular input 6 3 is small, we assume that the perturbed solution around and takes the form: PGRT (') 4 ' PGRT ') 4F ' S< We linearize the dynamics (12) (to first order in ) and equate the Fourier coefficients. This leads to M T * ! " T
' T 0 (16) where ' and ' . We determine and by solving the linearized dynamics of the differential mode
' ' PSR T ' 4F ' . Comparing once more the Fourier coefficients leads to 6 3 " ' ' T 0 (17) 6 3 " ' ' T 0 (18) where M 4 ' V 4
. By substituting and into Eq. (16), we find equation (8). References [Blair et al., 1998] Blair, H., Cho, J., and Sharp, P. (1998). Role of the lateral mammillary nucleus in the rat head direction circuit: A combined single unit recording and lesion study. Neuron, 21:1387–1397. [Blair and Sharp, 1995] Blair, H. and Sharp, P. (1995). Anticipatory head diirection signals in anterior thalamus: evidence for a thalamocortical circuit that integrates angular head motion to compute head direction. The Journal of Neuroscience, 15(9):6260–6270. [Ermentrout, 1994] Ermentrout, B. (1994). Reduction of conductance-based models with slow synapses to neural nets. Neural Computation, 6:679–695. [Goodridge and Touretzky, 2000] Goodridge, J. and Touretzky, D. (2000). Modeling attractor deformation in the rodent head-direction system. The Journal of Neurophysiology, 83:3402–3410. [Redish et al., 1996] Redish, A., Elga, A. N., and Touretzky, D. (1996). A coupled attractor model of the rodent head direction system. Network: Computation in Neural Systems, 7:671–685. [Seung, 1996] Seung, H. S. (1996). How the brain keeps the eyes still. Proc. Natl. Acad. Sci. USA, 93:13339–13344. [Zhang, 1996] Zhang, K. (1996). Representation of spatial orientation by the intrinsic dynamics of the head-direction cell ensemble: A theory. J. Neurosci., 16(6):2112–2126.
|
2001
|
112
|
1,920
|
(Not) Bounding the True Error John Langford Department of Computer Science Carnegie-Mellon University Pittsburgh, PA 15213 jcl+@cs.cmu.edu Rich Caruana Department of Computer Science Cornell University Ithaca, NY 14853 caruana@cs.cornell.edu Abstract We present a new approach to bounding the true error rate of a continuous valued classifier based upon PAC-Bayes bounds. The method first constructs a distribution over classifiers by determining how sensitive each parameter in the model is to noise. The true error rate of the stochastic classifier found with the sensitivity analysis can then be tightly bounded using a PAC-Bayes bound. In this paper we demonstrate the method on artificial neural networks with results of a order of magnitude improvement vs. the best deterministic neural net bounds. 1 Introduction In machine learning it is important to know the true error rate a classifier will achieve on future test cases. Estimating this error rate can be suprisingly difficult. For example, all known bounds on the true error rate of artificial neural networks tend to be extremely loose and often result in the meaningless bound of “always err” (error rate = 1.0). In this paper, we do not bound the true error rate of a neural network. Instead, we bound the true error rate of a distribution over neural networks which we create by analysing one neural network. (Hence, the title.) This approach proves to be much more fruitful than trying to bound the true error rate of an individual network. The best current approaches [1][2] often require , , or more examples before producing a nontrivial bound on the true error rate. We produce nontrivial bounds on the true error rate of a stochastic neural network with less than examples. A stochastic neural network is a neural network where each weight
is perturbed by a gaussian with variance every time it is evaluated. Our approach uses the PAC-Bayes bound [5]. The approach can be thought of as a redivision of the work between the experimenter and the theoretician: we make the experimenter work harder so that the theoretician’s true error bound becomes much tighter. This “extra work” on the part of the experimenter is significant, but tractable, and the resulting bounds are much tighter. An alternative viewpoint is that the classification problem is finding a hypothesis with a low upper bound on the future error rate. We present a post-processing phase for neural networks which results in a classifier with a much lower upper bound on the future error rate. The post-processing can be used with any artificial neural net trained with any optimization method; it does not require the learning procedure be modified, re-run, or even that the threshold function be differentiable. In fact, this post-processing step can easily be adapted to other learning algorithms. David MacKay [4] has done significant work to make approximate Bayesian learning tractable with a neural network. Our work here is complimentary rather than competitive. We exhibit a technique which will likely give nontrivial true error rate bounds for Bayesian neural networks regardless of approximation or prior modeling errors. Verification of this statement is work in progress. The post-processing step finds a “large” distribution over classifiers, which has a small average empirical error rate. Given the average empirical error rate, it is straightforward to apply the PAC-Bayes bound in order to find a bound on the average true error rate. We find this large distribution over classifiers by performing a simple noise sensitivy analysis on the learned model. The noise model allows us to generate a distribution of classifiers with a known, small, average empirical error rate. In this paper we refer to the distribution of neural nets that results from this noise analysis as a stochastic neural net model. Why do we expect the PAC-Bayes bound to be a significant improvement over standard covering number and VC bound approaches? There exist learning problems for which the difference between the lower bound and the PAC-Bayes upper bound are tight up to where is the number of training examples. This is superior to the guarantees which can be made for typical covering number bounds where the gap is, at best, known up to an (asymptotic) constant. The guarantee that PAC-Bayes bounds are sometimes quite tight encourages us to apply them here. The next sections will: 1. Describe the bounds we will compare. 2. Describe our algorithm for constructing a distribution over neural networks. 3. Present experimental results. 2 Theoretical setup We will work in the standard supervised batch learning setting. This setting starts with the assumption that all examples are drawn from some fixed (unknown) distribution, , over (input, output) pairs,
. The output is drawn from the space and the input space is arbitrary. The goal of machine learning is to use a sample set of pairs to find a classifier, , which maps the input space to the output space and has a small true error, ! #"%$&
(' ) . Since the distribution is unknown, the true error rate is not observable. However, we can observe the empirical error rate, * +,-. #"/
0' ) ) 1 32 54 1
6' ) . Now that the basic quantities of interest are defined, we will first present a modern neural network bound, then specialize the PAC-Bayes bound to a stochastic neural network. A stochastic neural network is simply a neural network where each weight in the neural network is drawn from some distribution whenever it is used. We will describe our technique for constructing the distribution of the stochastic neural network. 2.1 Neural Network bound We will compare a specialization of the best current neural network true error rate bound [2] with our approach. The neural network bound is described in terms of the following parameters: 1. A margin, 798:7 . 2. An arbitrary function (unrelated to the neural network sigmoid function) ; defined by ;<=
) if
:7 , ;<=
) if
?> , and linear in between. 3. @ , an upper bound on the sum of the magnitude of the weights in the A th layer of the neural network 4. B , a Lipschitz constant which holds for the A th layer of the neural network. A Lipschitz constant is a bound on the magnitude of the derivative. 5. C , the size of the input space. With these parameters defined, we get the following bound. Theorem 2.1 (2 layer feed-forward Neural Network true error bound) D" $ EGF (HJI K L>9M5NPO QSR T8UWVYX[Z where R 8 ) 1 [2 ; Q
Q 1 B 1 B @ 1 @ Proof: Given in [2]. The theorem is actually only given up to a universal constant. “ ” might be the right choice, but this is just an educated guess. The neural network true error bound above is (perhaps) the tightest known bound for general feed-forward neural networks and so it is the natural bound to compare with. This 2 layer feed-forward bound is not easily applied in a tight manner because we can’t calculate a priori what our weight bound @ should be. This can be patched up using the principle of structural risk minimization. In particular, we can state the bound for @ 1 )! #" where $ is some non-negative integer and > is a constant. If the $ th bound holds with probability % '& " , then all bounds will hold simultaneously with probability Z , since ( ) " 4 1 $ )+* , Applying this approach to the values of both @ 1 and @ , we get the following theorem: Theorem 2.2 (2 layer feed-forward Neural Network true error bound) #" $ E F HJI K L>3M NO Q ".R T8 /$0 V X9Z where R 8<1$ 0 ) 1 2 ; Q
Q 1 B 1 B "2354 7689 :; </= Proof: Apply the union bound to all possible values of $ and 0 as discussed above. In practice, we will use ) 2 ) > and report the value of the tightest applicable bound for all $.0 . 2.2 Stochastic Neural Network bound Our approach will start with a simple refinement [3] of the original PAC-Bayes bound [5]. We will first specialize this bound to stochastic neural networks and then show that the use of this bound in conjunction with a post-processing algorithm results in a much tighter true error rate upper bound. First, we will need to define some parameters of the theorem. 1. ? is a distribution over the hypotheses which can be found in an example dependent manner. 2. @ is a distribution over the hypotheses which is chosen a priori—without dependence on the examples. 3. BA )+CD E AD is the true error rate of the stochastic hypothesis which, in any evaluation, draws a hypothesis from ? , and outputs
. 4. * BA )FCG HE A * +, is the average empirical error rate of the same stochastic hypothesis. Now, we are ready to state the theorem. Theorem 2.3 (PAC-Bayes Relative Entropy Bound) For all priors, @ , #" $ E F ? K KL * A +,JII HA LK KL M?NIOI @ QP N & V X Z where KL /?NIOI @ )SR UT P NWV O X Y J C is the Kullback-Leibler divergence between the distributions ? and @ and KL * ZA IOI A is the KL divergence between a coin of bias * HA and a coin of bias [A . Proof: Given in [3]. We need to specialize this theorem for application to a stochastic neural network with a choice of the “prior”. Our “prior” will be zero on all neural net structures other than the one we train and a multidimensional isotropic gaussian on the values of the weights in our neural network. The multidimensional gaussian will have a mean of and a variance in each dimension of R . This choice is made for convenience and happens to work. The optimal value of R is unknown and dependent on the learning problem so we will wish to parameterize it in an example dependent manner. We can do this using the same trick as for the original neural net bound. Use a sequence of bounds where R ) " for and some constants and $ a nonnegative number. For the $ th bound set Z % & " . Now, the union bound will imply that all bounds hold simultaneously with probability at least Z . Now, assuming that our “posterior” ? is also defined by a multidimensional gaussian with the mean and variance in each dimension defined by and , we can specialize to the following corollary: Corollary 2.4 (Stochastic Neural Network bound) Let 0 be the number of weights in a neural net, be the A th weight and be the variance of the A th weight. Then, we have #" $ F ? K KL * A +,JII HA K9M5NPO " 2 4 1 P N 9
9 1 P N " & X Z (1) Proof: Analytic calculation of the KL divergence between two multidimensional Gaussians and the union bound applied for each value of $ . We will choose ) > and -) > as reasonable default values. One more step is necessary in order to apply this bound. The essential difficulty is evaluting * A . This quantity is observable although calculating it to high precision is difficult. We will avoid the need for a direct evaluation by a monte carlo evaluation and a bound on the tail of the monte carlo evaluation. Let * A +, #" A /
' ) be the observed rate of failure of a random hypotheses drawn according to ? and applied to a random training example. Then, the following simple bound holds: Theorem 2.5 (Sample Convergence Bound) For all distributions, ? , for all sample sets , #" A E KL * A IOI* HA +, K P N & VX[Z where is the number of evaluations of the stochastic hypothesis. Proof: This is simply an application of the Chernoff bound for the tail of a Binomial where a “head” occurs when an error is observed and the bias is * A +, . In order to calculate a bound on the expected true error rate, we will first bound the expected empirical error rate * ZA +, with confidence & then bound the expected true error rate A with confidence & , using our bound on * A +, . Since the total probability of failure is only & & ) Z our bound will hold with probability Z . In practice, we will use ) evaluations of the empirical error rate of the stochastic neural network. 2.3 Distribution Construction algorithm One critical step is missing in the description: How do we calculate the multidimensional gaussian, ? ? The variance of the posterior gaussian needs to be dependent on each weight in order to achieve a tight bound since we want any “meaningless” weights to not contribute significantly to the overall sample complexity. We use a simple greedy algorithm to find the appropriate variance in each dimension. 1. Train a neural net on the examples 2. For every weight, , search for the variance, , which reduces the empirical accuracy of the stochastic neural network by some fixed target percentage (we use ) while holding all other weights fixed. 0.01 0.1 1 10 100 10000 100000 error pattern presentations SNN bound NN bound SNN Train error NN Train error SNN Test error NN Test error 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 10000 100000 error pattern presentations Figure 1: Plot of measured errors and error bounds for the neural network (NN) and the stochastic neural network (SNN) on the synthetic problem. The training set has 100 cases and the reduction in empirical error is 5%. Note that a true error bound of “100” (visible in the graph on the left) implies that at least more examples are required in order to make a nonvacuous bound. The graph on the right expands the vertical scale by excluding the poor true error bound that has error above 100. The curves for NN and SNN are qualitatively similar on the train and test sets. As expected, the SNN consistently performs 5% worse than the NN on the train set (easier to see in the graph on the right). Surprisingly, the SNN performs worse than the NN by less than 5% on the test sets. Both NN and SNN exhibit overfitting after about 6000-12000 pattern presentations (600-1200 epochs). The shape of the SNN bound roughly mimics the shape of the empirically measured true error (this is more visible in the graph on the right) and thus might be useful for indicating where the net begins overfitting. 3. The stochastic neural network defined by
will generally have a too-large empirical error. Therefore, we calculate a global multiplier such that the stochastic neural network defined by
decreases the empirical accuracy by only the same (absolute error rate) used in Step 2. 4. Then, we evaluate the empirical error rate of the resulting stochastic neural net by repeatedly drawing samples from the stochastic neural network. In the work reported here we use samples. 3 Experimental Results How well can we bound the true error rate of a stochastic neural network? The answer is much better than we can bound the true error rate of a neural network. We use two datasets to empirically evaluate the quality of the new bound. The first is a synthetic dataset which has 25 input dimensions and one output dimension. Most of these dimensions are useless—simply random numbers drawn from a !" # $ Gaussian. One of the 25 input dimensions is dependent on the label. First, the label % is drawn uniformly from & |