index
int64
0
20.3k
text
stringlengths
0
1.3M
year
stringdate
1987-01-01 00:00:00
2024-01-01 00:00:00
No
stringlengths
1
4
1,200
Multi-Grid Methods for Reinforcement Learning in Controlled Diffusion Processes Stephan Pareigis stp@numerik.uni-kiel.de Lehrstuhl Praktische Mathematik Christian-Albrechts-U niversi tat Kiel Kiel, Germany Abstract Reinforcement learning methods for discrete and semi-Markov decision problems such as Real-Time Dynamic Programming can be generalized for Controlled Diffusion Processes. The optimal control problem reduces to a boundary value problem for a fully nonlinear second-order elliptic differential equation of HamiltonJacobi-Bellman (HJB-) type. Numerical analysis provides multigrid methods for this kind of equation. In the case of Learning Control, however, the systems of equations on the various grid-levels are obtained using observed information (transitions and local cost). To ensure consistency, special attention needs to be directed toward the type of time and space discretization during the observation. An algorithm for multi-grid observation is proposed. The multi-grid algorithm is demonstrated on a simple queuing problem. 1 Introduction Controlled Diffusion Processes (CDP) are the analogy to Markov Decision Problems in continuous state space and continuous time. A CDP can always be discretized in state space and time and thus reduced to a Markov Decision Problem. Algorithms like Q-Iearning and RTDP as described in [1] can then be applied to produce controls or optimal value functions for a fixed discretization. Problems arise when the discretization needs to be refined, or when multi-grid information needs to be extracted to accelerate the algorithm. The relation of time to state space discretization parameters is crucial in both cases. Therefore 1034 S. Pareigis a mathematical model of the discretized process is introduced, which reflects the properties of the converged empirical process. In this model, transition probabilities of the discrete process can be expressed in terms of the transition probabilities of the continuous process. Recent results in numerical methods for stochastic control problems in continuous time can be applied to give assumptions that guarantee a local consistency condition which is needed for convergence. The same assumptions allow application of multi-grid methods. In section 2 Controlled Diffusion Processes are introduced. A model for the discretized process is suggested in section 3 and the main theorem is stated. Section 4 presents an algorithm for multi-grid observation according to the results in the preceding section. Section 5 shows an application of multi-grid techniques for observed processes. 2 Controlled Diffusion Processes Consider a Controlled Diffusion Process (CDP) ~(t) in some bounded domain 0 C ffi. n fulfilling the diffusion equation ~(t) = b(~(t), u(t))dt + (7(~(t))dw. (1) The control u(t) takes values in some finite set U. The immediate reinforcement (cost) for state ~(t) and control u(t) is r(t) = r(~(t),u(t)). (2) The control objective is to find a feedback control law u(t) = u(~(t)), (3) that minimizes the total discounted cost J(x, u) = IE~ 1 00 e-/3tr(~(t), u(t)dt, (4) where IE~ is the expectation starting in x E 0 and applying the control law u(.). (3 > 0 is the discount. The transition probabilities of the CDP are given for any initial state x E 0 and subset A c 0 by the stochastic kernels PtU(x, A) :=prob{~(t) E AI~(O) =x,u}. It is known that the kernels have the properties l (y - x)PtU(x, dy) l (y - x)(y - xf PtU(x, dy) t . b(x, u) + o(t) t· (7(x)(7(xf + o(t). (5) (6) (7) For the optimal control it is sufficient to calculate the optimal value function V : O-tffi. V(x) := inf J(x, u). u(.) (8) Multi-Grid Methods for Reinforcement Learning in Diffusion Processes 1035 Under appropriate smoothness assumptions V is a solution of the Hamilton-JacobiBellman (HJB-) equation min {C:tV(x) - ,i3V(x) + r(x, an = 0, x E O. aEU (9) Let a(x) = O"(x)O"(x)T be the diffusion matrix, then La, a E U is defined as the elliptic differential operator n n La := L aij(x)ox/Jx; + Lbi(x,a)oxi. (10) i,j=l i=l 3 A Model for Observed CDP's Let Ohi be the centers of cells of a cell-centered grid on 0 with cell sizes ho, hI = ho/2, h2 = hI/2, .... For any x E Ohi we shall denote by A(x) the cell of x. Let 6.t > 0 be a parameter for the time discretization. · · · · ,· t" r----J.,c r J.'" · · r ~ J0 · :J · lD· · · · Figure 1: The picture depicts three cell-centered grid levels and the trajectory of a diffusion process. The approximating value function is represented locally constant on each cell. The triangles on the path denote the position of the diffusion at sample times 0, /It, 2/lt, 3/lt, . . .. Transitions between respective cells are then counted in matrices Q't, for each control a and grid i. By counting the transitions between cells and calculating the empirical probabilities as defined in (20) we obtain empirical processes on every grid. By the law of great numbers the empirical processes will converge towards observed CDPs as subsequently defined. Definition 1 An observed process ~hi,Lldt) is a Controlled Markov Chain (i.e. discrete state-space and discrete time) on Ohi and interpolation time 6.ti with the transition probabilities prob{~(6.ti) E A(Y)I~(O) E A(x), u} :n { PXti (z, A(y»dz, i J A(x) (11) where x, y E Ohi and ~(t) is a solution of (1). Also define the observed reinforcement p as (12) 1036 s. Pareigis On every grid Ohi the respective process ehi ,Llti has its own value function Vhi ,Llti . By theorem 10.4.1. in Kushner, Dupuis ([5], 1992) it holds, that Vhi,Llti (x) -+ V(x) for all x E 0, (13) if the following local consistency conditions hold. Definition 2 Let D.eh,Llt = eh,Llt(D.t) - eh,Llt(O). eh,Llt is called locally consistent to a solution e(.) of (1), iff IE~ D.eh,Llt IE~[D.eh,Llt IE~D.eh,Llt][D.eh,Llt IE~D.eh,LltlT sup lD.eh,Llt(nD.t)I n b(x, a)D.t + o(D.t) (14) a(x)D.t + o(D.t) (15) -+ 0 as h -+ O. (16) To verify these conditions for the observed CDP, the expectation and variance can be calculated. For the expectation we get L Phi,Llti(x,y)(y - x) yEOhi :n L l (y - x)PXdz,A(y))dz. (17) i yEOhi A(x) Recalling properties (6) and (7) and doing a similar calculation for the variance we obtain the following theorem. Theorem 3 For observed CDPs ehi,Llti let hi and D.ti be such that (18) Furthermore, ehi ,Llti shall be truncated at some radius R, such that R -+ 0 for hi -+ 0 and expectation and variance of the truncated process differ only in the order o(D.t) from expectation and variance of ehi,Llti. Then the observed processes ehi,Llti truncated at R are locally consistent to the diffusion process e(.) and therefore the value functions Vhi ,Llti converge to the value function V. 4 Identification by Multi-Grid Observation The condition in Theorem 3 provides information as how to choose parameters in the algorithm with empirical data. Choose discretization values ho, D.to for the coarsest grid no. D.to should typically be of order Ilbllsup/ho. Then choose for the finer grids grid space time o ho D.to 1 ~ 2 ~ 2 2 ~ 4 ~ 2 3 4 ~ 16 ~ 4 5 ~ 32 ~ 8 (19) The sequences verify the assumption (18). We may now formulate the algorithm for Multi-Grid Observation of the CDP e(.). Note that only observation is being carried out. The actual calculation of the value function may be done separately as described in the next section. The choice of the control is assumed to be done Multi-Grid Methodsfor Reinforcement Learning in Diffusion Processes 1037 by a separate controller. Let Ok be the finest grid, Le. Dotk and hk the finest discretizations. Let U, = u~t';~t,. = U x ... xU, Dotl! Dotk times. Qr' is a 10,1 x 10,1matrix (a, E U,), containing the number of transitions between cells in 0" Rr' is a lO,l-vector containing the empirical cost for every cell in 0 ,. The immediate cost is given by the system as r, = Jo~t' e-/3tr(~(t), a,)dt. T denotes current time. O. Initialize 0 " Qr', Rr' for all a, E U" 1 = 0, ... , k 1. repeat { 2. choose a = a(T) E U and apply a constantly on [T; T + Dotk) 3. T := T + Dotk 4. for I = 0 to k do { 5. determine cell Xl E 0, with ~(T - Dot,) E A(XI) 6. determine cell Yl E 0, with ~(T) E A(Yl) 7. if Ilxk - Ykll ~ R (truncation radius) then goto 2. else 8. a, := (a(T - Dot,) , a(T + Dotk - Dot,), .. . ,a(T - Dotk)) 9. receive immediate cost r, 10. Qr'(Xl,Yl) := Qr'(Xl,Yt) + 1 11. Rr' (Xl) := (rl + Rr' (Xl) . EZEn, Qr' (Xl, z)) / (1 + EZEn, Qr' (Xl, z)) } (for-do) } (repeat) Before applying a multi-grid algorithm for the calculation of the value function on the basis of the observations, one should make sure that every box has at least some data for every control. Especially in the early stages of learning only the two coarsest grids 00, 0 1 could be used for computation of the optimal value function and finer grids may be added (possibly locally) as learning evolves. 5 Application of Multi-Grid Techniques The identification algorithm produces matrices Qr' containing the number of transitions between boxes in 0 ,. We will calculate from the matrices Q the transition matrices P by the formula p,a' (x, y) = Qr' (x, Y)/ (L Qr' (x, Z)) , x, Y E 0" a, E U" 1 = 0, .. . , k. (20) zEn, Now we define matrices A and right hand sides I as Ar' := ({31 p,a' - I) / Dot, It':= Rr' / Dotl , (21) where {31 = e-/3~t,. The discrete Bellman equation takes the following form (22) The problem is now in a form to which the multi-grid method due to Hoppe, BloB ([2], 1989) can be applied. For prolongation and restriction we choose bilinear interpolation and full weighted restriction for cell-centered grids. We point out, that for any cell X E 0 , only those neighboring cells shall be used for prolongation and restriction for which the minimum in (22) is attained for the same control as the minimizing control in X (see [2], 1989 and [3], 1996 for details). On every grid 1038 s. Pareigis 0 1 the defect in equation (22) is calculated and used for a correction on grid 0 /- 1 . As a smoother nonlinear Gauss-Seidel iteration applied to (22) is used. Our approach differs from the algorithm in Hoppe, BloB ([2], 1989) in the special form of the matrices A~' in equation (22). The stars are generally larger than nine-point, in fact the stars grow with decreasing h although the matrices remain sparse. Also, when working with empirical information the relationship between the matrices Ar' on the various grids is based on observation of a process, which implies that coarse grid corrections do not always correct the equation of the finest grid (especially in the early stages of learning). However, using the observed transition matrices Ar' on the coarse grids saves the computing time which would otherwise be needed to calculate these matrices by the Galerkin product (see Hackbusch [4], 1985). 6 Simulation with precomputed transitions Consider a homogeneous server problem with two servers holding data (Xl, X2) E [0,1] x [0,1]. Two independent data streams arrive, one at each server. A controller has to decide to which server to route. The modeling equation for the stream shall be dx = b(x, u)dt + CT(x)dw, u E {I, 2} (23) with b(x,l) = (!1) b(x,2) = (~1) CT= (~ ~) (24) The boundaries at Xl = 0 and X2 = 0 are reflecting. The exceeding data on either server Xl, X2 > 1 is rejected from the system and penalized with g(Xl, 1) = g(1,x2) = 10, 9 = 0 otherwise. The objective of the control policy shall be to minimize IE 10 00 e-i3t(xI(t) + X2(t) + g(Xl,X2))dt. (25) The plots of the value function show, that in case of high load (Le. Xl, X2 close to 1) a maximum of cost is assumed. Therefore it is cheaper to overload a server and pay penalty than to stay close to the diagonal as is optimal in the low load case. For simulation we used preco~puted (Le. converged heuristic) transition probabilities to test the multi-grid performance. The discount f3 was set to .7. The multi-grid algorithm reduces the error in each iteration by' a factor 0.21, using 5 grid levels and a V -cycle and two smoothing iterations on the coarsest grid. For comparison, the iteration on the finest grid converges with a reduction factor 0.63. 7 Discussion We have given a condition for sampling controlled diffusion processes such that the value functions will converge while the discretization tends to zero. Rigorous numerical methods can now be applied to reinforcement learning algorithms in continuous-time, continuous-state as is demonstrated with a multi-grid algorithm for the HJB-equation. Ongoing work is directed towards adaptive grid refinement algorithms and application to systems that include hysteresis. Multi-Grid Methodsfor Reinforcement Leaming in Diffusion Processes 1039 Figure 2: Contour plots of the predicted reward in a homogeneous server problem with nonlinear costs are shown on different grid levels. On the coarsest 4 x 4 grid a sampling rate of one second is used with 9-point-star transition matrices. At the finest grid (64 x 64) a sampling rate of t second is used with observation on 81-point-stars. Inside the egg-shaped area the value function assumes its maximum. References [lJ A. Barto, S. Bradtke, S. Singh. Learning to Act using Real-Time Dynamic Programming, AI Journal on Computational Theories of Interaction and Agency, 1993. [2J M. BloB and R. Hoppe. Numerical Computation of the Value Function of Optimally Controlled Stochastic Switching Processes by Multi-Grid Techniques, Numer Funct Anal And Optim 10(3+4), 275-304, 1989. [3] S. Pareigis. Lernen der Lasung der Bellman-Gleichung durch Beobachtung von kontinuierlichen Prozessen, PhD Thesis, 1996. [4J W. Hackbusch. Multi-Grid Methods and Applications, Springer-Verlag, 1985. [5] H. Kushner and P. Dupuis. Numerical Methods for Stochastic Control Problems in Continuous Time, Springer-Verlag, 1992.
1996
145
1,201
Noisy Spiking Neurons with Temporal Coding have more Computational Power than Sigmoidal Neurons Wolfgang Maass Institute for Theoretical Computer Science Technische Universitaet Graz, Klosterwiesgasse 32/2 A-80lO Graz, Austria, e-mail: maass@igLtu-graz.ac.at Abstract We exhibit a novel way of simulating sigmoidal neural nets by networks of noisy spiking neurons in temporal coding. Furthermore it is shown that networks of noisy spiking neurons with temporal coding have a strictly larger computational power than sigmoidal neural nets with the same number of units. 1 Introduction and Definitions We consider a formal model SNN for a §piking neuron network that is basically a reformulation of the spike response model (and of the leaky integrate and fire model) without using 6-functions (see [Maass, 1996a] or [Maass, 1996b] for further backgrou nd). An SNN consists of a finite set V of spiking neurons, a set E ~ V x V of synapses, a weight wu,v 2: 0 and a response function cu,v : R+ --+ R for each synapse {u, v} E E (where R+ := {x E R: x 2: O}) , and a threshold function 8 v : R+ --+ R+ for each neuron v E V . If Fu ~ R+ is the set of firing times of a neuron u , then the potential at the trigger zone of neuron v at time t is given by Pv(t) := L L Wu,v . cu,v(t - s) . u:{u,v)EE sEF,,:s<t In a noise-free model a neuron v fires at time t as soon as Pv(t) reaches 8 v(t - t'), where t' is the time of the most recent firing of v . One says then that neuron v sends out an "action potential" or "spike" at time t . 212 W Maass For some specified subset Vin ~ V of input neurons one assumes that the firing times ("spike trains") Fu for neurons u E Vin are not defined by the preceding convention, but are given from the outside. The firing times Fv for all other neurons v E V are determined by the previously described rule, and the output of the network is given in the form of the spike trains Fv for a specified set of output neurons Vout ~ V . y y __ ~e~ £11," (I-S) o s I 0 s I ~ I' t Figure 1: Typical shapes of response functions cu,v (EPSP and IPSP) and threshold functions 8 v for biological neurons. We will assume in our subsequent constructions that all response functions cu,v and threshold functions 8 v in an SNN are "stereotyped", i.e. that the response functions differ apart from their "sign" (EPSP or IPSP) only in their delay du,v (where du,v := inf {t ~ 0 : cu,v(t) =J O}) , and that the threshold functions 8 v only differ by an additive constant (i.e. for all u and v there exists a constant cu,v such that 8 u(t) = 8 v(t) + Cu,v for all t ~ 0) . We refer to a term of the form wu,v . cu,v(t - s) as an ~xcitatory respectively inhibitory post~naptic potential (abbreviated: EPSP respectively IPSP). - Since biological neurons do not always fire in a reliable manner one also considers the related model of noisy spiking neurons, where Pv(t) is replaced by P ~Oi8Y(t) := Pv(t) +av(t) and 8 v(t-t') is replaced by 8~oi8Y(t_t') : = 8 v(t-t')+!3v(t-t'). av(t) and !3v(t - t') are allowed to be arbitrary functions with bounded absolute value (hence they can also represent "systematic noise"). Furthermore one allows that the current value of the difference D(t) := pvnOi8Y(t) 8~oi8Y(t - t') does not determine directly the firing time of neuron v, but only its current firing probability. We assume that the firing probability approaches 1 if D --+ 00 , and 0 if D --+ -00 . We refer to spiking neurons with these two types of noise as "noisy spiking neurons". We will explore in this article the power of analog computations with noisy spiking neurons, and we refer to [Maass, 1996a] for results about digital computations in this model. Details to the results in this article appear in [Maass, 1996b] and [Maass, 1997]. 2 Fast Simulation of Sigmoidal Neural Nets with Noisy Spiking Neurons in Temporal Coding So far one has only considered simulations of sigmoidal neural nets by spiking neurons where each analog variahle in the sigmOidal neural net is represented by the firing rate of a spiking neuron. However this "firing rate interpretation" is inconsistent with a number of empirical results about computations in biological neural Spiking Neurons have more Computational Power than Sigmoidal Neurons 213 systems. For example [Thorpe & Imbert, 1989] have demonstrated that visual pattern analysis and pattern classification can be carried out by humans in just 150 ms, in spite of the fact that it involvefi a minimum of 10 synaptic stages from the retina to the temporal lobe. [de Ruyter van Steveninck & Bialek, 1988] have found that a blowfly can produce flight torques within 30 ms of a visual stimulus by a neural system with several synaptic stages. However the firing rates of neurons involved in all these computations are usually below 100 Hz, and interspike intervals tend to be quite irregular. Hence One cannot interpret these analog computations with spiking neurons on the basis of an encoding of analog variables by firing rates. On the other hand experimental evidence has accumulated during the last few years which indicates that many biological neural systems use the timing of action potentials to encode information (see e.g. [Bialek & Rieke, 1992]' [Bair & Koch, 1996]). We will now describe a new way of simulating sigmoidal neural nets by networks of spiking neurons that is based on temporal coding. The key mechanism for this alternative simulation is based on the well known fact that EPSP's and IPSP's are able to shift the firing time of a spiking neuron. This mechanism can be demonstrated very clearly in our formal model if one assumes that EPSP's rise (and IPSP's fall) linearly during a certain initial time period. Hence we assume in the following that there exists some constant D. > 0 such that each response function eu,v(X) is of the form Ctu,v . (x - du,v) with Ctu,v E {-I, 1} for x E [du,v, du,v + D.] , and eu,v(X) = 0 for x E [0, du,v] . Consider a spiking neuron v that receives postsynaptic potentials from n presynaptic neurons at, ... ,an' For simplicity we asfiume that interspike intervals are so large that the firing time tv of neuron v depends just on a single firing time ta• of each neuron ai, and 8 v has returned to its "resting value" 8 v(0) before v fires again. Then if the next firing of v occurs at a time when the postsynaptic potentials described by Wa;,V . ea.,v(t - ta.) are all in their initial linear phase, its firing time tv is determined in the noise-free model for Wi := Wa;,v . Cta.,v by the equation E~=t Wi . (tv - tao - da.,v) = 8 v(0) , or equivalently (1) This equation revealfi the fiomewhat fiurprifiing fact that (for a certain range of their parameters) spiking neurons can compute a weighted sum in terms of firing times, i.e. temporal coding. One fihould alfio note that in the case where all delays da.,v have the same value, the "weights" Wi of this weighted sum are encoded in the "strengths" wa.,v of the synapsefi and their "sign" Cta.,t, , as in the "firing rate interpretation". Finally according to (1) the coefficients of the presynaptic firing times tao are automatically normalized, which appears to be of biological interest. In the simplest scheme for temporal coding (which is closely related to that in [Hopfield, 1995]) an analog variable x E [0,1] is encoded by the firing time T -,' x of a neuron, where T is assumed to be independent of x (in a biological context T might be time-locked to the onset of a fitimulus, or to some oscillation) and , is some constant that ifi determined in the proof of Theorem 2.1 (e.g. , = tlj2 in the noifie-free case). In contrafit to [Hopfield, 1995] we afifiume that both the inputs and the outputs of computationfi are encoded in thifi fafihion. This has the advantage that one can compose computational modules. 214 W. Maass We will first focus in Theorem 2.1 on the simulation of sigmoidal neural nets that employ the piecewise linear "linear saturated" activation function 1r : R [0,1] defined by 1r(Y) = 0 ify < 0, 1r(Y) = y if 0 ~ y ~ 1, and 1r(Y) = 1 ify > 1 . The Theorem 3.1 in the next section will imply that one can simulate with spiking neurons also sigmoidal neural nets that employ arbitrary continuous activation functions. Apart from the previously mentioned assumptions we will assume for the proofs of Theorem 2.1 and 3.1 that any EPSP satisfies eu,v(X) = 0 for all sufficiently large x, and eu,v(X) ~ eu,v(du,v + ~) for all x E [du,v + Ll, du,v + Ll + 1'] . We assume that each IPSP is continuous, and has value 0 except for some interval of R. Furthermore we assume for each EPSP and IPSP that jeu,v(x)1 grows at most linearly during the interval [du,v + Ll, du,v + Ll + 1'] . In addition we assume that 0 v(x) = 0 v(0) for sufficiently large x , and that 0 v(x) is sufficiently large for O<x~'Y. Theorem 2.1 For any given e, 8> 0 one can simulate any given feed forward sigmoidal neural net N with activation function 1r by a network NN,e,6 of noisy spiking neurons in temporal coding. More precisely, for any network input Xl, . .• ,Xm E [O,IJ the output of NN,e,6 differs with probability 2:: 1 - 0 by at most e from that of N . Furthermore the computation time of NN,e,6 depends neither on the number of gates in N nor on the parameters e, 0, but only on the number of layers of the sigmoidal neural network N . -We refer to [Maass, 1997] for details of the somewhat complicated proof. One employs the mechanism described by (1) to simulate through the firing time of a spiking neuron v a sigmoidal gate with activation function 1r for those gate-inputs where 1r operates in its linearly rising range. With the help of an auxiliary spiking neuron that fires at time T one can avoid the automatic "normalization" of the weights Wi that is provided by (1), and thereby compute a weighted sum with arbitrary given weights. In order to simulate in temporal coding the behaviour of the gate in the input range where 1r is "saturated" (Le. constant), it suffices to employ some auxiliary spiking neurons which make sure that v fires exactly once during the relevant time window (and not shortly before that). Since inputs and outputs of the resulting modules for each single gate of N are all given in temporal coding, one can compose these modules to simulate the multilayer sigmoidal neural net N. With a bit of additional work one can ensure that this construction also works with noisy spiking neurons. • 3 Universal Approximation Property of Networks of Noisy Spiking Neurons with Temporal Coding It is known [Leshno et al., 1993J that feed forward sigmoidal neural nets whose gates employ the activation function 1r can approximate with a single hidden layer for any n, kEN any given continuous function F : [O,I]n [O,I]k within any e > 0 with regard to the Loo-norm (i.e. uniform convE'rgence). Hence we can derive the following result from Theorem 2.1: Theorem 3.1 Any given continuous function F : [0, l]n _ [O,I]k can be approximated within any given e > 0 with arbitrarily high reliability in temporal coding by Spiking Neurons have more Computational Power than Sigmoidal Neurons 215 a network of noisy spiking neurons (SNN) with a single hidden layer (and hence within 15 ms for biologically realistic values of their time-constants). • Because of its generality this Theorem implies the same result also for more general schemes of coding analog variables by the firing times of neurons, besides the particular one that we have considered so far. In fact it implies that the same result holds for any other coding scheme C that is "continuously related" to the previously considered one in the sense that the transformation between firing times that encode an analog variable x in the here considered coding scheme and in the coding scheme C can be described by uniformly continuous functions in both directions. 4 Spiking Neurons have more Computational Power than Sigmoidal Neurons We consider the "element distinctness function" EDn by { I, EDn(Sl, ... ,Sn) = 0, arbitrary, if s i = s i for some i =I=- j if lSi -ail ~ 1 for all i,j with i =l=-j else. If one encodes the value of input variable Si by a firing of input neuron ai at time Tin - c· Si , then for sufficiently large values of the constant C > 0 a single noisy spiking neuron v can compute EDn with arbitrarily high reliability. This holds for any reasonable type ofresponse functions, e.g. the ones shown in Fig. 1. The binary output of this computation is assumed to be encoded by the firing/non-firing of v . Hair-trigger situations are avoided since no assumptions have to be made about the firing or non-firing of v if EPSP's arrive with a temporal distance between 0 and c . On the other hand the following result shows that a fairly large sigmoidal neural net is needed to compute the same function. Its proof provides the first application for Sontag's recent results about a new type of "dimension" d of a neural network N , where d is chosen maximal so that every subset of d inputs is shattered by N . Furthermore it expands a method due to [Koiran, 1995] for llsing the VC-dimension to prove lower bounds on network size. Theorem 4.1 Any sigmoidal neural net N that computes EDn has at least n2"4 -1 hidden units. Proof: Let N be an arbitrary sigmoidal neural net with k gates that computes EDn. Consider any set S ~ R+ of size n -1. Let .x > 0 be sufficiently large so that the numbers in .x . S have pairwise distance ~ 2 . Let A be a set of n - 1 numbers > max (.x . S) + 2 with pairwise distance ~ 2 . By assumption N can decide for n arbitrary inputs from .x . SuA whether they are all different. Let N>. be a variation of N where all weights on edges from the first input variable are mUltiplied with .x. Then N>. can compute any function from 216 W Maass S into {O, I} after one has assigned a suitable fixed set of n - 1 pairwise different numbers from ..\ . SuA to the last n - 1 input variables. Thus if one considers a'l programmable parameters of N the factor ..\ in the weights on edges from the first input variable and the ~ k thresholds of gates that are connected to some of the other n - 1 input variables, then N shatters S with these k + 1 programmable parameters. Since S ~ R + of size n - 1 was chosen arbitrarily, we can now apply the result from [Sontag, 1996], which yields an upper bound of 2w + 1 for the maximal number d such that every set of d different inputs can be shattered by a sigmoidal neural net with w programmable parameters (note that this parameter d is in general much smaller than the VC-dimension of the neural net). For w := k + 1 this implies in our case that n - 1 ~ 2(k + 1) + 1 , hence k ~ (n - 4)/2 . Thus N has at least (n - 4) /2 computation nodes, and therefore at least (n - 4)/2 -1 hidden units. One should point out that due to the generality of Sontag's result this lower bound is valid for all common activation functions of sigmoidal gates, and even if N employs heaviside gates besides sigmoidal gates. • Theorem 4.1 yields a lower bound of 4997 for the number of hidden units in any sigmoidal neural net that computes EDn for n = 10 000 , where 10 000 is a common estimate for the number of inputs (i.e. synapses) of a biological neuron. Finally we would like to point out that to the best of our knowledge Theorem 4.1 provides the largest known lower bound for any concrete function with n inputs on a sigmoidal neural net. The largest previously known lower bound for sigmoidal neural nets wa<; O(nl/4) , due to [Koiran, 1995J. 5 Conclusions Theorems 2.1 and 3.1 provide a model for analog computations in network of spiking neurons that is consistent with experimental results on the maximal computation speed of biological neural systems. As explained after Theorem 3.1, this result holds for a large variety of possible schemes for encoding analog variables by firing times. These theoretical results hold rigoro'U.sly only for a rather small time window of length, for temporal coding. However a closer inspection of the construction shows that the actual shape of EPSP's and IPSP's in biological neurons provides an automatic adjustment of extreme values of the inputs tao towards their average, which allows them to carry out rather similar computations for a substantially larger window size. It also appears to be of interest from the biological point of view that the synaptic weights play for temporal coding in our construction basically the same role as for rate coding, and hence the .~ame network is in principle able to compute closely related analog functions in both coding schemes. We have focused in our constructions on feedforward nets, but our method can for example also be used to simulate a Hopfield net with graded response by a network of noisy spiking neurons in temporal coding. A stable state of the Hopfield net corresponds then to a firing pattern of the simulating SNN where all neurons fire at the same frequency, with the ((pattern" of the stahle state encoded in their phase differences. Spiking Neurons have more Computational Power than Sigmoidal Neurons 217 The theoretical results in this article may also provide additional goals and directions for a new computer technology based on artificial spiking neurons. Acknowledgement I would like to thank David Haussler, Pa.')cal Koiran, and Eduardo Sontag for helpful communications. References [Bair & Koch, 1996] W. Bair, C. Koch, "Temporal preCISIon of spike trains in extra.')triate cortex of the behaving macaque monkey", Neural Computation, vol. 8, pp 1185-1202, 1996. [Bialek & Rieke, 1992] W. Bialek, and F. Rieke, "Reliability and information transmission in spiking neurons", Trends in Neuroscience, vol. 15, pp 428-434,1992. [Hopfield, 1995J J. J. Hopfield, "Pattern recognition computation using action potential timing for stimulus representations", Nature, vol. 376, pp 33-36, 1995. [Koiran, 1995] P. Koiran, "VC-dimension in circuit complexity", Proc. of the 11th IEEE Conference on Computational Complexity, pp 81-85, 1996. [Leshno et aI., 1993] M. Leshno, V. Y. Lin, A. Pinkus, and S. Schocken, "Multilayer feed forward networks with a nonpolynomial activation function can approximate any function", Neural Networks, vol. 6, pp 861-867, 1993. [Maass, 1996a] W. Maass, "On the computational power of noisy spiking neurons", Advances in Neural Information Processing Systems, vol. 8, pp 211-217, MIT Press, Cambridge, 1996. [Maass, 1996b] W. Maass, "Networks of spiking neurons: the third generation of neural network models", FTP-host: archive.cis.ohio-state.edu, FTP-filename: /pub/neuroprose/maass.third-generation.ps.Z, Neural Networks, to appear. [Maass, 1997] W. Maa.')s, "Fa.')t sigmoidal networks via spiking neurons", to appear in Neural Computation. FTP-host: archive.cis.ohio-state.edu FTP-filename: /pub/neuroprose/maass.sigmoidal-spiking.ps.Z, Neural Computation, to appear in vol. 9, 1997. [de Ruyter van Steveninck & Bialek, 1988] R. de Ruyter van Steveninck, and W. Bialek, "Real-time performance of a movement sensitive neuron in the blowfly visual system", Proc. Roy. Soc. B, vol. 234, pp 379-414, 1988. [Sontag, 1996] E. D. Sontag, "Shattering all sets of k points in 'general position' requires (k -1)/2 parameters", http://www.math.rutgers.edu/'''sontag/ , follow links to FTP archive. [Thorpe & Imbert, 1989J S. T. Thorpe, and M. Imbert, "Biological constraints on connectionist modelling", In: Connectionism in Perspective, R. Pfeifer, Z. Schreter, F. Fogelman-Soulie, and 1. Steels, eds., Elsevier, North-Holland, 1989.
1996
146
1,202
Edges are the 'Independent Components' of Natural Scenes. Anthony J. Bell and Terrence J. Sejnowski Computational Neurobiology Laboratory The Salk Institute 10010 N. Torrey Pines Road La Jolla, California 92037 tony@salk.edu, terry@salk.edu Abstract Field (1994) has suggested that neurons with line and edge selectivities found in primary visual cortex of cats and monkeys form a sparse, distributed representation of natural scenes, and Barlow (1989) has reasoned that such responses should emerge from an unsupervised learning algorithm that attempts to find a factorial code of independent visual features. We show here that non-linear 'infomax', when applied to an ensemble of natural scenes, produces sets of visual filters that are localised and oriented. Some of these filters are Gabor-like and resemble those produced by the sparseness-maximisation network of Olshausen & Field (1996). In addition, the outputs of these filters are as independent as possible, since the infomax network is able to perform Independent Components Analysis (ICA). We compare the resulting ICA filters and their associated basis functions, with other decorrelating filters produced by Principal Components Analysis (PCA) and zero-phase whitening filters (ZCA). The ICA filters have more sparsely distributed (kurtotic) outputs on natural scenes. They also resemble the receptive fields of simple cells in visual cortex, which suggests that these neurons form an information-theoretic co-ordinate system for images. 1 Introduction. Both the classic experiments of Rubel & Wiesel [8] on neurons in visual cortex, and several decades of theorising about feature detection in vision, have left open the question most succinctly phrased by Barlow "Why do we have edge detectors?" That is: are there any coding principles which would predict the formation of localised, oriented receptive 832 A. 1. Bell and T. 1. Sejnowski fields? Barlow's answer is that edges are suspicious coincidences in an image. Formalised information-theoretically, this means that our visual cortical feature detectors might be the end result of a redundancy reduction process [4, 2], in which the activation of each feature detector is supposed to be as statistically independent from the others as possible. Such a 'factorial code' potentially involves dependencies of all orders, but most studies [9, 10, 2] (and many others) have used only the second-order statistics required for decorrelating the outputs of a set of feature detectors. Yet there are multiple decorrelating solutions, including the 'global' unphysiological Fourier filters produced by PCA, so further constraints are required. Field [7] has argued for the importance of sparse, or 'Minimum Entropy', coding [4], in which each feature detector is activated as rarely as possible. Olshausen & Field demonstrated [12] that such a sparseness criterion could be used to self-organise localised, oriented receptive fields. Here we present similar results using a direct information-theoretic criterion which maximises the joint entropy of a non-linearly transformed output feature vector [5]. Under certain conditions, this process will perform Independent Component Analysis (or ICA) which is equivalent to Barlow's redundancy reduction problem. Since our ICA algorithm, applied to natural scenes, does produce local edge filters, Barlow's reasoning is vindicated. Our ICA filters are more sparsely distributed than those of other decorrelating filters, thus supporting some of the arguments of Field (1994) and helping to explain the results of Olshausen's network from an information-theoretic point of view. 2 Blind separation of natural images. A perceptual system is exposed to a series of small image patches, drawn from an ensemble of larger images. In the linear image synthesis model [12], each image patch, represented by the vector x, has been formed by the linear combination of N basis functions. The basis functions form the columns of a fixed matrix, A. The weighting of this linear combination (which varies with each image) is given by a vector, s. Each component of this vector has its own associated basis function, and represents an underlying 'cause' of the image. Thus: x=As. The goal of a perceptual system, in this simplified framework, is to linearly transform the images, x, with a matrix of filters, W, so that the resulting vector: u= Wx, recovers the underlying causes, s, possibly in a different order, and rescaled. For the sake of argument, we will define the ordering and scaling of the causes so that W = A -1. But what should determine their form? If we choose decorrelation, so that (uuT ) = I, then the solution for W must satisfy: WTW = (xxT)-1 There are several ways to constrain the solution to this: (1) (1) Principal Components Analysis W p (PCA), is the Orthogonal (global) solution [WWT = I]. The PCA solution to Eq.(l) is W p = n-tET, where n is the diagonal matrix of eigenvalues, and E is the matrix who's columns are the eigenvectors. The filters (rows of W p) are orthogonal, are thus the same as the PCA basis functions, and are typically global Fourier filters, ordered according to the amplitude spectrum of the image. Example PCA filters are shown in Fig.1a. (2) Zero-phase Components Analysis W z (ZCA), is the Symmetrical (local) solution [WWT = W 2]. The ZCA solution to Eq.(l) is W z = (XXT)-1/2 (matrix square root). ZCA is the polar opposite of PCA. It produces local (centre-surround type) whitening filEdges are the "Independent Components" of Natural Scenes 833 ters, which are ordered according to the phase spectrum of the image. That is, each filter whitens a given pixel in the image, preserving the spatial arrangement of the image and flattening its frequency (amplitude) spectrum. W z is related to the transforms described by Atick & Redlich [3]. Example ZCA filters and basis functions are shown in Fig.1b. (3) Independent Components Analysis WI (ICA), is the Factorised (semi-local) solution [/u(u) = TIi lUi (Ui)]. Please see [5] for full references. The 'infomax' solution we describe here is related to the approaches in [5, 1, 6]. As we will show, in Section 5, ICA on natural images produces decorrelating filters which are sensitive to both phase (locality) and spatial frequency information, just as in transforms involving oriented Gabor functions or wavelets. Example ICA filters are shown in Fig.1d and their corresponding basis functions are shown in Fig.1e. 3 An leA algorithm. It is important to recognise two differences between finding an ICA solution, WI, and other decorrelation methods. (1) there may be no ICA solution, and (2) a given ICA algorithm may not find the solution even if it exists, since there are approximations involved. In these senses, ICA is different from PCA and ZCA, and cannot be calculated analytically, for example, from second order statistics (the covariance matrix), except in the gaussian case. The approach 'which we developed in [5] (see there for further references to ICA) was to maximise by stochastic gradient ascent the joint entropy, H[g(u)], of the linear transform squashed by a sigmoidal function, g. When the non-linear function is the same (up to scaling and shifting) as the cumulative density functions (c.d.f.s) of the underlying independent components, it can be shown (Nadal & Parga 1995) that such a non-linear 'infomax' procedure also minimises the mutual information between the Ui, exactly what is required for ICA. In most cases, however, we must pick a non-linearity, g, without any detailed knowledge of the probability density functions (p.d.f.s) of the underlying independent components. In cases where the p.d.f.s are super-gaussian (meaning they are peakier and longer-tailed than a gaussian, having kurtosis greater than 0), we have repeatedly observed, using the logistic or tanh nonlinearities, that maximisation of H[g(u)] still leads to ICA solutions, when they exist, as with our experiments on speech signal separation [5]. Although the infomax algorithm is described here as an ICA algorithm, a fuller understanding needs to be developed of under exactly what conditions it may fail to converge to an ICA solution. The basic infomax algorithm changes weights according to the entropy gradient. Defining Yi = g( Ui) to be the sigmoidally transformed output variables, the stochastic gradient learning rule is: (2) In this, E denotes expected value, y = [g(ud ... g(UN)V, and IJI is the absolute value of the determinant of the Jacobian matrix: J = det [8Yi/8xj]ij' and y = [Y1 ... YN]T, the elements of which depend on the nonlinearity according to: Yi = 8/8Yi(8yi/8ui). Amari, Cichocki & Yang [1] have proposed a modification of this rule which utilises the natural gradient rather than the absolute gradient of H(y). The natural gradient exists for objective functions which are functions of matrices, as in this case, and is the same as the relative gradient concept developed by Cardoso & Laheld (1996). It amounts to multiplying 834 A. 1. Bell and T. 1. Sejnowski the absolute gradient by WTW, giving, in our case, the following altered version of Eq.(2): (3) This rule has the twin advantages over Eq.(2) of avoiding the matrix inverse, and of converging several orders of magnitude more quickly, for data, x, that is not prewhitened. The speedup is explained by the fact that convergence is no longer dependent on the conditioning of the underlying basis function matrix, A. Writing Eq.(3) for one weight gives ~Wij ex: Wij + iii l:k WkjUk. This rule is 'almost local' requiring a backwards pass. (a) peA 1 5 7 11 15 22 37 60 63 89 109 144 (b) (c) (d) zeA w leA (e) A D Figure 1: Selected decorrelating filters and their basis functions extracted from the natural scene data. Each type of decorrelating filter yielded 144 12x12 filters, of which we only display a subset here. Each column contains filters or basis functions of a particular type, and each of the rows has a number relating to which row of the filter or basis function matrix is displayed. (a) PCA (W p): The 1st, 5th, 7th etc Principal Components, showing increasing spatial frequency. There is no need to show basis functions and filters separately here since for PCA, they are the same thing. (b) ZCA (W z): The first 6 entries in this column show the one-pixel wide centresurround filter which whitens while preserving the phase spectrum. All are identical, but shifted. The lower 6 entries (37, 60 show the basis functions instead, which are the columns of the inverse of the W z matrix. ( c) W: the weights learnt by the lCA network trained on W z-whitened data, showing (in descending order) the DC filter, localised oriented filters, and localised checkerboard filters. (d) WI: The corresponding lCA filters, calculated according to WI = WW z, looking like whitened versions of the Wfilters. (e) A: the corresponding basis functions, or columns of WI!' These are the patterns which optimally stimulate their corresponding lCA filters, while not stimulating any other lCA filter, so that W1A = I. Edges are the "Independent Components" of Natural Scenes 835 4 Methods. We took four natural scenes involving trees, leafs etc., and converted them to greyscale values between 0 and 255. A training set, {x}, was then generated of 17,59512x12 samples from the images. This was 'sphered' by subtracting the mean and multiplying by twice the local symmetrical (zero-phase) whitening filter: {x} +- 2Wz({x} - (x)). This removes both first and second order statistics from the data, and makes the covariance matrix of x equal to 41. This is an appropriately scaled starting point for further training since infomax (Eq.(3)) on raw data, with the logistic function, Yi = (1 + e-Ui)-t, produces au-vector which approximately satisfies (uuT ) = 41. Therefore, by prewhitening x in this way, we can ensure that the subsequent transformation, u = Wx, to be learnt should approximate an orthonormal matrix (rotation without scaling), roughly satisfying the relation WTW = I. The matrix, W, is then initialised to the identity matrix, and trained using the logistic function version of Eq.(3), in which Yi = 1 - 2Yi. Thirty sweeps through the data were performed, at the end of each of which, the order of the data vectors was permuted to avoid cyclical behaviour in the learning. In each sweep, the weights were updated in batches of 50 presentations. The learning rate (proportionality constant in Eq.(3)) followed 21 sweeps at 0.001, and 3 sweeps at each of 0.0005,0.0002 and 0.0001, taking 2 hours running MATLAB on a Sparc-20 machine, though a reasonable result for 12x12 filters can be achieved in 30 minutes. To verify that the result was not affected by the starting condition of W = I, the training was repeated with several randomly initialised weight matrices, and also on data that was not prewhitened. The results were qualitatively similar, though convergence was much slower. The full ICA transform from the raw image was calculated as the product of the sphering (ZCA) matrix and the learnt matrix: WI = WW z. The basis function matrix, A, was calculated as wIt. A PCA matrix, W p, was calculated. The original (unsphered) data was then transformed by all three decorrelating transforms, and for each the kurtosis of each of the 144 filters was calculated. Then the mean kurtosis for each filter type (ICA, PCA, ZCA) was calculated, averaging over all filters and input data. 5 Results. The filters and basis functions resulting from training on natural scenes are displayed in Fig.1 and Fig.2. Fig.1 displays example filters and basis functions of each type. The PCA filters, Fig.1a, are spatially global and ordered in frequency. The ZCA filters and basis functions are spatially local and ordered in phase. The ICA filters, whether trained on the ZCA-whitened images, Fig.le, or the original images, Fig.1d, are semi-local filters, most with a specific orientation preference. The basis functions, Fig.1e, calculated from the Fig.1d ICA filters, are not local, and naturally have the spatial frequency characteristics of the original images. Basis functions calculated from Fig.1d (as with PCA filters) are the same as the corresponding filters since the matrix W (as with W p) is orthogonal. In order to show the full variety of ICA filters, Fig.2 shows, with lower resolution, all 144 filters in the matrix W, in reverse order of the vector-lengths of the filters, so that the filters corresponding to higher-variance independent components appear at the top. The general result is that ICA filters are localised and mostly oriented. Unlike the basis functions displayed in Olshausen & Field (1996), they do not cover a broad range of spatial frequencies. However, the appropriate comparison to make is between the ICA basis functions, and the basis functions in Olshausen & Field's Figure 4. The ICA basis functions in our Fig.1e are 836 A. 1. Bell and T. 1. Sejnowski oriented, but not localised and therefore it is difficult to observe any multiscale properties. However, when we ran the ICA algorithm on Olshausen's images, which were preprocessed with a whitening/low pass filter, our algorithm yielded basis functions which were localised multiscale Gabor patches qualitively similar to those in Olshausen's Figure 4. Part of the difference in our results is therefore attributable to different preprocessing techniques. The distributions (image histograms) produced by PCA, ZCA and ICA are generally doubleexponential (e-1u;j), or 'sparse', meaning peaky with a long tail, when compared to a gaussian, as predicted by Field [7]. The log histograms are seen to be roughly linear across 5 orders of magnitude. The histogram for the ICA filters, however, departs slightly from linearity, being more peaked, and having a longer tail than the ZCA and PCA histograms. This spreading of the tail signals the greater sparseness of the outputs of the ICA filters, and this is reflected in a calculated kurtosis measure of 10.04 for ICA, compared to 3.74 for PCA, and 4.5 for ZCA. In conclusion, these simulations show that the filters found by the ICA algorithm of Eq.(3) with a logistic non-linearity are localised, oriented, and produce outputs distributions of very high sparseness. It is notable that this is achieved through an information theoretic learning rule which (1) has no noise model, (2) is sensitive to higher-order statistics (spatial coincidences), (3) is non-Hebbian (it is closer to anti-Hebbian) and (4) is simple enough to be almost locally implementable. Many other levels of higher-order invariance (translation, rotation, scaling, lighting) exist in natural scenes. It will be interesting to see if informationtheoretic techniques can be extended to address these invariances. Acknowledgements This work emerged through many extremely useful discussions with Bruno Olshausen and David Field. We are very grateful to them, and also to Paul Viola and Barak Pearlmutter. The work was supported by the Howard Hughes Medical Institute. References [1] Amari S. Cichocki A. & Yang H.H. 1996. A new learning algorithm for blind signal separation, Advances in Neural Information Processing Systems 8, MIT press. [2] Atick J.J. 1992. Could information theory provide an ecological theory of sensory processing? Network 3, 213-251 [3] Atick J.J. & Redlich A.N. 1993. Convergent algorithm for sensory receptive field development, Neural Computation 5, 45-60 [4] Barlow H.B. 1989. Unsupervised learning, Neural Computation 1, 295-311 [5] Bell A.J. & Sejnowski T.J. 1995. An information maximization approach to blind separation and blind deconvolution, Neural Computation, 7, 1129-1159 [6] Cardoso J-F. & Laheld B. 1996. Equivariant adaptive source separation, IEEE Trans. on Signal Proc., Dec.1996. [7] Field D.J. 1994. What is the goal of sensory coding? Neural Computation 6, 559-601 [8] Hubel D.H. & Wiesel T.N. 1968. Receptive fields and functional architecture of monkey striate cortex, J. Physiol., 195: 215-244 • Edges are the "Independent Components" of Natural Scenes 837 [9] Linsker R. 1988. Self-organization in a perceptual network. Computer, 21, 105-117 [10] Miller K.D. 1988. Correlation-based models of neural development, in Neuroscience and Connectionist Theory, M. Gluck & D. Rumelhart, eds., 267-353, L.Erlbaum, NJ [11] Nadal J-P. & Parga N. 1994. Non-linear neurons in the low noise limit: a factorial code maximises information transfer. Network, 5,565-581. [12] Olshausen B.A. & Field D.J. 1996. Natural image statistics and efficient coding, Network: Computation in Neural Systems, 7, 2 . • Figure 2: The matrix of 144 filters obtained by training on ZCA-whitened natural images. Each filter is a row of the matrix W, and they are ordered left-to-right, top-to-bottom in reverse order of the length of the filter vectors. In a rough characterisation, and more-or-Iess in order of appearance, the filters consist of one DC filter (top left), 106 oriented filters (of which 35 were diagonal, 37 were vertical and 34 horizontal), and 37 localised checkerboard patterns. The diagonal filters are longer than the vertical and horizontal due to the bias induced by having square, rather than circular, receptive fields.
1996
147
1,203
For valid generalization, the size of the weights is more important than the size of the network Peter L. Bartlett Department of Systems Engineering Research School of Information Sciences and Engineering Australian National University Canberra, 0200 Australia Peter .BartlettClanu .edu.au Abstract This paper shows that if a large neural network is used for a pattern classification problem, and the learning algorithm finds a network with small weights that has small squared error on the training patterns, then the generalization performance depends on the size of the weights rather than the number of weights. More specifically, consider an i-layer feed-forward network of sigmoid units, in which the sum of the magnitudes of the weights associated with each unit is bounded by A. The misclassification probability converges to an error estimate (that is closely related to squared error on the training set) at rate O((cA)l(l+1)/2J(log n)jm) ignoring log factors, where m is the number of training patterns, n is the input dimension, and c is a constant. This may explain the generalization performance of neural networks, particularly when the number of training examples is considerably smaller than the number of weights. It also supports heuristics (such as weight decay and early stopping) that attempt to keep the weights small during training. 1 Introduction Results from statistical learning theory give bounds on the number of training examples that are necessary for satisfactory generalization performance in classification problems, in terms of the Vapnik-Chervonenkis dimension of the class of functions used by the learning system (see, for example, [13, 5]). Baum and Haussler [4] used these results to give sample size bounds for multi-layer threshold networks Generalization and the Size of the Weights in Neural Networks 135 that grow at least as quickly as the number of weights (see also [7]). However, for pattern classification applications the VC-bounds seem loose; neural networks often perform successfully with training sets that are considerably smaller than the number of weights. This paper shows that for classification problems on which neural networks perform well, if the weights are not too big, the size of the weights determines the generalization performance. In contrast with the function classes and algorithms considered in the VC-theory, neural networks used for binary classification problems have real-valued outputs, and learning algorithms typically attempt to minimize the squared error of the network output over a training set. As well as encouraging the correct classification, this tends to push the output away from zero and towards the target values of { -1, I}. It is easy to see that if the total squared error of a hypothesis on m examples is no more than mf, then on no more than mf/(I- o? of these examples can the hypothesis have either the incorrect sign or magnitude less than a. The next section gives misclassification probability bounds for hypotheses that are "distinctly correct" in this way on most examples. These bounds are in terms of a scale-sensitive version of the VC-dimension, called the fat-shattering dimension. Section 3 gives bounds on this dimension for feedforward sigmoid networks, which imply the main results. The proofs are sketched in Section 4. Full proofs can be found in the full version [2]. 2 Notation and bounds on misclassification probability Denote the space of input patterns by X. The space of labels is {-I, I}. We assume that there is a probability distribution P on the product space X x { -1, I}, that reflects both the relative frequency of different input patterns and the relative frequency of an expert's classification of those patterns. The learning algorithm uses a class of real-valued functions, called the hypothesis class H. An hypothesis h is correct on an example (x, y) if sgn(h(x)) = y, where sgn(a) : 1R -+ {-I, I} takes value 1 iff a 2: 0, so the misclassification probability (or error) of h is defined as erp(h) = P {(x, y) E X x {-I, I} : sgn(h(x)) "# y} . The crucial quantity determining misclassification probability is the fat-shattering dimension of the hypothesis class H. We say that a sequence Xl, ... , X d of d points from X is shattered by H iffunctions in H can give all classifications of the sequence. That is, for all b = (bl , ... , bm) E {-I, l}m there is an h in H satisfying sgn(h(xi)) = bi . The VC-dimension of H is defined as the size of the largest shattered sequence. l For a given scale parameter, > 0, we say that a sequence Xl, ... , Xd of d points from X is ,-shattered by H if there is a sequence rl, ... , rd of real values such that for all b = (bl , ... , bm) E {-I, l}m there is an h in H satisfying (h(xd - rdbi 2: ,. The fat-shattering dimension of H at " denoted fatH(!), is the size of the largest ,-shattered sequence. This dimension reflects the complexity of the functions in the class H, when examined at scale ,. Notice that fatH(!) is a nonincreasing function of ,. The following theorem gives generalization error bounds in terms of fatH(!). A related result, that applies to the case of no errors on the training set, will appear in [12]. Theorem 1 Define the input space X, hypothesis class H, and probability distribution P on X x {-I, I} as above. Let ° < 0 < 1/2, and ° < , < 1. Then, with probability 1- 0 over the training sequence (Xl, YI), ... , (xm, Ym) of m labelled 1 In fact, according to the usual definition, this is the VO-dimension of the class of thresholded versions of functions in H. 136 P. L. Bartlett examples, every hypothesis h in H satisfies erp(h) < ~ I{i : Ih(xdl < , or sgn(h(xt)) I- Yi}1 + fb, m, 6), m where 2 f2b, m, 6) = (d In(50em/d) log2(1250m) + In(4/6)) , m (1) and d = fatHb/16). 2.1 Comments It is informative to compare this result with the standard VC-bound. In that case, the bound on misclassification probability is 1 ( C ) 1/2 erp(h) < m I{i : sgn(h(xi)) I- ydl + m (dlog(m/d) + log(I/6)) , where d = VCdim(H) and c is a constant. We shall see in the next section that there are function classes H for which VCdim(H) is infinite but fatHb) is finite for all, > 0; an example is the class of functions computed by any two-layer neural network with an arbitrary number of parameters but constraints on the size of the parameters. It is known that if the learning algorithm and error estimates are constrained to make use of the sample only by considering the proportion of training examples that hypotheses misclassify, there are distributions P for which the second term in the VC-bound above cannot be improved by more than log factors. Theorem 1 shows that it can be improved if the learning algorithm makes use of the sample by considering the proportion of training examples that are correctly classified and have Ih(xdl < ,. It is possible to give a lower bound (see the full paper [2]) which, for the function classes considered here, shows that Theorem 1 also cannot be improved by more than log factors. The idea of using the magnitudes of the values of h(xd to give a more precise estimate of the generalization performance was first proposed by Vapnik in [13] (and was further developed by Vapnik and co-workers). There it was used only for the case of linear hypothesis classes. Results in [13] give bounds on misclassification probability for a test sample, in terms of values of h on the training and test data. This result is extended in [11], to give bounds on misclassification probability (that is, for unseen data) in terms of the values of h on the training examples. This is further extended in [12] to more general function classes, to give error bounds that are applicable when there is a hypothesis with no errors on the training examples. Lugosi and Pinter [9] have also obtained bounds on misclassification probability in terms of similar properties of the class of functions containing the true regression function (conditional expectation of Y given x). However, their results do not extend to the case when the true regression function is not in the class of real-valued functions used by the estimator. It seems unnatural that the quantity, is specified in advance in Theorem 1, since it depends on the examples. The full paper [2] gives a similar result in which the statement is made uniform over all values of this quantity. 3 The fat-shattering dimension of neural networks Bounds on the VC-dimensionofvarious neural network classes have been established (see [10] for a review), but these are all at least linear in the number of parameters. In this section, we give bounds on the fat-shattering dimension for several neural network classes. Generalization and the Size of the Weights in Neural Networks 137 We assume that the input space X is some subset of ]Rn. Define a sigmoid unit as a function from ]R k to ]R, parametrized by a vector of weights w E ]R k • The unit computes x t-7 0'( X • w), where 0' is a fixed bounded function satisfying a Lipchitz condition. (For simplicity, we ignore the offset parameter. It is equivalent to including an extra input with a constant value.) A multi-layer feed-forward sigmoid network of depth £ is a network of sigmoid units with a single output unit, which can be arranged in a layered structure with £ layers, so that the output of a unit passes only to the inputs of units in later layers. We will consider networks in which the weights are bounded. The relevant norm is the £1 norm: for a vector w E ]Rk, define IIwl11 = 2:7=1 IWil. The following result gives a bound on the fatshattering dimension of a (bounded) linear combination of real-valued functions, in terms of the fat-shattering dimension of the basis function class. We can apply this result in a recursive fashion to give bounds for single output feed-forward networks. Theorem 2 Let F be a class of functions that map from X to [-M/2, M/2], such that 0 E F and, for all f in F, - f E F. For A > 0, define the class H of weight-bounded linear combinations of functions from F as H = {t Wdi : k E W, fi E F, "will ~ A} . ,=1 Suppose , > 0 is such that d = fatFb/(32A)) 2: 1. Then fatHb) ~ (cM2 A 2d/,2) log2(M Ad/,), for some constant c. Gurvits and Koiran [6] have shown that the fat-shattering dimension of the class of two-layer networks with bounded output weights and linear threshold hidden units is a ((A2n2 h 2) log(n/,y)), when X = lRn. As a special case, Theorem 2 improves this result. Notice that the fat-shattering dimension of a function class is not changed by more than a constant factor if we compose the functions with a fixed function satisfying a Lipschitz condition (like the standard sigmoid function) . Also, for X = ]Rn and H = {x t-7 xd we have fatHb) ~ logn for all 'Y. Finally, for H = {x t-7 W· x: w E ]Rn} we have fatHb) ~ n for all 'Y. These observations, together with Theorem 2, give the following corollary. The 0(·) notation suppresses log factors. (Formally, f = O(g) if f = o(g1+O:) for all 0' > 0.) Corollary 3 If X C lR n and H is the class of two-layer sigmoid networks with the weights in the outp;;i unit satisfying IIwlh ~ A, then fatHb) = 6 (A2n/'Y2). If X = {x E lRn : Ilxlioo ~ B} and the hidden unit weights are also bounded, then fatHb) = 0 (B2 A6 (log n)h4). Applying Theorem 2 to this result gives the following result for deeper networks. Notice that there is no constraint on the number of hidden units in any layer, only on the total magnitude of the weights associated with a processing unit. Corollary 4 For some constant c, if X C ]Rn and H is the class of depth £ sigmoid networks in which the weight vector w associated with each unit beyond the first layer satisfies IIwlll ~ A, then fatHb) = () (n(cA)l(l-1)h2(l-1)) . If X = {x E lRn : IIxlioo ~ B} and the weights in the first layer units also satisfy IIwll1 ~ A, then fatHb) = () (B2(cA)l(l+1) /'Y2llog n). In the first part of this corollary, the network has fat-shattering dimension similar to the VC-dimension of a linear network. This formalizes the intuition that when the weights are small, the network operates in the "linear part" of the sigmoid, and so behaves like a linear network. 138 P. L. Bartlett 3.1 Comments Consider a depth '- sigmoid network with bounded weights. The last corollary and Theorem 1 imply that if the training sample size grows roughly as B2 Al2 /f2, then the misclassification probability of a network is within f of the proportion of training examples that the network classifies as "distinctly correct." These results give a plausible explanation for the generalization performance of neural networks. If, in applications, networks with many units have small weights and small squared error on the training examples, then the VC-dimension (and hence number of parameters) is not as important as the magnitude of the weights for generalization performance. It is possible to give a version of Theorem 1 in which the probability bound is uniform over all values of a complexity parameter indexing the function classes (using the same technique mentioned at the end of Section 2.1). For the case of sigmoid network classes, indexed by a weight bound, minimizing the resulting bound on misclassification probability is equivalent to minimizing the sum of a sample error term and a penalty term involving the weight bound. This supports the use of two popular heuristic techniques, weight decay and early stopping (see, for example, [8]), which aim to minimize squared error while maintaining small weights. These techniques give bounds on the fat-shattering dimension and hence generalization performance for any function class that can be expressed as a bounded number of compositions of either bounded-weight linear combinations or scalar Lipschitz functions with functions in a class that has finite fat-shattering dimension. This includes, for example, radial basis function networks. 4 Proofs 4.1 Proof sketch of Theorem 1 For A f S,' where (S, p) is a pseudometric space, a set T <; S is an f-cover of A if for all a in A there is a tin T with p(t, a) < f.. We define N(A, f, p) as the size of the smallest f-cover of A . For x = (Xl, . . . , Xm) E X m , define the pseudometric dloo(x) on the set S of functions defined on X by dloo(x)(f,g) = max; If(xd - g(x;)I. For a set A of functions, denote maxxEx",N(A,f,dloo(x)) by Noo(A,f,m). Alon et al. [1] have obtained the following bound on Noo in terms of the fat-shattering dimension. Lemma 5 For a class F of functions that map from {I, .. . , n} to {I, ... , b} with fatF(l) ~ d, log2Noo(F,2,n) < 1 + log2(nb2)log2 (L:1=o (7)bi ), provided that n ~ 1 + log2 (L:1=o (7)bi ). For, > 0 define 7r"'( : lR -+ [-",] as the piecewise-linear squashing function satisfying 7r"'((a) = , if a ~ " 7r"'((a) = -, if a ~ -" and 7r"'((a) = a otherwise. For a class H of real-valued functions, define 7r"'((H) as the set of compositions of 7r"'( with functions in H. Lemma 6 For X, H, P, 6, and, as in Theorem 1, pm { z : 3h E H, erp(h) 2: (! In CNoo(",(H), '1/2, 2m)) ) 1/2 + ~ I{i: Ih(z.)1 < 'I orsgn(h(z.)) # !Ii}I} < 6. Generalization and the Size o/the Weights in Neural Networks 139 The proof of the lemma relies on the observation that pm {z : 3h E H, erp(h) 2: f + ! I{i : Ih(xi)1 <, or sgn(h(xd) =f. ydl} < pm {z : 3h E H, P(11I""((h(x)) -,yl2: ,) 2: f+! I{i : 1I""((h(xd) =f. ,ydl}. We then use a standard symmetrization argument and the permutation argument introduced by Vapnik and Chervonenkis to bound this probability by the probability under a random permutation of a double-length sample that a related property holds. For any fixed sample, we can then use Pollard's approach of approximating the hypothesis class using a {J/2)-cover, except that in this case the appropriate cover is with respect to the /00 pseudometric. Applying Hoeffding's inequality gives the lemma. To prove Theorem 1, we need to bound the covering numbers in terms of the fatshattering dimension. It is easy to apply Lemma 5 to a quantized version of the function class to get such a bound, taking advantage of the range constraint imposed by the squashing function 11""( . 4.2 Proof sketch of Theorem 2 For x = (Xl"'" Xm) E X m, define the pseudometric dl1(x) on the class of functions defined on X by dl1(x)(f,g) = ~ L:;:l I/(xi) - g(Xi)/. Similarly, define dl2(x)(f,g) = (~ L:;:I(f(Xi) - g(xd)2)1/2 . If A is a set of functions defined on X, denote maxxEXm N(A", dl1(x)) by Nt{A", m), and similarly for N 2(A , " m) . The idea of the proof of Theorem 2 is to first derive a general upper bound on an /1 covering number of the class H, and then apply the following result (which is implicit in the proof of Theorem 2 in [3]) to give a bound on the fat-shattering dimension. Lemma 7 For a class F 01 [0, I]-valued functions on X satisfying fatF(4,) 2: d, we have log2 Nt{F" , d) 2: d/32. To derive an upper bound on N1{J, H, m), we start with the bound that Lemma 5 implies on the /00 covering number Noo(F", m) for the class F of hidden unit functions. Since dl.A/, g) :::; dl oo (f, g) , this implies the following bound on the /2 covering number for F (provided m satisfies the condition required by Lemma 5, and it turns out that the theorem is trivial otherwise). ( 4emM) (9mM2) log2 N2(F", m) < 1 + dlog2 d, log2 ,2 . (2) Next, we use the following result on approximation in /2 , which A. Barron attributes to Maurey. Lemma 8 (Maurey) Suppose G is a Hilbert space and F ~ G has 11/11:::; b for all I in F. Let I be an element from the convex closure of F. Then for all k 2: 1 and all c> b2_1//I/2, there are functions {It , · · . Jd ~ F such that III - ~ L:7=1 lil1 2 :::; ~ . This implies that any element of H can be approximated to a particular accuracy (with respect to /2) using a fixed linear combination of a small number of elements of F . It follows that we can construct an /2 cover of H from the /2 cover of F; using Lemma 8 and Inequality 2 shows that 2M2A2 ( (8emMA) (36mM2A2)) log2 N2(H", m) < ,2 1 + dlog2 d, log2 ,2 . 140 P. L. Bartlett Now, Jensen's inequality implies that dI1(x)(f,g) ~ dI2(x)(f,g), which gives a bound on Nt (H, 1, m). Comparing with the lower bound given by Lemma 7 and solving for m gives the result. A more refined analysis for the neural network case involves bounding N2 for successive layers, and solving to give a bound on the fat-shattering dimension of the network. Acknowledgements Thanks to Andrew Barron, Jonathan Baxter, Mike Jordan, Adam Kowalczyk, Wee Sun Lee, Phil Long, John Shawe-Taylor, and Robert Slaviero for helpful discussions and comments. References [1] N. Alon, S. Ben-David, N. Cesa-Bianchi, and D. Haussler. Scale-sensitive dimensions, uniform convergence, and learn ability. In Proceedings of the 1993 IEEE Symposium on Foundations of Computer Science. IEEE Press, 1993. [2] P. L. Bartlett. The sample complexity of pattern classification with neural networks: the size of the weights is more important than the size of the network. Technical report, Department of Systems Engineering, Australian National University, 1996. (available by anonymous ftp from syseng.anu.edu.au:pub/peter/TR96d.ps). [3] P. L. Bartlett, S. R. Kulkarni, and S. E. Posner. Covering numbers for realvalued function classes. Technical report, Australian National University and Princeton University, 1996. [4] E. Baum and D. Haussler. What size net gives valid generalization? Neural Computation, 1(1):151-160,1989. [5] A. Blumer, A. Ehrenfeucht, D. Haussler, and M. K. Warmuth. Learnability and the Vapnik-Chervorienkis dimension. J. ACM, 36(4):929-965, 1989. [6] L. Gurvits and P. Koiran. Approximation and learning of convex superpositions. In Computational Learning Theory: EUROCOLT'95, 1995. [7] D. Haussler. Decision theoretic generalizations of the PAC model for neural net and other learning applications. Inform. Comput., 100(1):78-150,1992. [8] J. Hertz, A. Krogh, and R. G. Palmer. Introduction to the Theory of Neural Computation. Addison-Wesley, 1991. [9] G. Lugosi and M. Pinter. A data-dependent skeleton estimate for learning. In Proc. 9th Annu. Conference on Comput. Learning Theory. ACM Press, New York, NY, 1996. [10] W. Maass. Vapnik-Chervonenkis dimension of neural nets. In M. A. Arbib, editor, The Handbook of Brain Theory and Neural Networks, pages 1000-1003. MIT Press, Cambridge, 1995. [11] J. Shawe-Taylor, P. 1. Bartlett, R. C. Williamson, and M. Anthony. A framework for structural risk minimisation. In Proc. 9th Annu. Conference on Comput. Learning Theory. ACM Press, New York, NY, 1996. [12] J. Shawe-Taylor, P. 1. Bartlett, R. C. Williamson, and M. Anthony. Structural risk minimization over data-dependent hierarchies. Technical report, 1996. [13] V. N. Vapnik. Estimation of Dependences Based on Empirical Data. SpringerVerlag, New York, 1982.
1996
148
1,204
Neural Network Modeling of Speech and Music Signals Axel Robel Technical University Berlin, Einsteinufer 17, Sekr. EN-8, 10587 Berlin, Germany Tel: +49-30-31425699, FAX: +49-30-31421143, email: roebel@kgw.tu-berlin.de Abstract Time series prediction is one of the major applications of neural networks. After a short introduction into the basic theoretical foundations we argue that the iterated prediction of a dynamical system may be interpreted as a model of the system dynamics. By means of RBF neural networks we describe a modeling approach and extend it to be able to model instationary systems. As a practical test for the capabilities of the method we investigate the modeling of musical and speech signals and demonstrate that the model may be used for synthesis of musical and speech signals. 1 Introduction Since the formulation of the reconstruction theorem by Takens [10] it has been clear that a nonlinear predictor of a dynamical system may be directly derived from a systems time series. The method has been investigated extensively and with good success for the prediction of time series of nonlinear systems. Especially the combination of reconstruction techniques and neural networks has shown good results [12]. In our work we extend the ideas of predicting nonlinear systems by the more demanding task of building system models, which are able to resynthesize the systems time series. In the case of chaotic or strange attractors the resynthesis of identical time series is known to be impossible. However, the modeling of the underlying attractor leads to the possibility to resynthesis time series which are consistent with the system dynamics. Moreover, the models may be used for the analysis of the system dynamics, for example the estimation of the Lyapunov exponents [6]. In the following we investigate the modeling of music and speech signals, where the system dynamics are known to be instationary. Therefore, we 780 A. Robel develop an extension of the modeling approach, such that we are able to handle instationary systems. In the following, we first give a short review concerning the state space reconstruction from time series by delay coordinate vectors, a method that has been introduced by Takens [10] and later extended by Sauer et al. [9]. Then we explain the structure of the neural networks we used in the experiments and the enhancements necessary to be able to model instationary dynamics. As an example we apply the neural models to a saxophone tone and a speech signal and demonstrate that the signals may be resynthesized using the neural models. Furthermore, we discuss some of the problems and outline further developments of the application. 2 Reconstructing attractors Assume an n-dimensional dynamical system f(-) evolving on an attractor A. A has fractal dimension d, which often is considerably smaller then n. The system state z is observed through a sequence of measurements h(Z), resulting in a time series of measurements Yt = h(z(t)). Under weak assumptions concerning h(-) and fe) the fractal embedding theorem[9] ensures that, for D > 2d, the set of all delayed coordinate vectors (1) with an arbitrary delay time T, forms an embedding of A in the D-dimensional reconstruction space. We call the minimal D, which yields an embedding of A, the embedding dimension De. Because an embedding preserves characteristic features of A, especially it is one to one, it may be employed for building a system model. For this purpose the reconstruction of the attractor is used to uniquely identify the systems state thereby establishing the possibility of uniquely predicting the systems evolution. The prediction function may be represented by a hyperplane over the attractor in an (D + 1) dimensional space. By iterating this prediction function we obtain a vector valued system model which, however, is valid only at the respective attractor. For the reconstruction of instationary systems dynamics we confine ourselves to the case of slowly varying parameters and model the in stationary system using a sequence of attractors. 3 RBF neural networks There are different topologies of neural networks that may be employed for time series modeling. In our investigation we used radial basis function networks which have shown considerably better scaling properties, when increasing the number of hidden units, than networks with sigmoid activation function [8]. As proposed by Verley sen et. al [11] we initialize the network using a vector quantization procedure and then apply backpropagation training to finally tune the network parameters. The tuning of the parameters yields an improvement factor of about ten in prediction error compared to the standard RBF network approach [8, 3]. Compared to earlier results [7] the normalization of the hidden layer activations yields a small improvement in the stability of the models. The resulting network function for m-dimensional vector valued output is of the form _ exp (_(C'(7-X)2) _ N(x) = "w· .I + b 7 JZ:::iexp(-(Cl(7~X)2) , (2) Neural Network Modeling of Speech and Music Signals k(i) Control x(i+1) x(i+ TL-----~, ,'~, " "-, / ~" \, " " , ,-,' x(i-3Tj - x(i-T) x(i-2T) x(i) Fig. I: Input/Output structure of the neural model. , , , , , , , , , , 781 where (T J represents the standard deviation of the Gaussian, the input x and the centers c are n-dimensional vectors and band Wj are m-dimensional parameters of the network. Networks of the form eq. (2) with a finite number of hidden units are able to approximate arbitrary closely all continuous mappings Rn -+ Rm [4]. This universal approximation property is the foundation of using neural networks for time series modeling, where we denote them as neural models. In the context of the previous section the neural models are approximating the systems prediction function. To be able to represent instationary dynamics, we extend the network according to figure 1 to have an additional input, that enables the control of the actual mapping (3) This model is close to the Hidden Control Neural Network described in [2]. From the universal approximation properties of the RBF-networks stated above it follows, that eq. (3) with appropriate control sequence k( i) is able to approximate any sequence of functions. In the context of time series prediction the value i represents the actual sample time. The control sequence may be optimized during training, as described in [2], The optimization of k( i) requires prohibitively large computational power if the number of different control values, the domain of k is large. However, as long as the systems instationarity is described by a smooth function of time, we argue that it is possible to select k( i) to be a fixed linear function of i. With the preselected k(i) the training of the network adapts the parameters tj and (Ttj such that the model evolution closely follows the systems instationarity. 4 Neural models As is shown in figure I we use the delayed coordinate vectors and a selected control sequence to train the network to predict the sequence of the following T time samples. The 782 A. Robel vector valued prediction avoids the need for a further interpolation of the predicted samples. Otherwise, an interpolation would be necessary to obtain the original sample frequency, but, because the Nyquist frequency is not regarded in choosing T, is not straightforward to achieve. After training we initialize the network input with the first input vector (X'o, k(O)) of the time series and iterate the network function shifting the network input and using the latest output unit to complete the new input. The control input may be copied from the training phase to resynthesize the training signal or may be varied to emulate another sequence of system dynamics. The question that has to be posed in this context is concerned with the stability of the model. Due to the prediction error of the model the iteration will soon leave the reconstructed attractor. Because there exists no training data from the neighborhood of the attractor the minimization of the prediction error of the network does not guaranty the stability of the model [5). Nevertheless, as we will see in the examples, the neural models are stable for at least some parameters D and T. Due to the high density of training data the method for stabilizing dynamical models presented in [5) is difficult to apply in our situation. Another approach to increase the model stability is to lower the gradient of the prediction function for the directions normal to the attractor. This may be obtained by disturbing the network input during training with a small noise level. While conceptually straightforward, we found that this method is only partly successful. While the resulting prediction function is smoother in the neighborhood of the attractor, the prediction error for training with noise is considerably higher as expected from the noise free results, such that the overall effect often is negative. To circumvent the problems of training with noise further investigations will consider a optimization function with regularization that directly penalizes high derivatives of the network with respect to the input units [1). The stability of the models is a major subject of further research. 5 Practical results We have applied our method to two acoustic time series, a single saxophone tone, consisting of 16000 samples sampled at 32kHz and a speech signal of the word manna l . The latter time series consists of 23000 samples with a sampling rate of 44.1kHz. Both time series have been normalized to stay within the interval [-1, 1]. The estimation of the dimension of the underlying attractors yields a dimension of about 2-3 in both cases. We chose the control input k(i) to be linear increasing from -0.8 to 0.8. Stable models we found for both time series using D > 5. Namely for the parameter T we observed considerable impact on the model quality. While smaller T results in better one step ahead prediction, the iterated model often becomes unstable. This might be explained by the decrease in variation within the prediction hyperplane, that has to be learned. For small T the model tends to become linear and does not capture the nonlinear characteristics of the system. Therefore the iteration of those models failed. To large values of T results in an insufficient one step ahead prediction error, which pushes the model far away from the attractor also producing unstable behavior. 'The name of our parallel computer Neural Network Modeling of Speech and Music Signals Saxophone (synthesized) 0.8 0.6 0.4 0.2 Iii' « 0 ;g. is -0.2 Cl. -0.4 -0.6 -0.8 -1 SOOO 10000 15000 20 10 0 -10 -20 -30 -40 -50 -60 -70 0 0.1 Power spectrum estimation 0.2 Original ReSynth _. __ .. 0.3 0.4 fIfO 783 0.5 Fig. 2: Synthesized saxophone signal and power spectrum estimation for the original (solid) and synthesized (dashed) signal. control input 0.8 '---~--~----'-''---' -aJ~--''' / OriCg~~~: 70.6 0.4 0.2 o -0.2 -0.4 -0.6 . ________________________ .,/0 1/ / // -0.8 "----~--~-~---'-----' 5000 1 0000 15000 20000 2S000 Saxophone (synthesized with long cnt) 0.8 0.6 0.4 0.2 « 0 -0.2 -0.4 -0.6 -0.8 -1 5000 10000 15000 20000 25000 Fig. 3: Varying the synthesized tone by varying the control input sequence. 5.1 Modeling a saxophone In the following we consider the results for the saxophone model. The model we present consists of 10 input units, 200 hidden units and 5 output units and was trained with additional Gaussian noise at the input. The standard deviation of the noise is 0.0005 and the RMS training error obtained is 0.005. The resulting saxophone model is able to resynthesize a signal which is nearly indistinguishable from the original one. The resynthesized time series is shown in figure 2. The time series follows the original one with a small phase shift, which stems from a small difference in the onset of the model. Also in figure 2 the power spectrum of the saxophone signal and the neural model is shown. From the spectrum we see the close resemblance of the sound. One major demand for the practical application of the proposed musical instrument models is the possibility to control the synthesized sound. At the present state there exists only one control input to the model. Nevertheless, it is interesting to investigate the effect of varying the control input of the model. We tried different control input sequences to synthesize saxophone tones. It turns out that the model reinains stable such that we are able to control the envelope of the sound. An example of a tone with increased duration is shown in figure 3. In this example the control input first follows the trained version, then remains constant to produce a longer duration of the tone and then increases to reproduce the decay of the tone from the trained time series. 784 A. Robel training signal (word: manna) resynthsized signal (word: manna) 0.8 0.8 0.6 0.6 0.4 0.4 02 0.2 « 0 « 0 -0.2 -0.2 -0.4 -0.4 -0.6 -0.6 -0.8 -0.8 -1 -1 0 5000 1 0000 15000 20000 25000 0 5000 10000 15000 20000 25000 Fig. 4: Original and synthesized signal of the word manna. 5.2 Modeling a speech signal For modeling the time series of the spoken word manna we used a similar network compared to the saxophone model. Due to the increased instationarity in the signal we needed an increased number of RBF units in the network. The best results up to now has been obtained with a network of 400 hidden units, delay time T = 8, output dimension 8 and input dimension 11. In figure 4 we show the original and the resynthesized signal. The quality of the model is not as high as in the case of the saxophone_ Nevertheless, the word is quite understandable. From the figure we see, that the main problems stem from the transitions between consecutive phonemes. These transitions are rather quick in time and, therefore, there exists only a small amount of data describing the dynamics of the transitions. We assume that more training examples of the same word will cure the problem. However, it will probably require a well trained speaker to reproduce the dynamics in speaking the same word twice_ 6 Further developments There are two practical applications that directly follow from the presented results_ The first one is to synthesize music signals. To consider musicians demands, we need to enhance the control of the synthesized signals. Therefore, in the future we will try to enlarge the models, incorporating different flavors of sound into the same model and adding additional control inputs. Especially we plan to build models for different volume and pitch. As a second application we will further investigate the possibilities for using the neural models as a speech synthesizer. An interesting topic of further research would be the extension of the model with an intonation control input that incorporates the possibility to synthesize different intonations of the same word from one model. 7 Summary The article describes a methodology to build instationary models from time series of dynamical systems. We give theoretical arguments for the universality of the models and discuss some of the restrictions and actual problems. As practical test for the method we apply the models. to the demanding task of the synthesis of musical and speech signals. It is demonstrated that the models are capable to resynthesize the trained signals. At the present Neural Network Modeling of Speech and Music Signals 785 state the envelope and duration of the synthesized signals may be controlled. Intended further developments have been shortly described. References [1] C. M. Bishop. Training with noise is equivalent to tikhonov regularization. Neural Computation, 7(1): 108-116, 1995. [2] E. Levin. Hidden control neural architecture modelling of nonlinear time varying systems and its applications. IEEE Transactions on Neural Networks, 4(2):109-116, 1993. [3] J. Moody and C. Darken. Fast learning in networks of locally-tuned processing units. Neural Computation, 1 :281-294, 1989. [4] J. Park and I. Sandberg. Universal approximation using radial-basis-function networks. Neural Computation, 3(2):246-257, 1991. [5] J. C. Principe and J.-M. Kuo. Dynamic modelling of chaotic time series with neural networks. In G. Tesauro, D. S. Touretzky, and T. Leen, editors, Neural Information Processing Systems 7 (NIPS 94), 1995. [6] A. Robel. Neural models for estimating lyapunov exponents and embedding dimension from time series of nonlinear dynamical systems. In Proceedings of the Intern. Conference on Artijicial Neural Networks, ICANN'95, Vol. II, pages 533-538, Paris, 1995. [7] A. Robel. Rbf networks for synthesis of speech and music signals. In 3. Workshop Fuzzy-Neuro-Systeme'95, Darmstadt, 1995. Deutsche Gesellschaft fur Informatik e.v. [8] A. Robel. Scaling properties of neural networks for the prediction of time series. In Proceedings of the 1996 IEEE Workshop on Neural Networks for Signal Processing VI, 1996. [9] T. Sauer, J. A. Yorke, and M. Casdagli. Embedology. Journal of Statistical Physics, 65(3/4):579-616, 1991. [to] F. Takens. Detecting Strange Attractors in Turbulence, volume 898 of Lecture Notes in Mathematics (Dynamical Systems and Turbulence, Warwick 1980), pages 366381. D.A. Rand and L.S. Young, Eds. Berlin: Springer, 1981. [11] M. Verleysen and K. Hlavackova. An optimized RBF network for approximation of functions. In Proceedings of the European Symposium on A rtijicial Neural Networks, ESANN'94,1994. [12] A. S. Weigend and N. A. Gershenfeld. Time Series Prediction: Forecasting the Future and Understanding the Past. Addison-Wesley Pub. Comp., 1993.
1996
149
1,205
Unification of Information Maximization and Minimization Ryotaro Kamimura Information Science Laboratory Tokai University 1117 Kitakaname Hiratsuka Kanagawa 259-12, Japan E-mail: ryo@cc.u-tokaLac.jp Abstract In the present paper, we propose a method to unify information maximization and minimization in hidden units. The information maximization and minimization are performed on two different levels: collective and individual level. Thus, two kinds of information: collective and individual information are defined. By maximizing collective information and by minimizing individual information, simple networks can be generated in terms of the number of connections and the number of hidden units. Obtained networks are expected to give better generalization and improved interpretation of internal representations. This method was applied to the inference of the maximum onset principle of an artificial language. In this problem, it was shown that the individual information minimization is not contradictory to the collective information maximization. In addition, experimental results confirmed improved generalization performance, because over-training can significantly be suppressed. 1 Introduction There have been many attempts to interpret neural networks from the information theoretical point of view [2], [4], [5]. Applied to the supervised learning, information has been maximized and minimized, depending on problems. In these methods, information is defined by the outputs of hidden units. Thus, the methods aim to control hidden unit activity patterns in an optimal manner. Information maximization methods have been used to interpret explicitly internal representations and simultaneously to reduce the number of necessary hidden units [5]. On the other hand, information minimization methods have been especially used to improve generalization performance [2], [4] and to speed up learning. Thus, if it is possible to Unification of Information Maximization and Minimization 509 maximize and minimize information simultaneously, information theoretic methods are expected to be applied to a wide range of problems. In this paper, we unify the above mentioned two methods, namely, information maximization and minimization methods, into one framework to improve generalization performance and to interpret explicitly internal representations. However, it is apparently impossible to maximize and minimize simultaneously the information defined by the hidden unit activity. Our goal is to maximize and to minimize information on two different levels, namely, collective and individual levels. This means that information can be maximized in collective ways and information is minimized for individual input-hidden connections. The seeming contradictory proposition of the simultaneous information maximization and minimization can be overcome by assuming the existence of the two levels for the information control. Information is supposed to be controlled by an information controller located outside neural networks and used exclusively to control information. By assuming the information controller, we can clearly see how information appropriately defined can be maximized or minimized. In addition, the actual implementation of information methods is much easier by introducing a concept of the information controller. 2 Concept of Information In this section, we explain a concept of information in a general framework of an information theory. Let Y take on a finite number of possible values Yl, Y2, ... , YM with probabilities P(Yl), P(Y2), ... , p(YM), respectively. Then, initial uncertainty H(Y) of a random variable Y is defined by M H(Y) = - L p(Yj) 10gp(Yj). ( 1) j=l Now, consider conditional uncertainty after the observation of another random variable X, taking possible values Xl, X2, ... , Xs with probabilities p(Xt},P(X2), ... ,p(XM), respectively. Conditional uncertainty H(Y I X) can be defined as S M H(Y I X) = - LP(x8) LP(Yj I x8) logp(xj I Y8)' (2) 8=1 j=1 We can easily verify that conditional uncertainty is always less than or equal to initial uncertainty. Information is usually defined as the decrease of this uncertainty [1]. I(Y I X) H(Y) - H(Y I X) M S M - L p(Yj ) 10gp(Yj) + L p( X8) L p(Yj I X,) 10gp(Yj I X8) j=1 8=1 j=1 ~ ~ () I) p(Yj I X8) L..,.; L..,.;P X8 p(Yj X8 log (.) . P YJ 8 J = LP(x8)I(Y I x8) (3) where I(Y I x,) 510 R. Kamimura j j which is referred to as conditional information. Especially, when prior uncertainty is maximum, that is, a prior probability is equi-probable (1/ M), then informlttion is S M I(Y I X) 10gM + I:p(x 3 ) I:p(Yj I x3 )logp(Yj I X 3 ) (5) 3=1 j=1 where log M is maximum uncertainty concering A. 3 Formulation of Information Controller In this section, we apply a concept of information to actual network architectures and define collective information and individual information. The notation in the above section is changed into ordinary notation used in the neural network. 3.1 Unification by Information Controller Two kinds of information, collective information and individual information, are controlled by using an information controller. The information controller is devised to interpret the mechanism of the information maximization and minimization more explicitly. As shown in Figure 1, the information controller is composed of two subcomponents, that is, an individual information minimizer and collective information ma.'Cimizer. A collective information maximizer is used to increase collective information as much as possible. An individual information minimizer is used to decrease individual information. By this minimization, the majority of connections are pushed toward zero. Eventually, all the hidden units tend to be intermediately activated. Thus, when the collective information maximizer and individual information maximizer are simultaneously applied, a hidden unit activity pattern is a pattern of the maximum information in which only one hidden unit is on, while all the other hidden units are off. However, multiple strongly negative connections to produce a maximum information state, are replaced by extremely weak inputhidden connections. Strongly negative connections are inhibited by the individual information minimization. This means that by the information controller, information can be maximized and at the same time one of the most important properties of the information minimization, namely, weight decay or weight elimination, can approximately be realized. Consequently, the information controller can generate much simplified networks in terms of hidden units and in terms of input-hidden connections. 3.2 Collective Information Maximizer A neural network to be controlled is composed of input, hidden and output units with bias, as shown in Figure 1. The jth hidden unit receives a net input from input units and at the same time from a collective information maximizer: L uj = Xj + I: Wjk~k (6) k=O where Xj is an information maximizer from the jth collective information maximizer to the jth hidden unit, L is the number of input units, Wjk is a connection from the kth input unit to the jth hidden unit and ~k is the kth element of the 8th input Unification of Information Maximization and Minimization Input- Hidden Connections Bias- Hidden Connections Bias-Output conn71ons Individual Information Minimizer Bias WiO . . . . • , ~ X \ \ j \. \ Information \ \ Maximizers ... \ '., .. • Collective Information Maximizer Information Controller .... ~~ I Target Figure 1: A network architecture, realizing the information controller. 511 pattern. The jth hidden unit produces an activity or an output by a sigmoidal activation function: vJ f(uJ) 1 (7) 1 + exp( -uJ) . The collective information maximizer is used to maximize the information contained in hidden units. For this purpose, we should define collective information. Now, suppose that in the previous formulation in information, a symbol X and Y represent a set of input patterns and hidden units respectively. Then, let us approximate a probability p(Yj I x$) by a normalized output pj of the jth hidden unit computed by (8) where the summation is over all the hidden units. Then, it is reasonable to suppose that at an initial stage all the hidden units are activated randomly or uniformly and all the input patterns are also randomly given to networks. Thus, a probability p(Yj) of the activation of hidden units at the initial stage is equi-probable, that is, 11M. A probability p(x$) of input patterns is also supposed to be equi-probable, namely, liS. Thus, information in the equation (3) is rewritten as I(Y I X) ~ M 1 1 1 S M -L M log M + S L L pJ 10gpJ j=l $=lj=l 512 R. Kamimura Input Unit (C) k Hidden Unit (0) J Fires Does not lire Figure 2: An interpretation of an input-hidden connection for defining the individual information. 1 S M log !If + S 2: 2:pj logpj 3=lj=1 (9) where log !If is maximum uncertainty. This information is considered to be the information acquired in a course oflearning. Information maximizers are updated to increase collective information. For obtaining update rules, we should differentiate the information function with respect to information maximizers Xj: where {3 is a parameter. {3soI(Y I X) OXj Ii ~ (lOg pj - f/:" log p:" ) pj (1 - vj) 3.3 Individual Information Minimization (10) For representing individual information by a concept of information discussed in the previous section, we consider an output Pjk from the jth hidden unit only with a connection from the kth input unit to the jth output unit: (11) which is supposed to be a probability of the firing of the jth hidden unit, given the firing ofthe kth input unit, as shown in Figure 2. Since this probability is considered to be a probability, given the firing of the kth input unit, conditional information is appropriate for measuring the information. In addition, it is reasonable to suppose that a probability of the firing of the jth hidden unit is 1/2 at an initial stage of learning, because we have no knowledge on hidden units. Thus, conditional information for a pair of the kth unit and the jth hidden unit is formulated as Ij k (D I fires) ~ - Pj k log ~ - (1 - Pj k ) log ( 1 - ~ ) +Pjk logpjk + (1 - Pjk) 10g(1- Pjk) log2 + Pjk logpjk + (1 - Pjk)log(l - Pjk) (12) If connections are close to zero, this function is close to minimum information, meaning that it is impossible. to estimate the firing of the kth hidden unit. If Unification of Information Maximization and Minimization 513 Table 1: An example of obtained input-hidden connections Wjle by the information controller. The parameter {3, It and 1} were 0.015, 0.0008, and 0.01. Hidden Input Units e: Information Units vJ 1 2 3 4 Bias WjO Maximizer Xj 1 3.09 10.77 26.48 13.82 22.07 -60.88 2 -3.35 0.11 0.33 -3.08 -0.95 1.63 3 -0.01 0.00 0.00 -0.01 0.00 -10.93 4 0.00 0.00 0.00 0.00 0.00 -10.94 5 0.00 0.00 -0.01 0.00 0.00 -10.97 6 0.02 0.01 -0.04 0.01 0.06 -12.01 7 0.00 0.00 -0.01 0.00 0.00 -11.01 8 0.00 0.00 -0.01 0.00 0.00 -11.00 9 0.02 0.01 -0.03 0.01 0.03 -11.61 10 0.01 0.00 -0.02 0.00 0.07 -11.67 connections are larger, the information is larger and correlation between input and hidden units is larger. Total individual information is the sum of all the individual individual information, namely, M L I(D I fires) 2: 2: Ijle(D I fires), (13) j=lle=O because each connection is treated separately or independently. The individual information minimization directly controls the input-hidden connections. By differentiating the individual information function and a cross entropy cost function with respect to input-hidden connections Wjle, we have rules for updating concerning input-hidden connections: ol(D I fi1>es) oG -It -1}-8wjle OWjle S -It Wjle pjle(1- Pjle) + 1} 2:c5J~k (14) !=1 where c5J is an ordinary delta for the cross entropy function and 1} and It are parameters. Thus, rules for updating with respect to input-hidden connections are closely related to the weight decay method. Clearly, as the individual information minimization corresponds to diminishing the strength of input-hidden connections. 4 Results and Discussion The information controller was applied to the segmentation of strings of an artificial language into appropriate minimal elements, that is, syllables. Table 1 shows input-hidden connections with the bias and the information maximizers. Hidden units were ordered by the magnitude of the relevance of each hidden unit [6]. Collective information and individual information could sufficiently be maximized and minimized. Relative collective and individual information were 0.94 and 0.13. In this state, all the input-hidden connections except connections into the first two hidden units are almost zero. Information maximizers Xj are all strongly negative for these cases. These negative information maximizers make eight hidden units (from the third to tenth hidden unit) inactive, that is, close to zero. By carefully 514 R. Kamimura Table 2: Generalization performance comparison for 200 and 200 training patterns. Averages in the table are average generalization errors over seven errors of ten errors with ten different initial values. Methods Standard Weight Decay Weight Elimination Information Controller Methods Standard Weight Decay Weight Elimination Information Controller (a) 200 patterns Generalization Errors RM S Error Rates Averages Std. Dev. Averages Std. Dev. 0.188 0.010 0.087 0.015 0.183 0.004 0.082 0.009 0.172 0.014 0.064 0.015 0.167 0.011 0.052 0.008 (b) 300 patterns Generalization Errors RMS Error Rates A verages Std. Dev. A verages Std. Dev. 0.108 0.009 0.024 0.009 0.110 0.003 0.012 0.004 0.083 0.005 0.009 0.006 0.072 0.006 0.008 0.004 examing the first two hidden units, we could see that the first hidden unit and the second hidden unit are concerned with rules for syllabification and a exceptional case. Then, networks were trained to infer the well-formedness of strings in addition to the segmentation to examine generalization performance. Table 2 shows generalization errors for 200 and 300 training patterns. As clearly shown in the figure, the best generalization performance in terms of RMS and error rates is obtained by the information controller. Thus, experimental results confirmed that in all cases the generalization performance of the information controller is well over the other methods. In addition, experimental results explicitly confirmed that better generalization performance is due to the suppression of over-training by the information controller. References [1] R. Ash, Information Theo1'1), John Wiley & Sons: New York, 1965. [2] G. Deco, W. Finnof and H. G. Zimmermann, "Unsupervised mutual information criterion for elimination of overtraining in Supervised Multilayer N etworks," Neural Computation, Vol. 7, pp.86-107, 1995. [3] R. Kamimura "Entropy minimization to increase the selectivity: selection and competition in neural networks," Intelligent Engineering Systems through Artificial Neural Networks, ASME Press, pp.227-232, 1992. [4] R. Kamimura, T. Takagi and S. Nakanishi, "Improving generalization performance by information minimization," IEICE Transactions on Information and Systems, Vol. E78-D, No.2, pp.163-173, 1995. [5] R. Kamimura and S. Nakanishi, "Hidden information maximization for feature detection and rule discovery," Network: Computation in Neural Systems, Vo1.6, pp.577-602, 1995. [6] M. C. Mozer and P. Smolen sky, "Using relevance to reduce network size automatically," Connection Science, Vo.l, No.1, pp.3-16, 1989.
1996
15
1,206
Rapid Visual Processing using Spike Asynchrony Simon J. Thorpe & Jacques Gautrais Centre de Recherche Cerveau & Cognition F-31062 Toulouse France email thorpe@cerco.ups-tlseJr Abstract We have investigated the possibility that rapid processing in the visual system could be achieved by using the order of firing in different neurones as a code, rather than more conventional firing rate schemes. Using SPIKENET, a neural net simulator based on integrate-and-fire neurones and in which neurones in the input layer function as analogto-delay converters, we have modeled the initial stages of visual processing. Initial results are extremely promising. Even with activity in retinal output cells limited to one spike per neuron per image (effectively ruling out any form of rate coding), sophisticated processing based on asynchronous activation was nonetheless possible. 1. INTRODUCTION We recently demonstrated that the human visual system can process previously unseen natural images in under 150 ms (Thorpe et aI, 1996). Such data, together with previous studies on processing speeds in the primate visual system (see Thorpe & Imbert, 1989) put severe constraints on models of visual processing. For example, temporal lobe neurones respond selectively to faces only 80-100 ms after stimulus onset (Oram & Perrett, 1992; Rolls & Tovee, 1994). To reach the temporal lobe in this time, information from the retina has to pass through roughly ten processing stages (see Fig. I). If one takes into account the surprisingly slow conduction velocities of intracortical axons « 1 ms-l, see Nowak & Bullier, 1997) it appears that the computation time within any cortical stage will be as little as 5-10 ms. Given that most cortical neurones will be firing below 100 spikes.sol , it is difficult to escape the conclusion that processing can be achieved with only one spike per neuron. 902 S. J. Thorpe and J. Gautrais Retina LGN VI V2 V4 PIT AIT r. 0 .. -a-.o -a-.o...a -a.o..n..oL. ..(") ~ ~ a-.~ ,...... ,...... ~ .~ r. ,...... 0 0 r. r. "" ,...... roo. ........ '" 0 '-' ~ ~ '-' ~ 0 '-' .~ 0-.('). ~~ ~ 30-SOms 40-60ms SO-70ms 60-BOms 70-90 ms SO-lOOms Figure 1 : Approximate latencies for neurones in different stages of the visual primate visual system (see Thorpe & Imbert, 1989; Nowak & Bullier, 1997). Such constraints pose major problems for conventional firing rate codes since at least two spikes are needed to estimate a neuron's instantaneous firing rate. While it is possible to use the number of spikes generated by a population of cells to code analog values, this turns out to be expensive, since to code n analog values, one needs n-1 neurones. Furthermore, the roughly Poisson nature of spike generation would also seriously limit the amount of information that can be transmitted. Even at 100 spikes.s· l , there is a roughly 35% chance that the neuron will generate no spike at all within a particular 10 ms window, again forcing the system to use large numbers of redundant cells. An alternative is to use information encoded in the temporal pattern of firing produced in response to transient stimuli (Mainen & Sejnowski, 1995). In particular, one can treat neurones not as analog to frequency converters (as is normally the case) but rather as analog to delay converters(Thorpe, 1990, 1994). The idea is very simple and uses the fact that the time taken for an integrate-and-fire neuron to reach threshold depends on input strength. Thus, in response to an intensity profile, the 6 neurones in figure 2 will tend to fire in a particular order - the most strongly activated cells firing first. Since each neuron fires one and only one spike, the firing rates of the cells contain no information, but there is information in the order in which the cells fire (see also Hopfield, 1995). A B c F I INTENSITY Figure 2 : An example of spike order coding. Because of the intrinsic properties of neurones the most strongly activated neurones will fire first. The sequence B>A>F>C>E>D is one ot the 720 (i.e. 6!) possible orders in which the 6 neurones can fire, each of which reflects a different intensity profile. Note that such a code can be used to send information very quickly. To test the plausibility of using spike order rather than firing rate as a code, we have developed a neural network simulator "SPIKENET" and used it to model the initial stages of visual processing. Initial results are very encouraging and demonstrate that sophisticated visual processing can indeed be achieved in a visual system in which only one spike per neuron is available. Rapid Vzsual Processing using Spike Asynchrony 903 2. SPIKENET SIMULATIONS SPIKENET has been developed in order to simulate the activity of large numbers of integrate-and-fire neurones. The basic neuronal elements are simple, and involve only a limited number of parameters, namely, an activation level, a threshold and a membrane time constant. The basic propagation mechanism involves processing the list of neurones that fired during the previous time step. For each spiking neuron, we add a synaptic weight value to each of its targets, and, if the target neuron's activation level exceeds its threshold, we add it to the list of spiking neurones for the next time step and reset its activation level by subtracting the threshold value. When a target neuron is affected for the first time on any particular time step, its activation level is recalculated to simulate an exponential decay over time. One of the great advantages of this kind of "event-driven" simulator is its computational efficiency - even very large networks of neurones can be simulated because no processor time is wasted on inactive neurones. 2.1 ARCHITECTURE As an initial test of the possibility of single spike processing, we simulated the propagation of activity in a visual system architecture with three levels (see Figure 3). Starting from a gray-scale image (180 x 214 pixels) we calculate the levels of activation in two retinal maps, one corresponding to ON-center retinal ganglion cells, the other to OFF-center cells. This essentiaIly corresponds to convolving the image with two Mexican-hat type operators. However, unlike more classic neural net models, these activation levels are not used to determine a continuous output value for each neuron, nor to calculate a firing rate. Instead, we treat the cells as analog-to-delay converters and calculate at which time step each retinal unit will fire. Because of their receptive field organization, cells which fire at the shortest latencies will correspond to regions in the image where the local contrast is high. Note, however, that each retinal ganglion cell will fire once and once only. While this is clearly not physiologically realistic (normally, cells firing at a short latencies go on to fire further spikes at short intervals) our aim was to see what sort of processing can be achieved in the absence of rate coding. The ON- and OFF-center cells each make excitatory connections to a large number of cortical maps in the second level of the network. Each map contains neurones with a different pattern of afferent connections which results in orientation and spatial frequency selectivity similar to that described for simple-type neurones in striate cortex. In these simulations we used 32 different filters corresponding to 8 different orientations (each separated by 45°) and four different scales or spatial frequencies. This is functionally equivalent to having one single cortical map (equivalent to area VI) in which each point in visual space corresponds to a hypercolumn containing a complete set of orientation and spatial frequency tuned filters. Units in the third layer receive weighted inputs from all the simple units corresponding to a particular region of space with the same orientation preference and thus roughly correspond to complex cells in area VI. 904 Layer 3 Orientation and Spatial Frequency tuned Complex cells (::::205 000 units) lJoyer1 \ , t I ON- and OFF-center cells £r 7 (::::77 000 units) (!!YI .". 'I !J Image 180 x 214 pixels S. J. Thorpe and J. Gautrais Figure 3 : Architecture used in the present simulations One unusual feature of the propagation process used in SPIKENET is that the postsynaptic effect of activating a synapse is not fixed, but depends on how many inputs have already been activated. Thus, the earliest firing cells produce a maximal post-synaptic effect (100%), but those which fire later produce less and less response. Specifically, the sensitivity of the post-synaptic neuron decreases by a fixed percentage each time one of its inputs fires. The phenomenon is somewhat similar to the sorts of activity-dependent synaptic depression described recently by Markram & Tsodyks (1996) and others, but differs in that the depression affects all the inputs to a particular neuron. The net result is to make the post-synaptic cell sensitive to the order in which its inputs are activated. 2.2 SIMULATION RESULTS When a new image is presented to the network, spikes are generated asynchronously in the ON- and OFF-center cells of the retina in such a way that information about regions of the image with high local contrast (i.e. where there are contours present) are sent to the cortex first. Progressively, neurons in the second layer become more and more excited, and, after a variable number of time steps, the first cells in the second layer will start to Rapid Visual Processing using Spike Asynchrony 905 reach threshold and fire. Note that, as in the first layer, the earliest firing units will be those for whom the pattern of input activation best matches their receptive field structure. Layerl "ON-center" Cells Layer 2 Simple cells Orientation 45° Layer 3 Complex cells Orientation 45° 40rns 45rns 50rns 80rns Figure 4 : Development of activity in 3 of the maps Figure 4 illustrates this process for just three maps. The top row shows the location of units in the ON-center retinal map that have fired after various delays. After 40 msec, the main outlines of the figure can be seen but progressively more details are seen after 45 and then 50 ms. Note that the representation used here uses pixel intensity to code the order in which the cells have fired - bright white spots correspond to places in the image where the cells fired earliest. In the final frame (taken at 80 ms) the vast majority of ONcenter cells have already fired and the resulting image is quite similar to a high spatial frequency filtered version of the original image. The middle row of images shows activity in one of the second level maps - the one corresponding to medium spatial frequency components oriented at 45°. Note that in the first times lice (40 ms) very few cells have fired, but that the proportion increases progressively over the next 10 or so milliseconds. However, even at the end of the propagation process, the proportion of cells that have actually fired remains low. Finally, the lowest row shows activity in the corresponding third layer map - again corresponding to contours oriented at 45°, but this time with less precise position specificity as a result of the grouping process. Figure 5 plots the total number of spikes occurring per millisecond in each of the three layers during the first 100 ms following the onset of processing. It is perhaps not 906 s. 1. Thorpe and 1. Gautrais surprising that the onset of firing occurs later for layers 2 and 3. However, there is a huge amount of overlap in the onset latencies of cells in the three layers, and indeed, it is doubtful whether there would be any systematic differences in mean onset latency between the three layers. 1250 en 1000 E .... Q) c.. ~ 750 ~ .0. en 500 250 o o 10 20 30 40 50 60 Time (ms) 70 On centre cells Simple cells Complex cells 80 90 100 Figure 5 : Amount of activity measured in spikes/ms for the three layers of neurones as a function of time But perhaps one of the most striking features of these simulations is the way in which the onset latency of cells can be seen to vary with the stimulus. The small number of cells in each layer which fire early are in fact very special because only the most optimally activated cells will fire at such short latencies. The implications of this effect for visual processing are far reaching because it means that the earliest information arriving at later processing stages will be particularly informative because the cells involved are very unambiguous. Interestingly, such changes in onset latency have been observed experimentally in neurones in area VI of the awake primate in response to changes in orientation. In these experiments it was shown that shifting the orientation of a grating away from a neuron's preferred orientation could result in changes in not only the firing rate of the cell, but also increases in onset latency of as much as 20-30 ms (Celebrini, Thorpe, Trotter & Imbert, 1993). 3. CONCLUSIONS A number of points can be made on the basis of these results. Perhaps the most important is that visual processing can indeed be performed under conditions in which spike frequency coding is effectively ruled out. Clearly, under normal conditions, neurones in the visual system that respond to a visual input will almost invariably generate more Rapid VISual Processing using Spike Asynchrony 907 than one spike. However, as we have argued previously, processing in the visual system is so fast that most cells will not have time to generate more than one spike before processing in later stages has to be initiated. The present results indicate that the use of temporal order coding may provide a key to understanding this remarkable efficiency. The simulations presented here are clearly very limited, but we are currently looking at spike propagation in more complex architectures that include extensive horizontal connections between neurones in a particular layer as well as additional layers of processing. As an example, we have recently developed an application capable of digit recognition. SPIKENET is well suited for such large scale simulations because of the event-driven nature of the propagation process. For instance, the propagation presented here, which involved roughly 700 000 neurones and over 35 million connections, took roughly 15 seconds on a 150 MHz PowerMac, and even faster simulations are possible using parallel processing. With this is view we have developed a version of SPIKENET that uses PVM (Parallel Virtual Machine) to run on a cluster of workstations. References Celebrini S., Thorpe S., Trotter Y. & Imbert M. (1993). Dynamics of orientation coding in area VI of the awake primate Visual Neuroscience 10, 811-25. Hopfie1d J. J. (1995). Pattern recognition computation using action potential timing for stimulus representation. Nature, 376, 33-36. Mainen Z. F. & Sejnowski T. J. (1995). Reliability of spike timing in neocortical neurons Science, 268, 1503-6. Markram H. & Tsodyks M. (1996) Redistribution of synaptic efficacy between neocortical pyramidal neurons. Nature, 382,807-810 Nowak L.G. & Bullier J (1997) The timing of information transfer in the visual system. In Kaas J., Rocklund K. & Peters A. (eds). Extrastriate Cortex in Primates (in press). Plenum Press. Oram M. W. & Perrett D. I. (1992). Time course of neural responses discriminating different views of the face and head Journal of Neurophysiology, 68, 70-84. Rolls E. T. & Tovee M. J. (1994). Processing speed in the cerebral cortex and the neurophysiology of visual masking Proc R Soc Lond B Bioi Sci, 257,9-15. Thorpe S., Fize D. & Marlot C. (1996). Speed of processing in the human visual system Nature, 381, 520-522. Thorpe S. J. (1990). Spike arrival times: A highly efficient coding scheme for neural networks. In R. Eckmiller, G. Hartman & G. Hauske (Eds.), Parallel processing in neural systems (pp. 91-94). North-Holland: Elsevier. Reprinted in H. Gutfreund & G. Toulouse (1994), Biology and computation: A physicist's choice. Singapour: World Scientific. Thorpe S. J. & Imbert M. (1989). Biological constraints on connectionist models. In R. Pfeifer, Z. Schreter, F. Fogelman-Soulie & L. Steels (Eds.), Connectionism in Perspective. (pp. 63-92). Amsterdam: Elsevier.
1996
150
1,207
Visual Cortex Circuitry and Orientation Tuning Trevor M undel Department of Neurology University of Chicago Chicago, IL 60637 mundel@math.uchicago.edu Alexander Dimitrov Department of Mathematics University of Chicago Chicago, IL 60637 a-dimitrov@ucllicago.edu Jack D. Cowan Departments of Mathematics and Neurology University of Chicago Chicago, IL 60637 cowan@math.uchicago.edu Abstract A simple mathematical model for the large-scale circuitry of primary visual cortex is introduced. It is shown that a basic cortical architecture of recurrent local excitation and lateral inhibition can account quantitatively for such properties as orientation tuning. The model can also account for such local effects as cross-orientation suppression. It is also shown that nonlocal state-dependent coupling between similar orientation patches, when added to the model, can satisfactorily reproduce such effects as non-local iso--orientation suppression, and non-local crossorientation enhancement. Following this an account is given of perceptual phenomena involving object segmentation, such as "popout", and the direct and indirect tilt illusions. 1 INTRODUCTION The edge detection mechanism in the primate visual cortex (VI) involves at least two fairly well characterized circuits. There is a local circuit operating at subhypercolumn dimensions comprising strong orientation specific recurrent excitation and weakly orientation specific inhibition. The other circuit operates between hyper columns , connecting cells with similar orientation preferences separated by several millimetres of cortical tissue. The horizontal connections which mediate this 888 T. Mundel. A. Dimitrov and J. D. Cowan circuit have been extensively studied. These connections are ideally structured to provide local cortical processes with information about the global nature of stimuli. Thus they have been invoked to explain a wide variety of context dependent visual processing. A good example of this is the tilt illusion (TI), where surround stimulation causes a misperception of the angle of tilt of a grating. The interaction between such local and long-range circuits has also been investigated. Typically these experiments involve the separate stimulation of a cells receptive field (the classical receptive field or "center") and the immediate region outside the receptive field (the non-classical receptive field or "surround"). In the first part of this work we present a simple model of cortical center-surround interaction. Despite the simplicity of the model we are able to quantitatively reproduce many experimental findings. We then apply the model to the TI. We are able to reproduce the principle features of both the direct and indirect TI with the model. 2 PRINCIPLES OF CORTICAL OPERATION Recent work with voltage-sensitive dyes (Blasdel, 1992) augments the early work of Rubel & Wiesel (1962) which indicated that clusters of cortical neurons corresponding to cortical columns have similar orientation preferences. An examination of local field potentials (Victor et al., 1994) which represent potentials averaged over cortical volumes containing many hundreds of cells show orientation preferences. These considerations suggest that the appropriate units for an analysis of orientation selectivity are the localized clusters of neurons preferring the same orientation. This choice of a population model immediately simplifies both analysis and computation with the model. For brevity we will refer to elements or edge detectors, however these are to be understood as referring to localized populations of neurons with a common orientation preference. We view the cortex as a lattice of hypercolumns, in which each hypercolumn comprises a continuum of iso-orientation patches distinguished by their preferred orientation ¢. All space coordinates refer to distances between hypercolumn centers. The popUlation model we adopt throughout this work is a simplified form of the Wilson-Cowan equations. 2.1 LOCAL MODEL Our local model is a ring (¢ = -900to + 90°) of coupled iso-orientation patches and inhibitors with the following characteristics • Weakly tuned orientation biased inputs to VI. These may arise either from slight orientation biases of lateral geniculate nucleus (LGN) neurons or from converging thalamocortical afferents • Sharply tuned (space constant ±7.5°) recurrent excitation between isoorientation populations • Broadly tuned inhibition to all iso-orientation populations with a cut-off of inhibition interactions at between 45° and 60° separation The principle constraint is that of a critical balance between excitatory and inhibitory currents. Recent theoretical studies (Tsodyks & Sejnowski 1995; Vreeswijk & Sompolinsky 1996) have focused on this condition as an explanation for certain features of the dynamics of natural neuronal assemblies. These features include the irregular temporal firing patterns of cortical neurons, the sensitivity of neuronal assemblies in vivo to small fluctuations in total synaptic input and the distribution of firing rates in cortical networks which is markedly skewed towards low mean VISual Cortex Circuitry and Orientation Tuning 889 rates. Vreeswijk & Sompolinsky demonstrate that such a balance emerges naturally in certain large networks of excitatory and inhibitory populations. We implement this critical balance by explicitly tuning the strength of connection weights between excitatory and inhibitory populations so that the system state is subcritical to a bifurcation point with respect to the relative strength of excitation/inhibition. 2.2 HORIZONTAL CONNECTIONS We distinguish three potential patterns of horizontal connectivity • connections between edge detectors along an axis parallel to the detectors preferred orientation (visuotopic connection) • connections along an axis orthogonal to the detectors preferred orientation, with or without visuotopic connections • radially symmetric connection to all detectors of the same orientation in surrounding hypercolumns Recent experimental work in the tree shrew (Fitzpatrick et al., 1996) and preliminary work in the macaque (Blasdel, personal communication) indicate that visuotopic connection is the predominant pattern of long-range connectivity. This connectivity pattern allows for the following reduction in dimension of the problem for certain experimental conditions. Consider the following experiment. A particular hypercolumn designated the "center" is stimulated with a grating at orientation </> resulting in a response from the hdge detector. The region outside the receptive area of this hypercolumn (in the "surround") is also stimulated with a grating at some uniform orientation </>' resulting in responses from <,i>'-edge detectors at each hypercolumn in the surround. In order to study the interactions between center and surround, then to first order approximation only the center hypercolumn and interaction with the surround along the </> visuotopic axis (defined by the center) and the </>' visuotopic axis (once again defined by the center) need be considered. In fact, except when </> = </>' the effect of the center on the surround will be negligible in view of the modulatory nature of the horizontal connections detailed above. Thus we can reduce the problem (a priori three dimensional - one angle and two space dimensions) to two dimensions (one angle and one space dimension) with respect to a fixed center. This reduction is the key to providing a simple analysis of complex neurophysiological and psychophysical data. 3 RESULTS 3.1 CENTER-SURROUND INTERACTIONS Although, we have modeled the state-dependence of the horizontal connections, many of the center-surround experiments we wish to model have not taken this dependence explicitly into account. In general the surround has been found to be suppressive on the center, which accords with the fact that the center is usually stimulated with high contrast stimuli. A typical example of the surround suppressive effect is shown in figure 1. The basic finding is that stimulation in the surround of a visual cortical cell's receptive field generally results in a suppression of the cell's tuning response that is maximal for surround stimulation at the orientation of the cell's peak tuning 890 T. Mundel, A. Dimitrov and 1. D. Cowan 60 ® 30 ~ 40 A -60 -30 0 +30 +60 +90 ORJENTATIONOFBAR (dell B 10 -60 -30 0 +30 +60 +90 ORJENfATION OF GRATINO (dell Figure 1: Non-local effects on orientation tuning - experimental data. Response to constant center stimulation at 15° and surround stimulation at angles [-90° , 900J (open circles), Local tuning curve (filled circles). Redrawn from Blakemore and Tobin (1972) response and falls off with stimulation at other orientations in a characteristic manner. Further examples of surround suppression can be found in the paper of Sillito et al. (1995). Figure 2 depicts simulations in which long-range connections to local inhibitory populations are strong compared to connections to local excitatory populations. These experiments and simulations appear to conflict with the consistent experimental finding that stimulating a hypercolumn with an orthogonal stimulus suppresses the response to the original stimulus. The relevant results can be summarised as follows: cross-orientation suppression (with orthogonal gratings) originates within the receptive field of most cells examined and is a consistent finding in both complex and simple cells. The degree of suppression depends linearly on the size of the orthogonal grating up to a critical dimension which is smaller than the classical receptive field dimension. It is possible to suppress a response to the baseline firing rate by either increasing the size or the contrast of the orthogonal grating. The model outlined earlier can account for all these observations, and similar measurements recently described by Sillitto et. al. (1995), in a strikingly simple fashion in the setting of single mode bifurcations. Orthogonal inputs are of the form a/2[1 + cos2s(cP - cPo)J + b/2[1 + cos2s(cP - cPo + 90°)]' where a and b are amplitudes with a > band cP E [-900,900J. By simple trigonometry this simplifies to (a + b) /2 + (a - b) /2 cos 2s( cP - cPo) Thus the input of amplitude b reduces the amplitude of the orthogonal input and hence gives rise to a smaller response. This is then the mechanism by which local double orthogonal stimulation leads to suppression. The center-surround case is different in that the orthogonal input originates from the horizontal connections and (in the suppressive setting) is input primarily to the orthogonal inhibitory population. It can be shown rigorously that for small amplitude stimuli this is equivalent to an orthogonal input to the excitatory population with opposite sign. Thus we have a total input (a + b)/2[1 + cos(2s(cP - cPo)J where b arises from the horizontal input and hence increases the amplitude of the fundamental component of the input. VISual Cortex Circuitry and Orientation Tuning -90 -«> -30 ° 30 60 90 , B -90 -«> -30 ° 30 60 90 .~1.5 ~ ~ 1 : ::10.5 , -90 -«> -30 ° 30 60 90 ~ 891 Figure 2: Non-local effects on orientation tuning. (<r<ro) = Center response to preferred orientation at different surround orientations, (x-x-x) = Center orientation tuning without surround stimulation, (-) = Center response to surround stimulation alone. A - response of population with 20° orientation preference. B, C and D - response of populations with 5° orientation preference. More realistically, multiple modes may bifurcate simultaneously, though in the small amplitude linear regime where orientation preference is determined these can all be treated separately in the manner detailed above and lead to the same result. 3.2 POPOUT AND GAIN CONTROL "Stimulus" dependent horizontal connections have recently been used to model a possible physiological correlate for the psychophysical phenomena of " pop-out " and enhancement (Stemmler, Usher & Niebur, 1995). In this model the horizontal input to excitatory neurons is via sodium channels which are dependent in the conventional manner on the differential between the membrane voltage and the sodium equilibrium potential. For such sodium channels the synaptic currents are attenuated as the membrane depolarizes towards the sodium equilibrium potential. This effect is opposite to that observed for the sodium channels mediating the horizontal input (Hirsch & Gilbert, 1991) which show increased synaptic currents as the membrane is depolarized. Thus although this model reproduces the phenomena described above it does not do so on the basis of the known physiology. We have confirmed with our model that weak stimulus enhancement and strong stimulus pop-out can be modelled with a variety of formulations for the excitatory neuron horizontal input, including a formulation which attenuates with increasing activity as in the model of Stemmler et. al. It is interesting to note that one overall effect of the horizontal connections is to act as a type of gain control uniformizing the local response over a range of stimulus strengths. This gain control function has been discussed by Somers, Nelson & Sur (1994). 3.3 THE TILT ILLUSION The tilt illusion (TI) is one of the basic orientation-based visual illusion. A TI occurs when viewing a test line against an inducing grating of uniformly oriented lines with an angle of (J between the orientation of the test line and the inducing 892 T. Mundel, A. Dimitrov and J. D. Cowan grating. Two components of the TI have been described (Wenderoth & Johnstone, 1988), the direct TI where the test line appears to be repelled by the gratingthe orientation differential appears increased, and the indirect TI where the test line appears attracted to the orientation of the grating-the orientation differential appears decreased. Figure 3 depicts a typical plot of magnitude of the tilt effect versus the angle differential between inducing grating and test line reproduced from Wenderoth & Beh (1977). The TI thus provides compelling evidence that local I.: I 0 -- -- -.---- -- - - •• - - • •• -.- - --- - -.. , Figure 3: Direct (positive) and indirect (negative) tilt effects detection of edges is dependent on information from more distant points in the visual field. It is generally believed, that the direct TI is due to lateral inhibition between cortical neurones (Wenderoth & Johnstone, 1988: Carpenter & Blakemore, 1973). It has been postulated that the indirect TI occurs at a higher level of visual processing. We show here that both the direct and indirect TI are a consequence of the lateral and local connections in our model. = :E A en 0 -0.5'-------:-___ -10 30 50 70 90 -P c -0.5'---10--30--5-0--70--90 ~ / / ., ~ :a. E « -90 ~ -30 / " I \ I .,0 t ~ -90 ~ -30 B \ \ J \ J \ I \ I \I 0 30 60 90 ~ 0 / " I , I \ \ \ \ J " 0 30 60 90 ~ Figure 4: Model simulations of the tilt effect (A and C). Band D show the corresponding kernels mediating long-range interactions. Solid lines indicate the absolute kernel and dashed lines indicate the effective kernel In figure 4 we give examples of the TI obtained from the model system. The effective kernels for long-range interactions are obtained by filtering the absolute kernels with the local filter which has a band-pass characteristic. It is this effective kernel which determines the tilt effect in keeping with our simulations and analysis which show that orientation preference is determined at the small amplitude linear stage of system development. Visual Cortex Circuitry and Orientation Tuning 893 4 SUMMARY AND DISCUSSION We have shown that a very simple center-surround organization, operating in the orientation domain can successfully account for a wide range of neurophysiological and psychophysical phenomena, all involving the effects of visual context on the responses of assemblies of spiking neurons. We expect to be able to show that such an organization can be seen in many parts of the cortex, and that it plays an important role in many forms of information processing in the brain. Acknowledgements Supported in part by Grant # 96-24 from the James S. McDonnell Foundation. References C. Blakemore and E.A. Tobin, Lateral Inhibition Between Orientation Detectors in the Cat's Visual Cortex, Exp. Brain Res., 15, 439-440, (1972) G.G. Blasdel, Orientation selectivity, preference, and continuity in monkey striate cortex, J. Neurosci. 12 No 8, 3139-3161 (1992) R.H.S. Carpenter and C. Blakemore, Interactions between orientations in human vision, Expl. Brain. Res., 18, 287- 303, (1973) G.C. DeAngelis, J.G. Robson, I. Ohzawa and R.D. Freeman, Organization of Suppression in Receptive Fields of Neurons in Cat Visual Cortex, J. Neurophysiol. , 68 No 1, 144-163, (1992) D. Fitzpatrick, The Functional Organization of Local Circuits in Visual Cortex: Insights from the Study of Tree Shrew Striate Cortex, Cerebral Cortex 6, 329-341, (1996) J.D. Hirsch and C.D. Gilbert, Synaptic physiology of horizontal connections in the cat's visual cortex, J. Neurosci., 11, 1800-1809, (1991) A.M. Sillito, K.L. Grieve, H.E. Jones, J. Cudeiro & J. Davis, Visual cortical mechanisms detecting focal orientation discontinuities, Nature, 378, 492-496, (1995) D.C. Somers, S. Nelson and M. Sur, Effects of long-range connections on gain control in an emergent model of visual cortical orientation selectivity, Soc. Neurosci. ,20, 646.7, (1994) M. Stemmler, M. Usher and E. Niebur, Lateral Interactions in Primary Visual Cortex: A Model Bridging Physiology and Psychophysics, Science, 269, 1877-1880, (1995) M. V. Tsodyks and T. Sejnowski, Rapid state switching in balanced cortical network models, Network, 6 No 2, 111-124, (1995) J.D. Victor, K. Purpura, E. Katz and B. Mao, Population encoding of spatial frequency, orientation and color in macaque VI, J. Neurophysiol., 72 No 5, (1994) C. Vreeswijk and H. Sompolinsky, Chaos in neuronal networks with balanced excitatory and inhibitory activity. Science 274, 1724-1726, (1996) P. Wenderoth and H. Beh, Component analysis of orientation illusions, Perception, 6 57-75, (1977) P. Wenderoth and S. Johnstone, The different mechanisms of the direct and indirect tilt illusions, Vision Res., 28 No 2, 301-312, (1988)
1996
151
1,208
Orientation contrast sensitivity from long-range interactions in visual cortex Klaus R. Pawelzik, Udo Ernst, Fred Wolf, Theo Geisel Institut fur Theoretische Physik and SFB 185 Nichtlineare Dynamik, Universitat Frankfurt, D-60054 Frankfurt/M., and MPI fur Stromungsforschung, D-37018 Gottingen, Germany email: {klaus.udo.fred.geisel}@chaos.uni-frankfurt.de Abstract Recently Sill ito and coworkers (Nature 378, pp. 492,1995) demonstrated that stimulation beyond the classical receptive field (cRF) can not only modulate, but radically change a neuron's response to oriented stimuli. They revealed that patch-suppressed cells when stimulated with contrasting orientations inside and outside their cRF can strongly respond to stimuli oriented orthogonal to their nominal preferred orientation. Here we analyze the emergence of such complex response patterns in a simple model of primary visual cortex. We show that the observed sensitivity for orientation contrast can be explained by a delicate interplay between local isotropic interactions and patchy long-range connectivity between distant iso-orientation domains. In particular we demonstrate that the observed properties might arise without specific connections between sites with cross-oriented cRFs. 1 Introduction Long range horizontal connections form a ubiquitous structural element of intracortical circuitry. In the primary visual cortex long range horizontal connections extend over distances spanning several hypercolumns and preferentially connect cells of similar orientation preference [1, 2, 3, 4]. Recent evidence suggests that Modeling Orientation Contrast Sensitivity 91 their physiological effect depends on the level of postsynaptic depolarization; acting exitatory on weakly activated and inhibitory on strongly activated cells [5, 6]. This differ ental influence possibly underlies perceptual phenomena as 'pop out' and 'fill in' [9]. Previous modeling studies demonstrated that such differential interactions may arise from a single set of long range excitatory connections terminating both on excitatory and inhibitory neurons in a given target column [7,8]. By and large these results suggest that long range horizontal connections between columns of like stimulus preference provide a central mechanism for the context dependent regulation of activation in cortical networks. Recent experiments by Sillito et al. suggest, however, that lateral connections in primary visual cortex can also induce more radical changes in receptive field organization [10]. Most importantly this study shows that patch- suppressed cells can respond selectively to orientation contrast between center and surround of a stimulus even if they are centrally stimulated orthogonal to their preferred orientation. Sillito et al. argued, that these response properties require specific connections between orthogonally tuned columns for which, however, presently there is only weak evidence. Here we demonstrate that such nonclassical receptive field properties might instead arise as an emergent property of the known intracortical circuitry. We investigate a simple model for intracortical activity dynamics driven by weakly orientation tuned afferent excitation. The cortical actitvity dynamics is based on a continous firing rate description and incorporates both a local center- surrond type interaction and long range connections between distant columns of like orientation preference. The connections of distant orientation columns are assumed to act either excitatory or inhibitory depending on the activation of their target neurons. It turns out that this set of interactions not only leads to the emergence of patch-suppressed cells, but also that a large fraction of these cells exhibits a selectivity for orientation contrast very similar to the one observed by Sillito et al. . 2 Model Our target is the analysis of basic rate modulations emerging from local and long range feedback interactions in a simple model of visual cortex. It is therefore appropriate to consider a simple rate dynamics x = -c' x + F(x), where x = {Xi, i = L.N} are the activation levels of N neurons. F(x) = g(Imex(x) + Ilat(x) + Iext ), where g(I) = Co' (I - Ithres) if I> Ithres, and g(I) = 0 otherwise, denotes the firing rate or gain function in dependence of the input I . The neurons are arranged in a quadratic array representing a small part of the visual cortex. Within this layer, neuron i has a position r i and a preferred orientation <Pi E [0,180]. The input to neuron i has three contributions Ii = Iimex + Ifat + Ifxt. Iinex = tmex . 2:f=l wiJex Xj is due to a mexican-hat shaped coupling structure with weights wiJex, Ilat = tlat . WL(Xi) . 2:f=l wi~rxj denotes input from long-range orientation-specific interactions with weights w~jt, and the third term models the orientation dependent external input Iext = text . 16~t . (1 + TIi), where "li denotes the noise added to the external input. wL(x) regulates the strength and sign ofthe long92 K. R. Pawelzik, U. Ernst. F. Wolf and T. Geisel a) Figure 1: Structure and response properties of the model network. a) Coupling structure from one neuron on a grid of N = 1600 elements projected on the orientation preference map which was used for stimulation (<<I> i ,i = 1... N). Inhibitory and excitatory couplings are marked with black and white squares, respectively, the sizes of which represent the coupling strength. b) Activation pattern of the network driven by a central stimulus of radius rc = 11 and horizontal orientation. c) Self-consistent orientation map calculated from the activation patterns for all stimulus orientations. Note that this map matches the input orientation preference map shown in a) and b). range lateral interaction in dependence of the postsynaptic activation, summarizing the behavior of a local circuit in which the inhibitory population has a larger gain. In particular, Wl(X) can be derived analytically from a simple cortical microcircuit (Fig.2). This circuit consists of an inhibitory and excitatory cell population connected reciprocally. Each population receives lateral input and is driven by the external stimulus I. The effective interaction WL depends on the lateral input L and external input I. Assuming a piecewise linear gain function for each population, similar as those for the Xi'S, the phase-space I-L is partitioned in some regions. Only if both I and L are small, WL is positive; justifying the choice WL = Xsh - tanh (0.55· (x - Xa)/Xb) which we used for our simulations. The weights wijex are given by (1) and wijex = 0 otherwise. In this representation of the local interactions weights and scales are independently controllable. In particular if we define 2 2· ere! 2 ( 2 a ex = --4' ain = ( )( )3' bex = aexrex , bin = ain rin rex) 1rrex 1r rin + rex rin rex (2) rex and rin denote the range of the excitatory and inhibitory part of the mexican hat, respectively. Here we used rex = 2.5 and rin = 4.0. erel controls the balances of inhibition and excitation. With constant activation level Xi = Xo Vi the inhibition is erel times higher than the excitation and Imex = €mex . (1 - erel) . Xo· Modeling Orientation Contrast Sensitivity 93 Figure 2: Local cortical circuit (left), consisting of an inhibitory and an excitatory population of neurons interconnected with weights Wie, Wei , and stimulated with lateral input L and external input I . By substituting this circuit with one single excitatory unit, we need a differential instead of a fixed lateral coupling strength (WL' right), which is positive only for small I and L. The weights wtit are and 0 otherwise. alat ,1/J and alat ,r provide the orientation selectivity and the range of the long-range lateral interaction, respectively. The additional parameter Clat 1· lat h th t '\;""'N lat 1 norma lzes wij suc a ~j=l wij ~ . The spatial width and the orientation selectivity of the input fields are modeled by a convolution with a Gaussian kernel before projected onto the cortical layer I ext . 1 ~ [ (I ri -rj 12) (I <Pi <Pj 12)] O,i 2 2 ~ exp 2 2' exp 2 2 . 7rar ecp,r j=1 arecp ,r are cp , ~ In our simulations, the orientation preference of a cortical neuron i was given by the orientation preference map displayed in Figla. 3 Results We analyzed stationary states of our model depending on stimulus conditions. The external input, a center stimulus of radius r c = 6 at orientation <Pc and an annulus IThe following parameters have been used in our simulations leading to the results shown in Figs. 1-4. "1i = 0.1 (external input noise) , f m ex = 2.2, fLat = 1.5, f ext = 1.3, Ash = 0.0, Aa = 0.2, Ab = 0.05, t e = 0.6, Se = 0.5, Cre! = 2.0 (balance between inhibition d . t' ) al' Lat h th t '\;""'N Lat 1 20 15 an eXClta lOn , C/at norm lzes w i] suc a ~j=l Wij ~ , 00Lnt ,q, = , O'/at ,r = O'r ecp ,r = 5, O'rec p,4> = 40, r ex = 2.5, and rin = 5.0. 94 K. R. Pawelzik, U. Ernst, F. Wolf and T. Geisel 0) b) c) Figure 3: Changes in patterns of activity induced by the additional presentation of a surround stimulus. Grey levels encode increase (darker grey) or decrease (lighter grey) in activation x. a) center and surround parallel, b) and c) center and surround orthogonal. While in b), the center is stimulated with the preferred orientation, in c), the center is stimulated with the non-preferred orientation. of inner radius rc and outer radius rs = 18 at orientation «I>s , was projected onto a grid of N = 40 x 40 elements (Fig.1a). Simulations were performed for 20 orientations equally spaced in the interval [0, 1800 J. When an oriented stimulus was presented to the center we found blob-like activity patterns centered on the corresponding iso-orientation domains (Fig.1b). Simulations utilizing the full set of orientations recovered the input-specificity and demonstrated the self-consistency of the parameter set chosen (Fig.1c). While in this case there were still some deviations, stimulation of the whole field yielded perfect self-consistency (not shown) in the sense of virtually identical afferent and response based orientation preferences. For combinations of center and surround inputs we observed patch- suppressed regions. These regions exhibited substantial responses for cross- oriented stimuli which often exceeded the response to an optimal center stimulus alone. Figs.3 and 4 summarize these results. Fig.3 shows responses to center-surround combinations compared to activity patterns resulting from center stimulation only. Obviously certain regions within the model were strongly patch-suppressed (Fig.3, light patches for same orientations of center and surround). Interestingly a large fraction of these locations exhibited enhanced activation when center and surround stimulus were orthogonal. Fig.4 displays tuning curves of patch-suppressed cells for variing the orientation of the surround stimulus. Clearly these cells exhibited an enhancement of most responses and a substantial selectivity for orientation contrast. Parameter variation indicated that qualitatively these results do not depend sensitively of the set of parameters chosen. 4 Summary and Discussion Our model implements only elementary assumptions about intracortical interactions. A local sombrero shaped feedback is well known to induce localized blobs of activity with a stereotyped shape [12J. This effect lies at the basis of many models of visual cortex, as e.g. for the explanation of contrast independence of orientation Modeling Orientation Contrast Sensitivity 0.8 x 0.6 \ c: 0 \ :OJ 0 \ > ':oJ u 0.4 , 0 c 0 0.2 CI) ::E 0.0 -90 o Orientation r/J , , \ +90 95 Figure 4: Tuning curves for patch-suppressed cells preferring a horizontal stimulus within their cRF. The bold line shows the orientation tuning curve of the response to an isolated center stimulus. The dashed and dotted lines show the tuning curve when stimulating with a horizontal (dashed) and a vertical (dotted) center stimulus while rotating the surround stimulus. The curves have been averaged over 6 units. tuning [13, 14]. Long range connections selectively connect columns of similar orientation preference which is consistent with current anatomical knowledge [3, 4]. The differential effect of this set of connections onto the target population was modeled by a continuous sign change of their effective action depending on the level of postsynaptic input or activation. Orientation maps were used to determine the input specificity and we assumed a rather weak selectivity of the afferent connections and a restricted contrast which implies that every stimulus provides some input also to orthogonally tuned cells. This means that long-range excitatory connections, while not effective when only the surround is stimulated, can very well be sufficient for driving cells if the stimulus to the center is orthogonal to their preferred orientation (Contrast sensitivity). In our model we find a large fraction of cells that exhibit sensitivity for centersurround stimuli. It turns out that most of the patch- suppressed cells respond to orientation contrasts, i.e. they are strongly selective for orientation discontinuities between center and surround. We also find contrast enhancement, i.e. larger responses to the preferred orientation in the center when stimulated with an orthogonal surround than if stimulated only centrally (Fig.4). The latter constitutes a genuinely emergent property, since no selective cross- oriented connections are present. This phenomenon can be understood as a desinhibitory effect. Since no cells having long-range connections to the center unit are activated, the additional sub-threshold input from outside the classical receptive field can evoke a larger response (Contrast enhancement). Contrarily, if center and surround are stimulated with the same ori96 K. R. Pawelzik, U. Ernst, F. Wolf and T. Geisel entation, all the cells with similar orientation preference become activated such that the long-range connections can strongly inhibit the center unit (Patch suppression). In other words, while the lack of inhibitory influences from the surround should recover the response with an amplitude similar or higher to the local stimulation, the orthogonal surround effectively leads to a desinhibition for some of the cells. Our results show a surprising agreement with previous findings on non-classical receptive field properties which culminated in the paper by Sillito et al. [10]. Our simple model clearly demonstrates that the known intracortical interactions might lead to surprising effects on receptive fields. While this contribution concentrated on analyzing the origin of selectivities for orientation discontinuities we expect that the pursued level of abstraction has a large potential for analyzing a wide range of non-classical receptive fields. Despite its simplicity we believe that our model captures the main features of rate interactions. More detailed models based on spiking neurons, however, will exhibit additional dynamical effects like correlations and synchrony which will be at the focus of our future research. Acknowledgement: We acknowledge inspiring discussions with S. Lowel and J. Cowan. This work was supported by the Deutsche Forschungsgemeinschaft. References [1] D. Ts'o, C.D. Gilbert, and T.N. Wiesel, J. Neurosci 6, 1160-1170 (1986). [2] C.D. Gilbert and T.N. Wiesel, J. Neurosci. 9, 2432-2442 (1989). [3] S. Lowel and W. Singer, Science 255, 209 (1992). [4] R. Malach, Y. Amir, M. Harel, and A. Grinvald, PNAS 90, 1046910473 (1993). [5] J.A. Hirsch and C.D. Gilbert, J. Neurosci. 6, 1800-1809 (1991). [6] M. Weliky, K. Kandler, D. Fitzpatrick, and L.C. Katz, Neuron 15, 541-552 (1995). [7] M. Stemmler, M. Usher, and E. Niebur, Science 269, 1877-1880 (1995). [8] L.J. Toth, D.C. Sommers, S.C. Rao, E.V. Todorov, D.-S. Kim, S.B. Nelson, A.G. Siapas, and M. Sur, preprint 1995. [9] U. Polat, D. Sagi, Vision Res. 7,993-999 (1993). [10] A.M. Sillito, K.L. Grieve, H.E. Jones, J. Cudeiro, and J. Davis, Nature 378, 492-496 (1995). [11] J.J. Knierim and D.C. van Essen, J. Neurophys. 67,961-980 (1992). [12] H.R. Wilson and J. Cowan, BioI. Cyb. 13,55-80 (1973). [13] R. Ben-Yishai, R.L. Bar-Or, and H. Sompolinsky, Proc. Nat. Acad. Sci. 92,3844-3848 (1995). [14] D. Sommers, S.B. Nelson, and M. Sur, J. Neurosci. 15, 5448-5465 (1995).
1996
152
1,209
Multi-Task Learning for Stock Selection Joumana Ghosn Dept. Informatique et Recherche Operationnelle Universite de Montreal Montreal, Qc H3C-3J7 ghosn~iro.umontreal.ca Yoshua Bengio * Dept. Informatique et Recherche Operationnelle Universite de Montreal Montreal, Qc H3C-3J7 bengioy~iro.umontreal . ca Abstract Artificial Neural Networks can be used to predict future returns of stocks in order to take financial decisions. Should one build a separate network for each stock or share the same network for all the stocks? In this paper we also explore other alternatives, in which some layers are shared and others are not shared. When the prediction of future returns for different stocks are viewed as different tasks, sharing some parameters across stocks is a form of multi-task learning. In a series of experiments with Canadian stocks, we obtain yearly returns that are more than 14% above various benchmarks. 1 Introd uction Previous applications of ANNs to financial time-series suggest that several of these prediction and decision-taking tasks present sufficient non-linearities to justify the use of ANNs (Refenes, 1994; Moody, Levin and Rehfuss, 1993). These models can incorporate various types of explanatory variables: so-called technical variables (depending on the past price sequence), micro-economic stock-specific variables (such as measures of company profitability), and macro-economic variables (which give information about the business cycle). One question addressed in this paper is whether the way to treat these different variables should be different for different stocks, i.e., should one use the same network for all the stocks or a different network for each stock? To explore this question "also, AT&T Labs, Holmdel, NJ 07733 Multi-Task Learning/or Stock Selection 947 we performed a series of experiments in which different subsets of parameters are shared across the different stock models. When the prediction of future returns for different stocks are viewed as different tasks (which may nonetheless have something in common), sharing some parameters across stocks is a form of multi-task learning. These experiments were performed on 9 years of data concerning 35 large capitalization companies of the Toronto Stock Exchange (TSE). Following the results of previous experiments (Bengio, 1996), the networks were not trained to predict the future return of stocks, but instead to directly optimize a financial criterion. This has been found to yield returns that are significantly superior to training the ANNs to minimize the mean squared prediction error. In section 2, we review previous work on multi-task. In section 3, we describe the financial task that we have considered, and the experimental setup. In section 4, we present the results of these experiments. In section 5, we propose an extension of this work in which the models are re-parameterized so as to automatically learn what must be shared and what need not be shared. 2 Parameter Sharing and Multi-Task Learning Most research on ANNs has been concerned with tabula rasa learning. The learner is given a set of examples (Xl, yr), (X2' Y2), ... , (XN , YN) chosen according to some unknown probability distribution. Each pair (x, y) represents an input x, and a desired value y. One defines a training criterion C to be minimized in function of the desired outputs and of the outputs of the learner f(x). The function f is parameterized by the parameters of the network and belongs to a set of hypotheses H, that is the set of all functions that can be realized for different values of the parameters. The part of generalization error due to variance (due to the specific choice of training examples) can be controlled by making strong assumptions on the model, i.e., by choosing a small hypotheses space H . But using an incorrect model also worsens performance. Over the last few years, methods for automatically choosing H based on similar tasks have been studied. They consider that a learner is embedded in a world where it faces many related tasks and that the knowledge acquired when learning a task can be used to learn better and/or faster a new task. Some methods consider that the related tasks are not always all available at the same time (Pratt, 1993; Silver and Mercer, 1995): knowledge acquired when learning a previous task is transferred to a new task. Instead, all tasks may be learned in parallel (Baxter, 1995; Caruana, 1995), and this is the approach followed here. Our objective is not to use multi-task learning to improve the speed of learning the training data (Pratt, 1993; Silver and Mercer, 1995), but instead to improve generalization performance. For example, in (Baxter, 1995), several neural networks (one for each task) are trained simultaneouly. The networks share their first hidden layers, while all the remaining layers are specific to each network. The shared layers use the knowledge provided from the training examples of all the tasks to build an internal representation suitable for all these tasks. The remaining layers of each network use the internal representation to learn a specific task. In the multitask learning method used by Caruana (Caruana, 1995), all the hidden 948 J. Ghosn and Y. Bengio layers are shared. They serve as mutual sources of inductive bias. It was also suggested that besides the relevant tasks that are used for learning, it may be possible to use other related tasks that we do not want to learn but that may help to further bias the learner (Caruana, Baluja and Mitchell, 1996; Intrator and Edelman, 1996). In the family discovery method (Omohundro, 1996), a parameterized family of models is built. Several learners are trained separately on different but related tasks and their parameters are used to construct a manifold of parameters. When a new task has to be learned, the parameters are chosen so as to maximize the data likelihood on the one hand, and to maximize a "family prior" on the other hand which restricts the chosen parameters to lie on the manifold. In all these methods, the values of some or all the parameters are constrained. Such models restrict the size of the hypotheses space sufficiently to ensure good generalization performance from a small number of examples. 3 Application to Stock Selection We apply the ideas of multi-task learning to a problem of stock selection and portfolio management. We consider a universe of 36 assets, including 35 risky assets and one risk-free asset. The risky assets are 35 Canadian large-capitalization stocks from the Toronto Stock Exchange. The risk-free asset is represented by 90-days Canadian treasury bills. The data is monthly and spans 8 years, from February 1986 to January 1994 (96 months). Each month, one can buy or sell some of these assets in such a way as to distribute the current worth between these assets. We do not allow borrowing or short selling, so the weights of the resulting portfolio are all non-negative (and they sum to 1). We have selected 5 explanatory variables, 2 of which represent macro-economic variables which are known to influence the business cycle, and 3 of which are microeconomic variables representing the profitability of the company and previous price changes of the stock. The macro-economic variables were derived from yields of long-term bonds and from the Consumer Price Index. The micro-economic variables were derived from the series of dividend yields and from the series of ratios of stock price to book value of the company. Spline extrapolation (not interpolation) was used to obtain monthly data from the quarterly or annual company statements or macro-economic variables. For these variables, we used the dates at which their value was made public, not the dates to which they theoretically refer. To take into account the non-stationarity of the financial and economic time-series, and estimate performance over a variety of economic situations, multiple training experiments were performed on different training windows, each time testing on the following 12 months. For each architecture, 5 such trainings took place, with training sets of size 3, 4, 5, 6, and 7 years respectively. Furthermore, multiple such experiments with different initial weights were performed to verify that we did not obtain "lucky" results due to particular initial weights. The 5 concatenated test periods make an overall 5-year test period from February 1989 to January 1994. The training algorithm is described in (Bengio, 1996) and is based on the optimization of the neural network parameters with respect to a financial criterion (here maximizing the overall profit). The outputs of the neural network feed a trading Multi-Task Learning/or Stock Selection 949 module. The trading module has as input at each time step the output of the network, as well as, the weights giving the current distribution of worth between the assets. These weights depend on the previous portfolio weights and on the relative change in value of each asset (due to different price changes). The outputs of the trading module are the current portfolio weights for each of the assets. Based on the difference between these desired weights and the current distribution of worth, transactions are performed. Transaction costs of 1 % (of the absolute value of each buy or sell transaction) are taken into account. Because of transaction costs, the actions of the trading module at time t influence the profitability of its future actions. The financial criterion depends in a non-additive way on the performance of the network over the whole sequence. To obtain gradients of this criterion with respect to the network output we have to backpropagate gradients backward through time, through the trading module, which computes a differentiable function of its inputs. Therefore, a gradient step is performed only after presenting the whole training sequence (in order, of course). In (Bengio, 1996), we have found this procedure to yield significantly larger profits (around 4% better annual return), at comparable risks, in comparison to training the neural network to predict expected future returns with the mean squared error criterion. In the experiments, the ANN was trained for 120 epochs. 4 Experimental Results Four sets of experiments with different types of parameter sharing were performed, with two different architectures for the neural network: 5-3-1 (5 inputs, a hidden layer of 3 units, and 1 output), 5-3-2-1 (5 inputs, 3 units in the first hidden layer, 2 units in the second hidden layer, and 1 output). The output represents the belief that the value of the stock is going to increase (or the expected future return over three months when training with the MSE criterion). Four types of parameter sharing between the different models for each stock are compared in these experiments: sharing everything (the same parameters for all the stocks), sharing only the parameters (weights and biases) of the first hidden layers, sharing only the output layer parameters, and not sharing anything (independent models for each stock). The main results for the test period, using the 5-3-1 architecture, are summarized in Table 1, and graphically depicted in Figure 1 with the worth curves for the four types of sharing. The results for the test period, using the 5-3-2-1 architecture are summarized in Table 2. The ANNs were compared to two benchmarks: a buy-andhold benchmark (with uniform initial weights over all 35 stocks), and the TSE300 Index. Since the buy-and-hold benchmark performed better (8.3% yearly return) than the TSE300 Index (4.4% yearly return) during the 02/89-01/94 test period, Tables 1, and 2 give comparisons with the buy-and-hold benchmark. Variations of average yearly return on the test period due to different initial weights were computed by performing each of the experiments 18 times with different random seeds. The resulting standard deviations are less than 3.7 when no parameters or all the parameters are shared, less than 2.7 when the parameters of the first hidden layers are shared, and less than 4.2 when the output layer is shared. The values of beta and alpha are computed by fitting the monthly return of the portfolio r p to the return of the benchmark r M , both adjusted for the risk-free return 950 J. Ghosn and Y. Bengio Table 1: Comparative results for the 5-3-1 architecture: four types of sharing are compared with the buy-and-hold benchmark (see text). buy & share share share no hold all hidden output sharing A verage yearly return 8.3% 13% 23.4'70 24.8% 22.8% Standard deviation (monthly) 3.5% 4.3% 5.3% 5.3% 5.2% Beta 1 1.07 1.30 1.26 1.26 Alpha (yearly) 0 9% 20.6% 21.8% 19.9% t-statistic for alpha = 0 NA 11 14.9 15 14 Reward to variability 0.9% 9.6% 22.9% 24.7% 22.3% Excess return above benchmark 0 4.7% 15.1% 16.4% 14.5% Maximum drawdown 15.7% 13.3% 13.4% 13.3% 13.3% Table 2: Comparative results for the 5-3-2-1 architecture: three types of sharing are compared with the buy-and-hold benchmark (see text). buy & share share first share all no hold all hidden hidden sharing Average yearly return 8.3% 12.5% 22.7% 23% 9.1% Standard deviation (monthly) 3.5% 4.% 5.2% 5.2% 3.1% Beta 1 1.02 1.25 1.28 0.87 Alpha (yearly) 0 8.2% 19.7% 20.1% 4.% t-statistic for alpha = 0 NA 12.1 14.1 14.8 21.2 Reward to variability 0.9% 9.3% 22.2% 22.5% 2.5% Excess return above benchmark 0 4.2% 14.4% 14.7% 0.8% Maximum drawdown 15.7% 13% 12.6% 13.4% 10% ri (interest rates), according to the linear regression E(rp -rd = alpha + beta(rMr i). Beta is simply the ratio of the covariance between the portfolio return and the market return with the variance of the market. According to the Capital Asset Pricing Model (Sharpe, 1964), beta gives a measure of "systematic" risk, i.e., as it relates to the risk of the market, whereas the variance of the return gives a measure of total risk. The value of alpha in the tables is annualized (as a compound return): it represents a measure of excess return (over the market benchmark) adjusted for market risk (beta). The hypothesis that alpha = 0 is clearly rejected in all cases (with t-statistics above 9, and corresponding p-values very close to 0). The reward to variability (or "Sharpe ratio") as defined in (Sharpe, 1966), is another riskadjusted measure of performance: h.=2:J, where O"p is the standard deviation of Up the portfolio return (monthly returns were used here). The excess return above benchmark is the simple difference (not risk-adjusted) between the return of the portfolio and that of the benchmark. The maximum drawdown is another measure of risk, and it can be defined in terms of the worth curve: worth[t] is the ratio between the value of the portfolio at time t and its value at time o. The maximum drawdown is then defined as max ({max.<t worth[s])-worth[t)) t (max.:St worth[s]) Three conclusions clearly come out of the tables and figure: (1) The main improvement is obtained by allowing some parameters to be not shared (for the 5-3-1 a~chitecture, although the best results are obtained with a shared hidden and a free output layer, there are no significant differences between the different types of partial sharing, or no sharing at all). (2) Sharing some parameters yielded more consistent results (across architectures) than when not sharing at all. (3) The performance obtained in this way is very much better than that obtained by the benchmarks (buy-and-hold or TSE300), i.e., the yearly return is more than 14% above the best benchmark, while the risks are comparable (as measured by standard deviation of Multi-Task Learning/or Stock Selection 951 Share WO No Share Share WI Share A1l MSE Buy & Hold TSE300 3 .2 617 2 . 781 88 2 .30206 1. 82224 1. 34242 0 . 862602 8902 9002 91 02 9202 9302 Figure 1: Evolution of total worth in the 5-year test period 02/89-01/94, for the 5-3-1 architecture, and different types of sharing. From top to bottom: sharing the hidden layer, no sharing across stocks, sharing the output layer, sharing everything, sharing everything with MSE training, Buy and Hold benchmark, TSE300 benchmark. return or by maximum drawdown). 5 Future Work We will extend the results presented here in two directions. Firstly, given the impressive results obtained with the described approach, we would like to repeat the experiment on different data sets, for different markets. Secondly, we would like to generalize the type of multi-task learning by allowing for more freedom in the way the different tasks influence each other. Following (Omohundro, 1996), the basic idea is to re-parameterize the parameters (}i E Rnl of the ith model, for all n models in the following way: (}i = 1(Pi,W) where Pi E Rn2, W E Rn3 , and n x nl < n x n2 + n3 . For example, if 10 is an affine function, this forces the parameters of each the n different networks to lie on the same linear manifold. The position of a point on the manifold is given by a n 2dimensional vector Pi, and the manifold itself is specified by the n3 parameters of w. The expected advantage of this approach with respect to the one used in this paper is that different models (e.g., corresponding to different stocks) may "share" more or less depending on how far their Pi is from the Pj'S for other models. One does not have to specify which parameters are free and which are shared, but one has to specify how many are really free (n2) per model, and the shape of the manifold. 6 Conclusion The results presented of this paper show an interesting application of the ideas of multi-task learning to stock selection. In this paper we have addressed the question of whether ANNs trained for stock selection or portfolio management should be different for each stock or shared across all the stocks. We have found significantly better results when some or (sometimes) all of the parameters of the stock models are free (not shared). Since a parcimonuous model is always preferable, we conclude that partially sharing the parameters is even preferable, since it does not 952 J. Ghosn and Y. Bengio yield a deterioration in performance, and it yields more consistent results. Another interesting conclusion of this paper is that very large returns can be obtained at risks comparable to the market using a combination of partial parameter sharing and training with respect to a financial training criterion, with a small number of explanatory input features that include technical, micro-economic and macroeconomic information. References Baxter, J. (1995). Learning internal representations. In Proceedings of the Eighth International Conference on Computational Learning Theory, pages 311- 320, Santa Cruz, California. ACM Press. Bengio, Y. (1996). Using a financial training criterion rather than a prediction criterion. Technical Report #1019, Dept. Informatique et Recherche Operationnelle, Universite de Montreal. Caruana, R. (1995). Learning many related tasks at the same time with backpropagation. In Tesauro, G., Touretzky, D. S., and Leen, T. K., editors, Advances in Neural Information Processing Systems, volume 7, pages 657--664, Cambridge, MA. MIT Press. Caruana, R., Baluja, S., and Mitchell, T. (1996). Using the future to "sort out" the present: Rankprop and multitask learning for medical risk evaluation. In Advances in Neural Information Processing Systems, volume 8. Intrator, N. and Edelman, S. (1996). How to make a low-dimensional representation suitable for diverse tasks. Connection Science, Special issue on Transfer in Neural Networks. to appear. Moody, J., Levin, U., and Rehfuss, S. (1993). Predicting the U.S. index of industrial production. Neural Network World, 3(6):791-794. Omohundro, S. (1996). Family discovery. In Mozer, M., Touretzky, D., and Perrone, M., editors, Advances in Neural Information Processing Systems 8. MIT Press, Cambridge, MA. Pratt, L. Y. (1993). Discriminability-based transfer between neural networks. In Giles, C. L., Hanson, S. J., and Cowan, J., editors, Advances in Neural Information Processing Systems 5, pages 204-211, San Mateo, CA. Morgan Kaufmann. Refenes, A. (1994). Stock performance modeling using neural networks: a comparative study with regression models. Neural Networks, 7(2):375-388. Sharpe, W. (1964). Capital asset prices: A theory of market equilibrium under conditions of risk. Journal of Finance, 19:425-442. Sharpe, W. (1966). Mutual fund performance. Journal of Business, 39(1):119-138. Silver, D. L. and Mercer, R. E. (1995). Toward a model of consolidation: The retention and transfer of neural net task knowledge. In Proceedings of the INNS World Congress on Neural Networks, volume 3, pages 164-169, Washington, DC.
1996
16
1,210
Cholinergic Modulation Preserves Spike Timing Under Physiologically Realistic Fluctuating Input Akaysha C. Tang The Salk Institute Howard Hughes Medical Institute Computational Neurobiology Laboratory La Jolla, CA 92037 Terrence J. Sejnowski The Salk Institute Andreas M. Bartels Zoological Institute University of Zurich Ziirich Switzerland Howard Hughes Medical Institute Computational Neurobiology Laboratory La Jolla, CA 92037 Abstract Neuromodulation can change not only the mean firing rate of a neuron, but also its pattern of firing. Therefore, a reliable neural coding scheme, whether a rate coding or a spike time based coding, must be robust in a dynamic neuromodulatory environment. The common observation that cholinergic modulation leads to a reduction in spike frequency adaptation implies a modification of spike timing, which would make a neural code based on precise spike timing difficult to maintain. In this paper, the effects of cholinergic modulation were studied to test the hypothesis that precise spike timing can serve as a reliable neural code. Using the whole cell patch-clamp technique in rat neocortical slice preparation and compartmental modeling techniques, we show that cholinergic modulation, surprisingly, preserved spike timing in response to a fluctuating inputs that resembles in vivo conditions. This result suggests that in vivo spike timing may be much more resistant to changes in neuromodulator concentrations than previous physiological studies have implied. 112 A. C. Tang, A. M. Bartels and T. J. Sejnowski 1 Introduction Recently, there has been a vigorous debate concerning the nature of neural coding (Rieke et al. 1996; Stevens and Zador 1995; Shadlen and Newsome 1994). The prevailing view has been that the mean firing rate conveys all information about the sensory stimulus in a spike train and the precise timing of the individual spikes is noise. This belief is, in part, based on a lack of correlation between the precise timing of the spikes and the sensory qualities of the stimulus under study, particularly, on a lack of spike timing repeatability when identical stimulation is delivered. This view has been challenged by a number of recent studies, in which highly repeatable temporal patterns of spikes can be observed both in vivo (Bair and Koch 1996; Abeles et al. 1993) and in vitro (Mainen and Sejnowski 1994). Furthermore, application of information theory to the coding problem in the frog and house fly (Bialek et al. 1991; Bialek and Rieke 1992) suggested that additional information could be extracted from spike timing. In the absence of direct evidence for a timing code in the cerebral cortex, the role of spike timing in neural coding remains controversial. 1.1 A necessary condition for a spike timing code If spike timing is important in defining a stimulus, precisely timed spikes must be maintained under a range of physiological conditions. One important aspect of a neuron's environment is the presence of various neuromodulators. Due to their widespread projections in the nervous system, major neuromodulators, such as acetylcholine (ACh) and norepinephrine (NA), can have a profound influence on the firing properties of most neurons. If a change in concentration of a neuromodulator completely alters the temporal structure of the spike train, it would be unlikely that spike timing could serve as a reliable neural code. A major effect of cholinergic modulation on cortical neurons is a reduction in spike frequency adaptation, which is characterized by a shortening of inter-spike-intervals and an increase in neuronal excitability (McCormick 1993; Nicoll 1988). One obvious consequence of this cholinergic effect is a modification of spike timing (Fig. 1 A). This modification of spike timing due to a change in neuromodulator concentration would seem to preclude the possibility of a neural code based on precise spike timing. 1.2 Re-examination of the cholinergic modulation of spike timing Despite its popularity, the square pulse stimulus used in most eletrophysiological studies is rarely encountered by a cortical neuron under physiological conditions. The corresponding behavior of the neuron at the input/output level may have limited relevance to the behavior of the neuron under its natural condition, which is characterized in vivo by highly fluctuating synaptic inputs. In this paper, we reexamine the effect of cholinergic modulation on spike timing under two contrasting stimulus conditions: the physiologically unrealistic square pulse input versus the more plausible fluctuating input. We report that under physiologically more realistic fluctuating inputs, effects of cholinergic modulation preserved the timing of each individual spike (Fig. IB). This result is consistent with the hypothesis that spike timing may be relevant to information encoding. Cholinergic Modulation Preserves Spike Timing 113 2 Methods 2.1 Experimental Using the whole cell patch-clamp technique, we made somatic recordings from layer 2/3 neocortical neurons in the rat visual cortex. Coronal slices of 400 p.m were prepared from 14 to 18 days old Long Evans rats (for details see (Mainen and Sejnowski 1994). Spike trains elicited by current injection of 900 ms were recorded for the square pulse inputs and fluctuating inputs with equal mean synaptic inputs, in the absence and presence of a cholinergic agonist carbachol. The fluctuating inputs were constructed from Gaussian noise and convolved with an alpha function with a time constant of 3 ms, reflecting the time course of the synaptic events. The amplitude of fluctuation was such that the subthreshold membrane potential fluctuation observed in our experiments were comparable to that in whole-cell patch clamp study in vivo (Ferster and Jagadeesh 1992). The cholinergic agonist carbachol at concentrations of 5,7.5, 15,30 p.M was delivered through bath perfusion (perfusion time: between 1 and 6 min). For each cell, three sets of blocks were recorded before, during and after carbachol perfusion at a given concentration. Each block contained 20 trials of stimulation under identical experimental conditions. 2.2 Simulation We used a compartmental model of a neocortical neuron to explore the contribution of three potassium conductances affected by cholinergic modulation (Madison et ai. 1987). Simulations were performed in a reduced 9 compartment model, based on a layer 2 pyramidal cell reconstruction using the NEURON program. The model had five conductances: gNa, gK v , 9KM' gCa, gK(Ca)' Membrane resistivity was 40KOcm2 , capacitance was Ip.F/ p.m2 , and axial resistance was 2000cm. Intrinsic noise was simulated by injecting a randomly fluctuating current to fit the spike jitter observed experimentally. Different potassium conductances were manipulated as independent variables and the spike timing displacement was measured for multiple levels of conductance change corresponding to multiple concentrations of carbachol. 2.3 Data analysis For both experimental and simulation data, first derivatives were used to detect spikes and to determine the timing for spike initiation. Raster plots of the spike trains were derived from the series of membrane potentials for each trial, and a smoothed histogram was then constructed to reflect the instantaneous firing rate for each block of trials under identical stimulation and pharmacological conditions. An event was then defined as a period of increase in instantaneous firing rate that is greater than a threshold level (set at 3 times of the mean firing rate within the block of trials) (Mainen and Sejnowski 1994). The effect of carbachol on spike timing under fluctuating inputs was quantified by defining the displacement in spike timing for each event, dj, as the time difference between the nearest peaks of the events under carbachol and control condition. The weight for each event, Wi, is determined by the peak of the event. The higher the peak, the less the spike jitter. The mean displacement is (1) 114 A. C. Tang, A. M. Bartels and T. J. Sejnowski where i= 1, 2, ... nth event in the control condition. 3 Results 3.1 Experimental The effects of carbachol on spike timing under the square pulse and fluctuating inputs are shown in Fig. lA and B respectively. In the absence of carbachol, a square pulse input produced a spike train with clear spike frequency adaptation (Fig. lAl) . Similar to previous reports from the literature, addition of carbachol to the perfusion medium reduced spike frequency adaptation (Fig. lA2). This reduction in spike frequency adaptation is reflected in the shortening of inter-spikeintervals and an increase in the firing frequency. Most importantly, spike timing was altered by carbachol perfusion. When a fluctuating current was injected, the strong spike frequency adaptation observed under a square pulse input was no longer apparent (Fig. IBl). Unlike the results under the square pulse condition, addition of carbachol to the bath medium preserved the timing of the spikes (Fig. IB2). An increased excitability was achieved with the insertion of additional spikes between the existing spikes. At 8t 82 Figure 1: Response of a cortical neuron to square pulse current injection (A) and a fluctuating input (B). The membrane potential during the 1024 ms sampling period is plotted as a function of time for the two types of inputs (onset: 5 ms; duration: 900 ms). The grey lines show where the spikes occurred in the upper traces. Preservation of spike timing under carbachol was examined at concentrations of 5, 7.5, 15, and 30J.tM, here shown in one cell (Fig. 2, 5J.tM). The smoothed histograms (as described in section 2.3) were plotted for blocks of 20 identical trials under the same fluctuating input. The alignment of the events between the control and carbachol indicates that spike timing was well preserved. The table gives the mean spike displacement, D, for a range of carbachol concentrations. The spike jitter within the control and carbachol conditions was approximately 1 ms, and was not changed significantly by carbachol (control: 0.96 ± 0.3; carbachol: 0.94 ± 0.42 ms.) 3.2 Simulation The model captured the basic characteristics of experimental data. In response to fluctuating inputs, the model neurons showed reduced spike frequency adaptation and preservation of spike timing. The in vitro experiment were limited to only two levels of stimulus fluctuation. To show that reduced adaptation in response Cholinergic Modulation Preserves Spike Timing 115 Control Cbolinergic Modification of Spike Timing 1 D(ms) N Carbachol (microM) l 1 r 2.76± 0.38 15 5-7.5 3.33 1 15 9.3 2 30 Carbachol Figure 2: Preservation of spike timing for a range of carbachol concentrations. Left: the top portion is the histogram for the control condition; the bottom is the histogram for the carbachol condition shown inverted. The alignment of the events between the control and carbachol indicates preserved timing. Right: statistics of spike displacement. 55 ~ 48 c:: 41 Q '::I S Q, 34 .a < 27 20 0 50 100 150 Fluctuation (pA) Figure 3: Reduced adaptation as a function of increasing stimulus fluctuation. Adaptation measured as a normalized spike count difference between the first and second halves of the 900 ms stimulation: (C2-Cl)/Cl. to fluctuating inputs is a general phenomenon, in the model neuron we measured adaptation for multiple levels of stimulus fluctuation. As shown in Fig. 3, spike frequency adaptation decreased as a function of increasing stimulus fluctuation over a range of fluctuation amplitude. The effects cholinergic modulation on spike timing were studied under simulated cholinergic modulation. Similar to the experimental finding, increased neuronal excitability to fluctuating inputs was accompanied by insertion of additional spikes (Fig. 4 left) and spike timing was preserved simultaneously (Fig. 4 right). In real neurons, the total effects of cholinergic modulation depends on its effects on at least three potassium conductances. Using the model, we examined the effects of manipulating each of the three potassium conductances on spike displacement and spike jitter. We found that (1) spike displacement due to reduction in potassium conductances were all very small, on the order of a few milliseconds (Fig. 5 top row); (2) Compared to the conductances underlying 1M and I,eak , spike displacement was most sensitive to changes in the conductance underlying IAHP (Fig. 5 top row), whose reduction alone led to the best reproduction of the experimental data; (3) spike jitters of approximately 1 ms were independent of the values of the three 116 lOOms 1 30mV A. C. Tang, A. M. Bartels and T. 1. Sejnowski I Figure 4: Preservation of spike timing in the model neocortical neuron. Left: Responses of the model neuron to fluctuating input. Top: replicating data from the control condition. Bottom: reproducing the carbachol effect by blocking the adaptation current, IAHP. Right: histogram display of preservation of spike timing in a block of 20 trials. potassium conductances (Fig. 5 bottom row) . These results make predictions for new experiments where each individual current is blocked selectively. 4 Conclusions The results showed that under the physiologically realistic fluctuating input, the effects of cholinergic modulation on spike timing are rather different from that observed when unphysiological square pulse inputs were used. Instead of moving the spikes forward in time by shortening the inter-spike-intervals, cholinergic modulation preserved spike timing. This preservation of spike timing was achieved simultaneously with an increase in neuronal excitability. According to the classical view of neuromodulation, one would have expected that a spike timing based neural code would be difficult to maintain across a range of neuromodulator concentrations. The fact that spike timing was rather resistant to changes in the neuromodulatory environment raises the possibility that spike timing may serve some function in the cortex. The differential effect of cholinergic modulation on spike timing observed under the square pulse and fluctuating inputs also calls for caution in generalizing an observation from one set of parameter values to another, especially when generalizing from in vitro to in vivo. This concern for external validity is particularly important for computational neuroscientists whose work involves integrating phenomena from the cellular, systems and finally, to behavioral levels. Acknowledgments Supported by the Howard Hughes Medical Institute. We are grateful to Zachary Mainen, Barak Pearlmutter, Raphael Ritz, Anthony Zador, David Horn, Chuck Stevens, William Bialek, and Christof Koch for helpful discussions. References Abeles, M., Bergman, H., Margalit, E., and Vaadia, E. (1993). Spatiotemporal Cholinergic Modulation Preserves Spike Timing c u EL __ 8'" ",8 - '-' 0.. '" is 5 : ~ gKleak ~~ 00 5 10 15 20 25 1:~ 1:~1:~ 0.5 0.5 0.5 o 0 0 o 25 50 75 0 15 30 45 0 5 10 15 20 25 Reduction of conductance (%) 117 Figure 5: Effects of individual conductance changes on spike timing. Top: spike displacement as a function of changing conductances. Bottom: spike jitter as a function of changing conductances. Each conductance was reduced from its control value which was determined by fitting experimentally observed spike trains. The range of change for the leak conductance was constrained by the experimentally observed resting membrane potential changes (avg. 5 m V.) firing patterns in the frontal cortex of behaving monkeys. J. Neurophysiol., 70, 1629-1638. Bair, W. and Koch, C. (1996). Temporal precision of spike trains in extrastriate cortex of the behaving Macaque monkey. Neural Computation, 8(6), 11841202. Bialek, W. and Rieke, F. (1992). Reliability and information transmission in spiking neurons. Trends Neurosci., 15,428-434. Bialek, W., Rieke, F., de Ruyter van Stevenick, R. R., and Warland, D. (1991). Reading a neural code. Science, 252, 1854-7. Ferster, D. and Jagadeesh, B. (1992) . EPSP-IPSP interactions in cat visual cortex studied with in vivo whole-cell patch recording. J. Neurosci., 12(4), 12621274. Madison, D. V. , Lancaster, B., and Nicoll, R. A. (1987) . Voltage clamp analysis of cholinergic action in the hippocampus. J. Neurosci., 7(3), 733-741. Mainen, Z. F. and Sejnowski, T. J. (1994). Reliability of spike timing in neocortical neurons. Science, 268, 1503-6. McCormick, D. A. (1993). Actions of acetylecholine in the cerebral cortex and thalamus and implications for function .. Prog. Brain Res., 98, 303- 308. Nicoll, R. (1988). The coupling of neurotransmitter receptors to ion channels in the brain. Science, 241, 545-550. Rieke, F., Warland, D., de Ruyter van Steveninck, R., and Bialek, W . (1996). Spikes: Exploring the Neural Code. MIT Press. Shadlen, M. N. and Newsome, W . T. (1994) . Noise, neural codes and cortical organization. Current Opinion in Neurobiology, 4, 569-579. Stevens, C. and Zador, A. (1995) . The enigma of the brain. Current Biology, 5, 1-2.
1996
17
1,211
Analytical Mean Squared Error Curves in Temporal Difference Learning Satinder Singh Department of Computer Science University of Colorado Boulder, CO 80309-0430 baveja@cs.colorado.edu Abstract Peter Dayan Brain and Cognitive Sciences E25-210, MIT Cambridge, MA 02139 bertsekas@lids.mit.edu We have calculated analytical expressions for how the bias and variance of the estimators provided by various temporal difference value estimation algorithms change with offline updates over trials in absorbing Markov chains using lookup table representations. We illustrate classes of learning curve behavior in various chains, and show the manner in which TD is sensitive to the choice of its stepsize and eligibility trace parameters. 1 INTRODUCTION A reassuring theory of asymptotic convergence is available for many reinforcement learning (RL) algorithms. What is not available, however, is a theory that explains the finite-term learning curve behavior of RL algorithms, e.g., what are the different kinds of learning curves, what are their key determinants, and how do different problem parameters effect rate of convergence. Answering these questions is crucial not only for making useful comparisons between algorithms, but also for developing hybrid and new RL methods. In this paper we provide preliminary answers to some of the above questions for the case of absorbing Markov chains, where mean square error between the estimated and true predictions is used as the quantity of interest in learning curves. Our main contribution is in deriving the analytical update equations for the two components of MSE, bias and variance, for popular Monte Carlo (MC) and TD(A) (Sutton, 1988) algorithms. These derivations are presented in a larger paper. Here we apply our theoretical results to produce analytical learning curves for TD on two specific Markov chains chosen to highlight the effect of various problem and algorithm parameters, in particular the definite trade-offs between step-size, Q, and eligibility-trace parameter, A. Although these results are for specific problems, we Analytical MSE Curves/or TD Learning 1055 believe that many ofthe conclusions are intuitive or have previous empirical support, and may be more generally applicable. 2 ANALYTICAL RESULTS A random walk, or trial, in an absorbing Markov chain with only terminal payoffs produces a sequence of states terminated by a payoff. The prediction task is to determine the expected payoff as a function of the start state i, called the optimal value function, and denoted v.... Accordingly, vi = E {rls1 = i}, where St is the state at step t, and r is the random terminal payoff. The algorithms analysed are iterative and produce a sequence of estimates of v" by repeatedly combining the result from a new trial with the old estimate to produce a new estimate. They have the form: viet) = Viet - 1) + a(t)oi(t) where vet) = {Viet)} is the estimate of the optimal value function after t trials, Oi (t) is the result for state i based on random trial t, and the step-size a( t) determines how the old estimate and the new result are combined. The algorithms differ in the os produced from a trial. Monte Carlo algorithms use the final payoff that results from a trial to define the Oi(t) (e.g., Barto & Duff, 1994). Therefore in MC algorithms the estimated value ofa state is unaffected by the estimated value of any other state. The main contribution of TD algorithms (Sutton, 1988) over MC algorithms is that they update the value of a state based not only on the terminal payoff but also on the the estimated values of the intervening states. When a state is first visited, it initiates a shortterm memory process, an eligibility trace, which then decays exponentially over time with parameter A. The amount by which the value of an intervening state combines with the old estimate is determined in part by the magnitude of the eligibility trace at that point. In general, the initial estimate v(O) could be a random vector drawn from some distribution, but often v(O) is fixed to some initial value such as zero. In either case, subsequent estimates, vet); t > 0, will be random vectors because of the random os. The random vector v( t) has a bias vector b( t) d~ E {v( t) - v"'} and a covariance matrix G(t) d~ E{(v(t) - E{v(t)})(v(t) - E{v(t)})T}. The scalar quantity of interest for learning curves is the weighted MSE as a function of trial number t, and is defined as follows: MSE(t) = l:i Pi(E{(Vi(t) - vi?}) = l:i pi(bl(t) + Gii(t», where Pi = {JJT [I - QJ -1 )d l:j (J.'T [I - QJ -1)j is the weight for state i, which is the expected number of visits to i in a trial divided by the expected length of a trial1 (J.'i is the probability of starting in state i; Q is the transition matrix of the chain). In this paper we present results just for the standard TD(A) algorithm (Sutton, 1988), but we have analysed (Singh & Dayan, 1996) various other TD-like algorithms (e.g., Singh & Sutton, 1996) and comment on their behavior in the conclusions. Our analytical results are based on two non-trivial assumptions: first that lookup tables are used, and second that the algorithm parameters a and A are functions of the trial number alone rather than also depending on the state. We also make two assumptions that we believe would not change the general nature of the results obtained here: that the estimated values are updated offline (after the end of each trial), and that the only non-zero payoffs are on the transitions to the terminal states. With the above caveats, our analytical results allow rapid computation of exact mean square error (MSE) learning curves as a function of trial number. 10ther reasonable choices for the weights, PI, would not change the nature of the results presented here. 1056 S. Singh and P. Dayan 2.1 BIAS, VARIANCE, And MSE UPDATE EQUATIONS The analytical update equations for the bias, variance and MSE are complex and their details are in Singh & Dayan (1996) they take the following form in outline: b(t) C(t) am + Bmb(t - 1) AS + BSC(t - 1) + IS (b(t - 1» (1) (2) where matrix Bm depends linearly on a(t) and BS and IS depend at most quadratically on a(t). We coded this detail in the C programming language to develop a software tool2 whose rapid computation of exact MSE error curves allowed us to experiment with many different algorithm and problem parameters on many Markov chains. Of course, one could have averaged together many empirical MSE curves obtained via simulation of these Markov chains to get approximations to the analytical MSE error curves, but in many cases MSE curves that take minutes to compute analytically take days to derive empirically on the same computer for five significant digit accuracy. Empirical simulation is particularly slow in cases where the variance converges to non-zero values (because of constant step-sizes) with long tails in the asymptotic distribution of estimated values (we present an example in Figure lc). Our analytical method, on the other hand, computes exact MSE curves for L trials in O( Istate space 13 L) steps regardless of the behavior of the variance and bias curves. 2.2 ANALYTICAL METHODS Two consequences of having the analytical forms of the equations for the update of the mean and variance are that it is possible to optimize schedules for setting a and A and, for fixed A and a, work out terminal rates of convergence for band C. Computing one-step optimal a's: Given a particular A, the effect on the MSE of a single step for any of the algorithms is quadratic in a. It is therefore straightforward to calculate the value of a that minimises MSE(t) at the next time step. This is called the greedy value of a. It is not clear that if one were interested in minimising MSE(t + t'), one would choose successive a(u) that greedily minimise MSE(t); MSE(t + 1); .... In general, one could use our formulre and dynamic programming to optimise a whole schedule for a( u), but this is computationally challenging. Note that this technique for setting greedy a assumes complete knowledge about the Markov chain and the initial bias and covariance of v(O), and is therefore not directly applicable to realistic applications of reinforcement learning. Nevertheless, it is a good analysis tool to approximate omniscient optimal step-size schedules, eliminating the effect of the choice of a when studying the effect of the A. Computing one-step optimal A'S: Calculating analytically the A that would minimize MSE(t) given the bias and variance at trial t - 1 is substantially harder because terms such as [/ - A(t)Q]-l appear in the expressions. However, since it is possible to compute MSE(t) for any choice of A, it is straightforward to find to any desired accuracy the Ag(t) that gives the lowest resulting MSE(t). This is possible only because MSE(t) can be computed very cheaply using our analytical equations. The caveats about greediness in choosing ag{t) also apply to Ag(t). For one of the Markov chains, we used a stochastic gradient ascent method to optimise A( u) and 2The analytical MSE error curve software is available via anonymous ftp from the following address: ftp.cs.colorado.edu /users/baveja/ AMse. tar.Z Analytical MSE Curves/or TD Learning 1057 a( u) to minimise MSE( t + tf) and found that it was not optimal to choose Ag (t) and ag(t) at the first step. Computing terminal rates of convergence: In the update equations 1 and 2, . b(t) depends linearly on b(t - 1) through a matrix Bm; and C(t) depends linearly on C(t - 1) through a matrix B S . For the case of fixed a and A, the maximal and minimal eigenvalues of Bm and B S determine the fact and speed of convergence of the algorithms to finite endpoints. If the modulus of the real part of any of the eigenvalues is greater than 1, then the algorithms will not converge in general. We observed that the mean update is more stable than the mean square update, i.e., appropriate eigenvalues are obtained for larger values of a (we call the largest feasible a the largest learning rate for which TD will converge). Further, we know that the mean converges to v" if a is sufficiently small that it converges at all, and so we can determine the terminal covariance. Just like the delta rule, these algorithms converge at best to an (-ball for a constant finite step-size. This amounts to the MSE converging to a fixed value, which our equations also predict. Further, by calculating the eigenvalues of Bm , we can calculate an estimate of the rate of decrease of the bias. 3 LEARNING CURVES ON SPECIFIC MARKOV CHAINS We applied our software to two problems: a symmetric random walk (SRW), and a Markov chain for which we can control the frequency of returns to each state in a single run (we call this the cyclicity of the chain). a) ... i ,.,. • ... ..... 1,., -----..:-.---;,;;:;; ... .---=,;; ... ~:;;:_~, =iiJ .... , T .... N~ b) " •• .. I·. I. . • 10.0 tellO 110.0 _0 110.0 TNlN ...... e"....,., MIlE DiIdll.IIan ) _,. c •• ,--------------, . . Figure 1: Comparing Analytical and Empirical MSE curves. a) analytical and empirical learning curves obtained on the 19 state SRW problem with parameters Ci = 0.01, ,\ = 0.9. The empirical curve was obtained by averaging together more than three million simulation runs, and the analytical and empirical MSE curves agree up to the fourth decimal place; b) a case where the empirical method fails to match the analytical learning curve after more than 15 million runs on a 5 state SRW problem. The empirical learning curve is very spiky. c) Empirical distribution plot over 15.5 million runs for the MSE at trial 198. The inset shows impulses at actual sample values greater than 100. The largest value is greater than 200000. Agreement: First, we present empirical confirmation of our analytical equations on the 19 state SRW problem. We ran TD(A) for specific choices of a and A for more than three million simulation runs and averaged the resulting empIrical weighted MSE error curves. Figure la shows the analytical and empirical learning curves, which agree to within four decimal places. Long-Tails of Empirical MSE distribution: There are cases in which the agreement is apparently much worse (see Figure Ib). This is because of the surprisingly long tails for the empirical MSE distribution - Figure lc shows an example for a 5 1058 S. Singh and P. Dayan state SRW. This points to interesting structure that our analysis is unable to reveal. Effect of a and A: Extensive studies on the 19 state SRW that we do not have space to describe fully show that: HI) for each algorithm, increasing a while holding A fixed increases the asymptotic value of MSE, and similarly for increasing A whilst holding a constant; H2) larger values of a or A (except A very close to 1) lead to faster convergence to the asymptotic value of MSE if there exists one; H3) in general, for each algorithm as one decreases A the reasonable range of a shrinks, i.e., larger a can be used with larger A without causing excessive MSE. The effect in H3 is counter-intuitive because larger A tends to amplify the effective step-size and so one would expect the opposite effect. Indeed, this increase in the range of feasible a is not strictly true, especially very near A = 1, but it does seem to hold for a large range of A. Me versus TD(A): Sutton (1988) and others have investigated the effect of A on the empirical MSE at small trial numbers and consistently shown that TD is better for some A < 1 than MC (A = 1). Figure 2a shows substantial changes as a function of trial number in the value of A that leads to the lowest MSE. This effect is consistent with hypotheses HI-H3. Figure 2b confirms that this remains true even if greedy choices of a tailored for each value of A are used. Curves for different values of A yield minimum MSE over different trial number segments. We observed these effects on several Markov chains. a) b) Accumulate .,. , \ \ \ . • oo \". :--... ••• ~'':: _____ .o--------.----_u)_ •. __ 0.8 0.8 ,.,..~ ... ----.... ------00000 50 1OQ.0· 150.0 100.0 HO.O Tn.! Numbor Figure 2: U-shaped Curves. a) Weighted MSE as a function of A and trial number for fixed ex = 0.05 (minimum in A shown as a black line). This is a 3-d version of the U-shaped curves in Sutton (1988), with trial number being the extra axis. b) Weighted MSE as a function of trial number for various A using greedy ex. Curves for different values of A yield minimum MSE over different trial number segments. Initial bias: Watkins (1989) suggested that A trades off bias for variance, since A ..... 1 has low bias, but potentially high variance, and conversely for A ..... 0. Figure 3a confirms this in a problem which is a little like a random walk, except that it is highly cyclic so that it returns to each state many times in a single trial. If the initial bias is high (low), then the initial greedy value of A is high (low). We had expected the asymptotic greedy value of A to be 0, since once b(t) ..... 0, then A = ° leads to lower variance updates. However, Figure 3a shows a non-zero asymptote - presumably because larger learning rates can be used for A > 0, because of covariance. Figure 3b shows, however, that there is little advantage in choosing A cleverly except in the first few trials, at least if good values of a are available. Eigenvalue stability analysis: We analysed the eigenvalues of the covariance update matrix BS (c.f. Equation 2) to determine maximal fixed a as a function of A. Note that larger a tends to lead to faster learning, provided that the values converge. Figure 4a shows the largest eigenvalue of B S as a function of A for various Analytical MSE Curvesfor TD Learning 1059 Accumulate a) 1.0 ,-----~-____ -----, b) 02 0.0 ':-------,~---:-:':":__----:-::'. 0.0 SO.O 100.0 lSO.0 -Figure 3: Greedy A for a highly cyclic problem. a) Greedy A for high and low initial bias (using greedy a). b) Ratio of MSE for given value of A to that for greedy A at each trial. The greedy A is used for every step. a) 4.0 ,---~-~-~-~---, 1 3.0 \\ i '. , " 1 2.0 ••• ~:.. 5 ... -.. 10 "-•• . ...., 0.0 ':-----:"0--......,...,.--:":--,.__----' 0.0 0.2 0.4 0.' 0.' 1.0 A b) Largest Feasible a c) 0.10 ... ,... ........ 0.1 ·_· .... ,... ....... ·o,01 '-0 .. " 'I 0.00 O·OIo.~O --------,,0A':"". ------:".0 0·900.'-0--02--0~ .4--0~.6--0.~8 --',.0 A Figure 4: Eigenvalue analysis of covariance reduction. a) Maximal modulus of the eigenvalues of B S • These determine the rate of convergence of the variance. Values greater than 1 lead to instability. b) Largest a such that the covariance is bounded. The inset shows a blowup for 0.9 ;:; A ;:; 1. Note that A = 1 is not optimal. c) Maximal bias reduction rates as a function of A, after controlling for asymptotic variance (to 0.1 and 0.01) by choosing appropriate a's. Again, A < 1 is optimal. a. If this eigenvalue is larger than 1, then the algorithm will diverge - a behavior that we observed in our simulations. The effect of hypothesis H3 above is evident for larger A, only smaller a can be used. Figure 4b shows this in more graphic form, indicating the largest a that leads to stable eigenvalues for B S . Note the reversal very dose to A = 1, which provides more evidence against the pure MC algorithm. The choice of a and A control both rate of convergence and the asymptotic MSE. In Figure 4c we control for the asymptotic variance by choosing appropriate as as a function of .x and plot maximal eigenvalues of 8 m (c.f. Equation 1; it controls the terminal rate of convergence of the bias to zero) as a function of A. Again, we see evidence for T Dover Me. 4 CONCLUSIONS We have provided analytical expressions for calculating how the bias and variance of various TD and Monte Carlo algorithms change over iterations. The expressions themselves seem not to be very revealing, but we have provided many illustrations of their behavior in some example Markov chains. We have also used the analysis to calculate one-step optimal values of the step-size a and eligibility trace A parameters. Further, we have calculated terminal mean square errors and maximal bias reduction rates. Since all these results depend on the precise Markov chains chosen, 1060 S. Singh and P. Dayan it is hard to make generalisations. We have posited four general conjectures: HI) for constant A, the larger a, the larger the terminal MSE; H2) the larger a or A (except for A very close to I), the faster the convergence to the asymptotic MSE, provided that this is finite; H3) the smaller A, the smaller the range of a for which the terminal MSE is not excessive; H4) higher values of A are good for cases with high initial biases. The third of these is somewhat surprising, because the effective value of the step-size is really al(l - A). However, the lower A, the more the value of a state is based on the value estimates for nearby states. We conjecture that with small A, large a can quickly lead to high correlation in the value estimates of nearby states and result in runaway variance updates. Two main lines of evidence suggest that using values of A other than 1 (Le., using a temporal difference rather than a Monte-Carlo algorithm) can be beneficial. First, the greedy value of A chosen to minimise the MSE at the end of the step (whilst using the associated greedy a) remains away from 1 (see Figure 3). Second, the eigenvalue analysis of B S showed that the largest value of a that can be used is higher for A < 1 (also the asymptotic speed with which the bias can be guaranteed to decrease fastest is higher for A < 1). Although in this paper we have only discussed results for the standard TD(A) algorithm (called Accumulate), we have also analysed Replace TD(A) of Singh & Sutton (1996) and various others. This analysis clearly provides only an early step to understanding the course of learning for TD algorithms, and has focused exclusively on prediction rather than control. The analytical expressions for MSE might lend themselves to general conclusions over whole classes of Markov chains, and our graphs also point to interesting unexplained phenomena, such as the apparent long tails in Figure 1c and the convergence of greedy values of A in Figure 3. Stronger analyses such as those providing large deviation rates would be desirable. References Barto, A.G. &. Duff, M. (1994). Monte Carlo matrix inversion and reinforcement learning. NIPS 6, pp 687-694. Singh, S.P. &. Dayan, P. (1996). Analytical mean squared error curves in temporal difference learning. Machine Learning, submitted. Singh, S.P. &. Sutton, R.S. (1996). Reinforcement learning with replacing eligibility traces. Machine Learning, to appear. Sutton, R.S. (1988). Learning to predict by the methods of temporal difference. Machine Learning, 3, pp 9-44. Watkins, C.J.C.H. (1989). Learning from Delayed Rewards. PhD Thesis. University of Cambridge, England.
1996
18
1,212
Dynamics of Training Siegfried Bos* Lab for Information Representation RIKEN, Hirosawa 2-1, Wako-shi Saitama 351-01, Japan Abstract Manfred Opper Theoretical Physics III University of Wiirzburg 97074 Wiirzburg, Germany A new method to calculate the full training process of a neural network is introduced. No sophisticated methods like the replica trick are used. The results are directly related to the actual number of training steps. Some results are presented here, like the maximal learning rate, an exact description of early stopping, and the necessary number of training steps. Further problems can be addressed with this approach. 1 INTRODUCTION Training guided by empirical risk minimization does not always minimize the expected risk. This phenomenon is called overfitting and is one of the major problems in neural network learning. In a previous work [Bos 1995] we developed an approximate description of the training process using statistical mechanics. To solve this problem exactly, we introduce a new description which is directly dependent on the actual training steps. As a first result we get analytical curves for empirical risk and expected risk as functions of the training time, like the ones shown in Fig. l. To make the method tractable we restrict ourselves to a quite simple neural network model, which nevertheless demonstrates some typical behavior of neural nets. The model is a single layer perceptron, which has one N -dim. layer of adjustable weights W between input x and output z. The outputs are linear, Le. 1 N Z = h = r;:;r L WiXi . vN i=l (1) We are interested in supervised learning, where examples xf (J-L = 1, ... , P) are given for which the correct output z* is known. To define the task more clearly and to monitor the training process, we assume that the examples are provided by another network, the so called teacher network. The teacher is not restricted to linear outputs, it can have a nonlinear output function 9*(h*). * email: boesClzoo.riken.go.jpandopperCiphysik.uni-wuerzburg.de 142 S. Bos and M. Opper Learning by examples attempts to minimize the error averaged over all examples, Le. ET := 1/2 < (z~ - ZIl)2 >{il'}' which is called training error or empirical risk. In fact what we are interested in is a minimal error averaged over all possible inputs X, i.e EG := 1/2 < (z* - z)2 > {xEInput}' called generalization error or expected risk. It can be shown [see Bos 1995] that for random inputs, Le. all components Xi are independent and have zero means and unit variance, the generalization error can be described by the order parameters R and Q, 1 EG(t) = 2" [G - 2H R(t) + Q(t)] (2) with the two parameters G =< [g*(h)]2 >h and H =<g*(h) h>h. The order parameters are defined as: 1 N R(t) =< N L WtWi(t) >{wt} , i=l (3) As a novelty in this paper we average the order parameters not as usual in statistical mechanics over many example realizations {xt}, but over many teacher realizations {Wt}, where we use a spherical distribution. This corresponds to a Bayesian average over the unknown teacher. A study of the static properties of this model was done by Saad [1996]. Further comments about the averages can be found in the appendix. In the next section we introduce our new method briefly. Readers, who do not wish to go into technical details in first reading, can tum directly to the results (15) and (16). The remainder of the section can be read later, as a proof. In the third section results will be presented and discussed. Finally, we conclude the paper with a summary and a perspective on further problems. 2 DYNAMICAL APPROACH Basically we exploit the gradient descent learning rule, using the linear student, i.e g'(h) = 1 and zJl. = hll = )wWxJl., For P < N, the weights are linear combinations of the example inputs xr, if Wi(O) = 0, (5) After some algebra a recursion for (1 Jl.(t) can be found, Le. (6) where the term in the round brackets defines the overlap matrix C IlV. From the geometric series we know the solution of this recursion, and therefore for the weights (7) Dynamics o/Training 143 It fulfills the initial conditions Wi(O) = ° and Wi (l) = 1] 'L:=1 z~ xr (Hebbian), and yields after infinite time steps the so called Pseudo-inverse weights, i.e. p Wi(oo) = ~ L z~ (C- 1)/-IV xr . /-1,£1=1 (8) This is valid as long as the examples are linearly independent, i.e. P < N. Remarks about the other case (P > N) will follow later. With the expression (7) we can calculate the behavior of the order parameters for the whole training process. For R(t) we get R(t) = 1 ~ [E - (E - 1]C)t] /-I ( 1 ~ W* v) N 6 C < Z* 'N 6 i xi >{W,"} /-1,£1=1 /-IV V lV a=1 (9) For the average we have used expression (21) from the appendix. Similarly we get for the other order parameter, ( N ) /-IT 1 £ItT X < Z. Z* N LXi Xi >{W:} a=1 (10) Again we have applied an identity (20) from the appendix and we did some matrix algebra. Note, up to this point the order parameters were calculated without any assumption about the statistics of the inputs. The results hold, even without the thermodynamic limit. The trace can be calculated by an integration over the eigenvalues, thus we attain integrals of the following form, p ~mu ~ L [(E - 1]C)1 cm] /-1/-1 = J d~ p(~)(l- 1]~)1 ~m =: I!n(t, 0:, 1]), /-1=1 (11) ~min with l = {O, t, 2t} and m = {-1,0, 1}. These integrals can be calculated once we know the density of the eigenvalues p(~). The determination of this density can be found in recent literature calculated by Opper [1989J using replicas, by Krogh [1992J using perturbation theory and by Sollich [1994J with matrix identities. We should note, that the thermodynamic limit and the special assumptions about the inputs enter the calculation here. All authors found (12) 144 S. Bos and M. Opper for a < 1. The maximal and the minimal eigenvalues are ~max,min := (1 ± fo)2. So all that remains now is a numerical integration. Similarly we can calculate the behavior of the training error from 1 p ET(t) =< 2P L (z~ - hl-')2 >{Wn . (13) 1-'=1 For the overdetermined case (P > N) we can find a recursion analog to (6), W;{t + I} = ~ [6;; - (~ ~ x'tx'J ) 1 Wj{t} + .IN ~ z~x't . {14} The term in the round brackets defines now the matrix B ij . The calculation is therefore quite similar to above with the matrix B playing the role of matrix C. The density of the eigenvalues p(A) for matrix B is the one from above (12) multiplied bya. Altogether, we find the following results in the case of a < 1, G G - H2 (1 rl 2t ) H2 ( 2t) EG(t, a, 77) = "2 + 2 a 1 _ a - 2.1':"'1 + L1 - T a 1- 10 , G - H2 2t H2 2t Er(t,a,77) = 2 10 + 2 II , (15) and in the case of a > 1, G - H2 ( 1 t 2t ) H2 2t EG(t,a,77) = 2 1+ a_I-2I-1+L1 +T1o, G - H2 ( 1 l~t ) H2 1ft ET ( t, a, 77) = 1 - - + + . 2 a a 2 a (16) If t ---+ 00 all the time-dependent integrals Ik and .t;t vanish. The remaining first two terms describe, in the limit a -+ 00, the optimal convergence rate of the errors. In the next section we discuss the implications of this result. 3 RESULTS First we illustrate how well the theoretical results describe the training process. If we compare the theory with simulations, we find a very good correspondence, see Fig. 1. Trying other values for the learning rate we can see that there is a maximal learning rate. It is twice the inverse of the maximal eigenvalue of the matrix B, i.e. 2 2 77max = ~max = (1 + fo)2 . (17) This is consistent with a more general result, that the maximalleaming is twice the inverse of the maximal eigenvalue of the Hessian. In the case of the linear perceptron the matrix B is identical to the Hessian. As our approach is directly related to the actual number of training steps we can examine how training time varies in different training scenarios. Training can be stopped if the training error reaches a certain minimal value, i.e if ET(t) ~ E¥!in +f. Or, in cross-validated early stopping, we will terminate training if the generalization error starts to increase, i.e. if EG(t + 1) > EG(t). Dynamics of Training 145 0.4 0.3 0.2 , , , ET , , , , , , , , 0.1 , , , 'l , ----~ -----±----- ~-----±-----~-----%----- ~ - - - --%-- - --~-- - --%-- ---~ - - - -~ - - - - -~ ---0.0 0 50 100 150 training steps Figure 1: Behavior of the generalization error EG (upper line) and the training error ET (lower line) during the training process. As the loading rate a = P / N = 1.5 is near the storage capacity (a = 1) of the net overfitting occurs. The theory describes the results of the simulations very well. Parameters: learning rate TJ = 0.1, system size N = 200, and g .. (h) = tanh(')'h) with gain 'Y = 5. Fig. 2 shows that in exhaustive training the training time diverges for a near 1, in the region where also the overfitting occurs. In the same region early stopping shows only a slight increase in training time. Furthermore, we can guess from Fig. 2 that asymptotically only a few training steps are necessary to fulfill the stopping criteria. This has to be specified more precisely. First we study the behavior of EG after only one training step, i.e. t = 1. Since we interested in the limit of many examples (a -+ 00) we choose the learning rate as a fraction of the maximal learning rate (17), i.e. TJ = TJO/~max' Then we can calculate the behavior of EG(t = 1, a, ~;ix) analytically. We find that only in the case of TJo = 1, the generalization error can reach its asymptotic minimum Einf. The rate of the convergence is a-l like in the optimal case, but the prefactor is different. However, already for t = 2 we find, EG (t = 2, a, TJ = ~;!x) := EG - E;.nf = G -2 H2 a ~ 1 + 0 (~2) . (18) If a is large, so that we can neglect the a- 2 term, then two batch training steps are already enough to get the optimal convergence rate. These results are illustrated in Fig. 3. 4 SUMMARY In this paper we have calculated the behavior of the learning and the training error during the whole training process. The novel approach relates the errors directly to the actual number of training steps. It was shown how good this theory describes the training process. Several results have been presented, such as the maximal learning rate and the training time in different scenarios, like early stopping. If the learning rate is chosen appropriately, then only two batch training steps are necessary to reach the optimal convergence rate for sufficiently large a. 146 S. Bas and M. Opper t le+02 le+OI 40 0 , , , , , , ~ : ~ \ / ,: \0 ~~" A'". ,? \, .\i>. 4):' I;J/ " '. ~ .co ... ' ..\~ l·'.\.a. 4-. 0,_,-' 1_ "'P. cr-..t;>-L-~-<>-+-'1' -4 tl. ~ ... i O __ ~_i0 "-t-R o 0: ___ : '---: ,----_.. -----. eps=O.OOl eps=O.OlO early stop ...... \ ..... ~ ...... ~ , , -' .. ---------1 ............... -..... . le+OO le-Ol le+OO le+OI PIN , , -----------------------""'i~ ........... -, , " le+02 Figure 2: Number of necessary training steps to fulfill certain stopping criteria. The upper lines show the result if training is stopped when the training error is lower than E~in+€, with € = 0.001 (dotted line) and € = 0.01 (dashed line). The solid line is the early stopping result where training is stopped, when the generalization error started to increase, EG(t + 1) > EG(t). Simulation results are indicated by marks. Parameters: learning rate 11 = 0.01, system size N = 200, and 9.(h) = tanh(-yh) with gain '"Y = 5. Further problems, like the dynamical description of weight decay, and the relation of the dynamical approach to the thermodynamic description of the training process [see Bos, 1995] can not be discussed here due to lack of space. These problems are examined in an extended version of this work [Bos and Opper 1996]. It would be very interesting if this method could be extended towards other, more realistic models. A APPENDIX Here we add some identities which are necessary for the averages over the teacher weight distributions, eqs. (9) and (10). In the statistical mechanics approach one assumes that the distribution of the local fields h is Gaussian. This becomes true, if one averages over random inputs Xi, with first moments zero and one, which is the usual approach [see Bos 1995 and ref.]. In principle it is also possible to average over many tasks, i.e many teacher realizations W·, which is done here. The Gaussian local fields h~ fulfill, < h~ >= 0, < h~h~ >=CjJ.v' (19) This implies 00 00 < z~ z~ >{Wtl J Dh~ J Dh~ 9.(V1 - (CjJ.v)2 h~ + CjJ.v h~) 9.(h~) -00 -00 (20) In the second identity we first calculated the diagonal term and for the non-diagonal term we made an expansion assuming small correlations. Similarly the following Dynamics o/Training le+OO Ie-OI le-02 EO le-03 le-04 Ie-OS Ie-OI ---Ie+OO ............... le+OI le+02 PIN exh. opt. t=1 t=2 t=3 le+03 147 le+04 Figure 3: Behavior of EG = EG Einf after t training steps. Results for t = 1, 2 and 3 are given. For large enough a it is already after t = 2 training steps possible to reach the optimal convergence (solid line). If t = 3 the optimal result is reached even faster. Parameters: learning rate 17 = ~;!x and 9*(h) = tanhbh) with gain ')' = 5. identity can be proved, < z~ h~ >{W;}= 8,.£1/ H + (G,.W 8,." .. ) H. (21) Acknowledgment: We thank Shun-ichi Amari for many discussions and E. Helle, A. Stevenin-Barbier for proofreading and valuable comments. References Bas S. (1995), 'Avoiding overfitting by finite temperature learning and crossvalidation', in Int. Conference on Artificial Neural Networks 95 (ICANN'95), edited by EC2 & Cie, Vo1.2, p.1l1- 1l6. Bas S., and Opper M. (1996), 'An exact description of early stopping and weight decay', submitted. Kinzel W., and Opper M. (1995), 'Dynamics of learning', in Models of Neural Networks I, edited by E. Domany, J. L. van Hemmen and K. Schulten, Springer, p.157-179. Krogh A. (1992), 'Learning with noise in a linear perceptron', J. Phys. A 25, p.1l35-1l47. Opper M. (1989), 'Learning in neural networks: Solvable dynamics', Europhys. Lett. 8, p.389-392. Saad D. (1996), 'General Gaussian priors for improved generalization', submitted to Neural Networks. Sollich P. (1995), 'Learning in large linear perceptrons and why the thermodynamic limit is relevant to the real world', in NIPS 7, p.207-214.
1996
19
1,213
A variational principle for model-based morphing Lawrence K. Saul'" and Michael I. Jordan Center for Biological and Computational Learning Massachusetts Institute of Technology 79 Amherst Street, EI0-034D Cambridge, MA 02139 Abstract Given a multidimensional data set and a model of its density, we consider how to define the optimal interpolation between two points. This is done by assigning a cost to each path through space, based on two competing goals-one to interpolate through regions of high density, the other to minimize arc length. From this path functional, we derive the Euler-Lagrange equations for extremal motionj given two points, the desired interpolation is found by solving a boundary value problem. We show that this interpolation can be done efficiently, in high dimensions, for Gaussian, Dirichlet, and mixture models. 1 Introduction The problem of non-linear interpolation arises frequently in image, speech, and signal processing. Consider the following two examples: (i) given two profiles of the same face, connect them by a smooth animation of intermediate poses[l]j (ii) given a telephone signal masked by intermittent noise, fill in the missing speech. Both these examples may be viewed as instances of the same abstract problem. In qualitative terms, we can state the problem as follows[2]: given a multidimensional data set, and two points from this set, find a smooth adjoining path that is consistent with available models of the data. We will refer to this as the problem of model-based morphing. In this paper, we examine this problem it arises from statistical models of multidimensional data. Specifically, our focus is on models that have been derived from Current address: AT&T Labs, 600 Mountain Ave 2D-439, Murray Hill, NJ 07974 268 L K. Saul and M. I. Jordan some form of density estimation. Though there exists a large body of work on the use of statistical models for regression and classification, there has been comparatively little work on the other types of operations that these models support. Non-linear morphing is an example of such an operation, one that has important applications to video email[3], low-bandwidth teleconferencing[4]' and audiovisual speech recognition [2] . A common way to describe multidimensional data is some form of mixture modeling. Mixture models represent the data as a collection of two or more clusters; thus, they are well-suited to handling complicated (multimodal) data sets. Roughly speaking, for these models the problem of interpolation can be divided into two tasks- how to interpolate between points in the same cluster, and how to interpolate between points in different clusters. Our paper will therefore be organized along these lines. Previous studies of morphing have exploited the properties of radial basis function networks[l] and locally linear models[2]. We have been influenced by both these works, especially in the abstract formulation of the problem. New features of our approach include: the fundamental role played by the density, the treatment of nonGaussian models, the use of a continuous variational principle, and the description of the interpolant by a differential equation. 2 Intracluster interpolation Let Q = {q(1), q(2), .. . , qlQI} denote a set of multidimensional data points, and let P( q) denote a model of the distribution from which these points were generated. Given two points, our problem is to find a smooth adjoining path that respects the statistical model of the data. In particular, the desired interpolant should not pass through regions of space that the modeled density P( q) assigns low probability. 2.1 Clusters and metrics To develop these ideas further, we begin by considering a special class of modelsnamely, those that represent clusters. We say that P( q) models a data cluster if P( q) has a unique (global) maximum; in turn, we identify the location of this maximum, q"', as the prototype. Let us now consider the geometry of the space inhabited by the data. To endow this space with a geometric structure, we must define a metric, ga,B( q), that provides a measure of the distance between two nearby points: 1 V[q, q + dq] = [~gap(q) dqadqp r + 0 (Idql') . (1) Intuitively speaking, the metric should reflect the fact that as one moves away from the center of the cluster, the density of the data dies off more quickly in some directions than in others. A natural choice for the metric, one that meets the above criteria, is the negative Hessian of the log-likelihood: (2) A Variational Principle for Model-based Morphing 269 This metric is positive-definite if In P( q) is concave; this will be true for all the examples we discuss. 2.2 From densities to paths The problem of model-based interpolation is to balance two competing goalsone to interpolate through regions of high density, the other to avoid excessive deformations. Using the metric in eq. (1), we can now assign a cost (or penalty) to each path based on these competing goals. Consider the path parameterized by q(t) . We begin by dividing the path into segments, each of which is traversed in some small time interval, dt. We assign a value to each segment by _ {[P(q(t»] _l}V[q(t),q(t+dt)] ¢(t) P( q*) e , (3) where f ~ O. For reasons that will become clear shortly, we refer to f as the line tension. The value assigned to each segment dep~n~s on two terms: a ratio of probabilities, P( q(t»j P( q*), which favors points near the prototype, and the constant multiplier, e-l . Both these terms are upper bo~nded by unity, and hence so is their product. The value of the segment also decays with its length, as a result of the exponent, V[q(t), q(t + dt)]. We derive a path functional by piecing these segments together, multiplying their individual contributions, and taking the continuum limit. A value for the entire path is obtained from the product: e-S = II ¢(t). (4) Taking the logarithm of both sides, and considering the limit dt -+ 0, we obtain the path functional 1 S[q( t)1 = J {-In [ ~~!i) 1 +l H ~ go~( q)4 0 <il dt, (5) where q == it [q] is the tangent vector to the path at time t. The terms in this functional balance the two competing goals for non-linear interpolation. The first favors paths that interpolate through regions of high density, while the second favors paths with small arc lengths; both are computed under the metric induced by the modeled density. The line tension f determines the cost per unit arc length and modulates the competition between the two terms. Note that the value of the functional does not depend on the rate at which the path is traversed. To minimize this functional, we use the following result from the calculus of variations. Let £(q, q) denote the integrand of eq. (5), such that S[q(t)] = f dt £( q, q). Then the path which minimizes this functional obeys the EulerLagrange equations[5]: (6) 270 L. K. Saul and M. I. Jordan We define the model-based interpolant between two points as the path which minimizes this functional; it is found by solving the associated boundary value problem. The function C( q, q) is known as the Lagrangian. In the next sections, we present eq. (5) for two distributions of interest-the multivariate Gaussian and the Dirichlet. 2.3 Gaussian cloud The simplest model of multidimensional data is the multivariate Gaussian. In this case, the data is modeled by _ IM11/2 {I T } P(x) (27r)N/2 exp -2" [x Mx] , (7) where M is the inverse covariance matrix and N is the dimensionality. Without loss of generality, we nave chosen the coordinate system so that the mean of the data coincides with the origin. For the Gaussian, the mean also defines the location of the prototype; moreover, from eq. (2), the metric induced by this model is just the inverse covariance matrix. From eq. (5), we obtain the path functional: (8) To find a model-based interpolant, we seek the path that minimizes this functional. Because the functional is parameterization-invariant, it suffices to consider paths that are traversed at a constant (unit) rate: xTMx = 1. From eq. (6), we find that the optimal path with this parameterization satisfies: {~ [xTMx] + f} x + [xTMx] x = x. (9) This is a set of coupled non-linear equations for the components of x(t). However, note that at any moment in time, the acceleration, x, can be expressed as a linear combination of the position, x, and the velocity, x. It follows that the motion of x lies in a plane; in particular, it lies in the plane spanned by the initial conditions, x and x, at time t = O. This enables one to solve the boundary value problem efficiently, even in very high dimensions. Figure la shows some solutions to this boundary value problem for different values of the line tension, f. Note how the paths bend toward the origin, with the degree of curvature determined by the line tension, f. 2.4 Dirichlet simplex For many types of data, the multivariate Gaussian distribution is not the most appropriate model. Suppose that the data points are vectors of positive numbers whose elements sum to one. In particular, we say that w is a probability vector if w = (Wi, W2, . .. , WN) E 'RN, Wct > 0 for all a, and I:Ct Wct = 1. Clearly, the multivariate Gaussian is not suited to data of this form, since no matter what the mean and covariance matrix, it cannot assign zero probability to vectors outside the simplex. Instead, a more natural model is the Dirichlet distribution: 8",-1 pew) = f(O) IJ ;(Oct) , (10) A Variational Principle for Model-based Morphing 271 where ()a > 0 for all Ct, and () == :La ()a. Here, f(.) is the gamma function, and ()a are parameters that determine the statistics of P(w). Note that P(w) = 0 for vectors that are not probability vectors; in particular, the simplex constraints on w are implicit assumptions of the model. We can rewrite the Dirichlet distribution in a more revealing form as follows. First, let w'" denote the probability vector with elements w~ = ()al(). Then, making a change of variables from w to In w, we have: P(ln w) = ;8 exp { () [KL (w"'llw)]}, (11) where Z8 is a normalization factor that depends on ()a (but not w), and the quantity in the exponent is () times the Kullback-Leibler (KL) divergence, KL(w"'llw)= Lw:ln [::J. a (12) The KL divergence measures the mismatch between wand w"', with KL(w"'llw) = 0 if and only if w = w·. Since KL(w"'llw) has no other minima besides the one at w"', we shall say that P(ln w) models a data cluster in the variable In w . The metric induced by this modeled density is computed by following the prescription of eq. (2). For two nearby points inside the simplex, wand w + dw, the result of this prescription is that the squared distance is given by ds2 = () L dw~ . a Wa (13) Up to a multiplicative factor of2(), eq. (13) measures the infinitesimal KL divergence between wand w + dw. This is a natural metric for vectors whose elements can be interpreted as probabilities. The functional for non-linear interpolation is found by substituting the modeled density and the induced metric into eq. (5). For the Dirichlet distribution, this gives: 1 S[w(t)] = J {O[KL(W'IlW )] H} [O~ ::r dt. (14) Our problem is to find the path that minimizes this functional. Because the functional is parameterization-invariant, it again suffices to consider paths that are traversed at a constant rate, or :La w;lwa = 1. In addition to this, however, we must also enforce the constraint that w remains inside the simplex; this is done by introducing a Lagrange multiplier. Following this procedure, we find that the optimal path is described by: [0 KL(w'llw) HJ {w. - 2~. + ~. } - 0 [~ :: Wp] w. = O(w. - w;). (15) Given two endpoints, this differential equation defines a boundary value problem for the optimal path. Unlike before, however, in this case the motion of w is not 272 L. K. Saul and M. I. Jordan confined to a plane. Hence, the boundary value problem for eq. (15) does not collapse to one dimension, as does its Gaussian counterpart, eq. (9). To remedy this situation, we have developed an efficient approximation that finds a near-optimal interpolant, in lieu of the optimal one. This is done in two steps: first, by solving eq. (15) exactly in the limit £ -+ 00; second, by using this limiting solution, WOO(t), to find the lowest-cost path that can be expressed as the convex combination: w(t) = m(t)w'" + [1 - m(t)) WOO(t). (16) The lowest-cost path of this form is found by substituting eq. (16) into the Dirichlet functional, eq. (14), and solving the Euler-Lagrange equations for m(t). The motivation for eq. (16) is that for finite £, we expect the optimal interpolant to deviate from WOO(t) and bend toward the prototype at w*. In practice, this approximation works very well, and by collapsing the boundary value problem to one dimension, it allows cheap computation of the Dirichlet interpolants. Some paths from eq. (16), as well as the £ -+ 00 paths on which they are based, are shown in figure lb. These paths were computed for the twelve dimensional simplex (N = 12), then projected onto the WI w2-plane. 3 Intercluster interpolation The Gaussian and Dirichlet distributions of the previous section are clearly inadequate for modeling for multimodal data sets. In this section, we extend the variational principle to mixture models, which describe the data as a collection of k 2: 2 clusters. In particular, suppose the data is modeled by k P(q) = L 7rzP(qlz). (17) z=1 Here, we have assumed that the conditional densities P( qlz) model data clusters as defined in section 2.1, and the coefficients 7rz = P(z) define prior probabilities for the latent variable, z E {I, 2, ... , k}. The crucial step for mixture models is to develop the appropriate generalization of eq. (5). To this end, let £z(q, q) denote the Lagrangian derived from the conditional density, P(qlz), and £z the line tensioni that appears in this Lagrangian. We now combine these Lagrangians into a single functional: S[q(t), z(t)) = J dt £z(t)(q, q). (18) Note that eq. (18) is a functional of two arguments, not one. For mixture models, which define a joint density P(q, z) = 7rzP(qlz), our goal is to find the optimal path in the joint space q ® z. Here, z(t) is a piecewise-constant function of time that assigns a discrete label to each point along the path; in other words, it provides a temporal segmentation of the path, q(t). The purpose of z(t) in eq. (18) is to select which Lagrangian is used to compute the contribution from the interval [t, t + dt). lTo respect the weighting of the mixture components in eq. (17), we set the line tensions according to iz = i-In 'Trz. Thus, components with higher weights have lower line tensions. A Variational Principle for Model-based Morphing 273 0.7 , * 0.0 '.CI> '. 'liS o. 0.5 -.---,.,\~.-.-.-. 0 .. 'lI ~0.3 '.0 -0. 0.2 .. ':1.(1) 0.1 '.1 j * • 00 0 .• 0.4 0.8 -6 0 W1 xl Figure 1: Model-based morphs for (a) Gaussian distribution; (b) Dirichlet distribution; (c) mixture of Gaussians. The prototypes are shown as asterisks; f denotes the line tension. Figure lc shows the convergence of the iterative algorithm; n denotes the number of iterations. As before, we define the model-based interpolant as the path q(t) that minimizes eq. (18). In this case, however, both q(t) and z(t) must be simultaneously optimized to recover this path. We have implemented an iterative scheme to perform this optimization, one that alternately (i) estimates the segmentation z(t), (ii) computes the model-based interpolant within each cluster based on this segmentation, and (iii) reestimates the points (along the cluster boundaries) where z(t) changes value. In short, the strategy is to optimize z(t) for fixed q(t), then optimize q(t) for fixed z(t). Figure lc shows how this algorithm operates on a simple mixture of Gaussians. In this example, the covariance matrices were set equal to the identity matrix, and the means of the Gaussians were distributed along a circle in the Xlx2-plane. Note that with each iteration, the interpolant converges more closely to the path that traverses this circle. The effect is similar to the manifold-snake algorithm of Bregler and Omohundro[2]. 4 Discussion In this paper we have proposed a variational principle for model-based interpolation. Our framework handles Gaussian, Dirichlet, and mixture models, and the resulting algorithms scale well to high dimensions. Future work will concentrate on the application to real images. References [1] T. Poggio and F. Girosi. Networks for approximation and learning. Proc. of IEEE, vol 78:9 (1990). [2] C. Bregler and S. Omohundro. Nonlinear image interpolation using manifold learning. In G. Tesauro, D. Touretzky, and T. Leen (eds.). Advances in Neural Information Processing Systems 7, 973-980. MIT Press, Cambridge, MA (1995). [3] T . Ezzat. Example based analysis and synthesis for images of faces. MIT EECS M.S. thesis (1996). [4] D. Beymer, A. Shashua, and T. Poggio. Example based image analysis and synthesis. AI Memo 1161, MIT (1993). [5] H. Goldstein. Classical Mechanics. Addison-Wesley, London (1980).
1996
2
1,214
A Constructive RBF Network for Writer Adaptation John C. Platt and Nada P. Matic Synaptics, Inc. 2698 Orchard Parkway San Jose, CA 95134 platt@synaptics.com, nada@synaptics.com Abstract This paper discusses a fairly general adaptation algorithm which augments a standard neural network to increase its recognition accuracy for a specific user. The basis for the algorithm is that the output of a neural network is characteristic of the input, even when the output is incorrect. We exploit this characteristic output by using an Output Adaptation Module (OAM) which maps this output into the correct user-dependent confidence vector. The OAM is a simplified Resource Allocating Network which constructs radial basis functions on-line. We applied the OAM to construct a writer-adaptive character recognition system for on-line handprinted characters. The OAM decreases the word error rate on a test set by an average of 45%, while creating only 3 to 25 basis functions for each writer in the test set. 1 Introduction One of the major difficulties in creating any statistical pattern recognition system is that the statistics of the training set is often different from the statistics in actual use. The creation of a statistical pattern recognizer is often considered as a regression problem, where class probabilities are estimated from a fixed training set. Statistical pattern recognizers tend to work well for typical data that is similar to the training set data, but do not work well for atypical data that is not well represented in the training set. Poor performance on atypical data is a problem for human interfaces, because people tend to provide drastically non-typical data (for example, see figure 1). The solution to this difficulty is to create an adaptive recognizer, instead of treating recognition as a static regression problem. The recognizer must adapt to new statistics during use. As applied to on-line handwriting recognition, an adaptive 766 atypical "rIO OAM ~> abc consistent (Incorrect) neural network response J. C. Platt and N. P. MafiC typical lOr" r t - - - - - - - - - - ' , ~'--abc r correct neural network response Figure 1: When given atypical input data, the neural network produces a consistent incorrect output pattern. The OAM recognizes the consistent pattern and produces a corrected user-adaptive output. recognizer improves the accuracy for a particular user by adapting the recognizer to that user. This paper proposes a novel method for creating an adaptive recognizer, which we call the Output Adaptation Module or OAM. The OAM was inspired by the development of a writer-independent neural network handwriting recognizer. We noticed that the output of this neural network was characteristic of the input: if a specific style of character was shown to the network, the network's output was almost always consistent for that specific style, even when the output was incorrect. To exploit the consistency of the incorrect outputs, we decided to add an OAM on top of the network. The OAM learns to recognize these consistent incorrect output vectors, and produces a more correct output vector (see figure 1). The units of the OAM are radial basis functions (RBF) [5]. Adaptation of these RBF units is performed using a simplified version of the Resource Allocating Network (RAN) algorithm of Platt [4][2]. The number of units that RAN allocates scales sub-linearly with the number of presented learning examples, in contrast to other algorithms which allocate a new unit for every learned example. The OAM has the following properties, which are useful for a user-adaptive recognIzer: • The adaptation is very fast: the user need only provide a few additional examples of his own data. • There is very little recognition speed degradation. • A modest amount of additional memory per user is required. • The OAM is not limited to neural network recognizers. • The output of the OAM is a corrected vector of confidences, which is more useful for contextual post-processing than a single label. A Constructive RBF Networkfor Writer Adaptation 767 1.1 Relationship to Previous Work The OAM is related to previous work in user adaptation of neural recognizers for both speech and handwriting. A previous example of user adaptation of a neural handwriting recognizer employed a Time Delay Neural Network (TDNN), where the last layer of a TDNN was replaced with a tunable classifier that is more appropriate for adaptation [1][3]. In Guyon, et al. [1], the last layer of a TDNN was replaced by a k-nearest neighbor classifier. This work was further extended in Matic, et al. [3], where the last layer of the TDNN was replaced with an optimal hyperplane classifier which is retrained for adaptation purposes. The optimal hyperplane classifier retained the same accuracy as the k-nearest neighbor classifier, while reducing the amount of computation and memory required for adaptation. The present work improves upon these previous user-adaptive handwriting systems in three ways. First, the OAM does not require the retraining and storage of the entire last layer of the network. The OAM thus further reduces both CPU and memory requirements. Second, the OAM produces an output vector of confidences, instead of simply an output label. This vector of confidences can be used effectively by a contextual post-processing step, while a label cannot. Third, our adaptation experiments are performed on a neural network which recognizes a full character set. These previous papers only experimented with neural networks that recognized character subsets, which is a less difficult adaptation problem. The OAM is related to stacking [6]. In stacking, outputs of multiple recognizers are combined via training on partitions of the training set. With the OAM, the multiple outputs of a recognizer are combined using memory-based learning. The OAM is trained on the idiosyncratic statistics of actual use, not on a pre-defined training set partition. 2 The Output Adaptation Module (OAM) Section 2 of this paper describes the OAM in detail, while section 3 describes its application to create a user-adaptive handwriting recognizer. The OAM maps the output of a neural network Vi into a user-adapted output Oi, by adding an adaptation vector Ai: Oi = Vi + Ai· (1) Depending on the neural network training algorithm used, both the output of the neural network Vi and the user-adapted output Oi can estimate a posteriori class probabilities, suitable for further post-processing. The goal of the OAM is to bring the output Oi closer to an ideal response Ii . In our experiments, the target Ii is 0.9 for the neuron corresponding to the correct character and 0.1 for all other neurons. The adaptation vector Ai is computed by a radial basis function network that takes Vi as an input: Oi = Vi + Ai Vi + L CijcI>jCV), (2) j cI>j(V) f (d(V;t;») . (3) where Mj is the center ofthe jth radial basis function, d is a distance metric between V and Mj, Rj is a parameter that controls the width of the jth basis function, f is 768 J. C. Platt and N. P Matie o Desired = max loil 10.551 A RBF M j 2: ) INPUT Figure 2: The architecture of the OAM. a decreasing function that controls the shape of the basis functions, and Gij is the amount of correction that the jth basis function adds to the output. We call Mj the memories in the adaptation module, and we call Cj the correction vectors (see figure 2). The function f is a decreasing polynomial function: f(x) = { (1_x2)2, if x < 1; (4) 0, otherwise. The distance function d is a Euclidean distance metric that first clips both of its input vectors to the range [0.1,0.9] in order to reduce spurious noise. The algorithm for constructing the radial basis functions is a simplification of the RAN algorithm [4][2]. The OAM starts with no memories or corrections. When the user corrects a recognition error, the OAM finds the distance dmin from the nearest memory to the vector Vi. If the distance dmin is greater than a threshold 6, then a new RBF unit is allocated, with a new memory that is set to the vector Vi, and a corresponding correction vector that is set to correct the error with a step size a: Gik = a(Ti - Oi). (5) If the distance dmin is less than 6, no new unit is allocated: the correction vector of the nearest memory to Vi is updated to correct the error by a step size b. (6) For our experiments, we set 6 = 0.1, a = 0.25, and b = 0.2. The values a and bare chosen to be less than 1 to sacrifice learning speed to gain learning stability. The number of radial basis functions grows sub-linearly with the number of errors, because units are only allocated for novel errors that the OAM has not seen before. A Constructive RBF Network/or Writer Adaptation 769 For errors similar to those the OAM has seen before, the algorithm updates one of the correction vectors using a simplified LMS rule (eq. 6). In the computation of the nearest memory, we always consider an additional phantom memory: the target Q that corresponds to the highest output in V. This phantom memory is considered in order to prevent the OAM from allocating memories when the neural network output is unambiguous. The phantom memory prevents the OAM from affecting the output for neatly written characters. The adaptation algorithm used is described as pseudo-code, below: For every character shown to the network { } If the user indicates an error { } f = target vector of the correct character Q = target vector of the highest element in V dmin = min(minj d(V, Mj ), d(V, Q)) If dmin > 0 { / / allocate a new memory k = index of the neW memory Cik = a(1i - Od Mk = V } Rj = dmin else if memories exist and minj d(V, M}) < d(V, Q) { k = arg minj d(V, Mj) Cik = Cik + b(1i - Oi)cf>k(V) } 3 Experiments and Results To test the effectiveness of the OAM, we used it to create a writer-adaptive handwriting recognition system. The OAM was connected to the outputs of a writerindependent neural network trained to recognize characters hand-printed in boxes. This neural network was a carefully tuned multilayer feed-forward network, trained with the back-propagation algorithm. The network has 510 inputs, 200 hidden units, and 72 outputs. The input to the OAM was a vector of 72 confidences, one per class. These confidences were in the range [0,1]. There is one input for every upper case character, lower case character, and digit. There is also one input for each member of a subset of punctuation characters ( !$&',-:;=? ). The OAM was tested interactively. Tests of the OAM were performed by five writers disjoint from the training writers for the writer-independent neural network. These writers had atypical writing styles which were difficult for the network to recognize. Test characters were entered word-by-word on a tablet. The writers were instructed to write more examples of characters that reflected their atypical writing style. The words that these writers used were not taken from a word list, and could consist of any combination of the 72 available characters. Users were shown the results of the OAM, combined into words and further processed by a dictionary. Whenever the system failed to recognize a word correctly, all misclassified characters and their corresponding desired labels were used by the OAM to adapt the system to the 770 ... o 1: GI " ~20 ~ 1 a; '3 10 E :::J U writer nm 20 writertw J. C. Platt and N. P. Matic no adaptation adaptation 30 o 50 word no. no adaptation adaptation 50 word no. Figure 3: The cumulative number of word errors for the writer "nm" and the writer "tw" , with and without adaptation Writer % word error % word error Memories stored Words written no OAM OAM during test during test rag 25% 17% 6 93 mfe 62% 36% 12 53 qm 42% 31% 25 80 nm 54% 21% 4 39 tw 28';70 10'10 3 60 Table 1: Quantitative test results for the OAM. particular user. Figure 3 shows the performance of the OAM for the writer "nm" and for the writer "tw". The total number of word errors since adaptation started is plotted against the number of words shown to the OAM. The baseline cumulative error without the OAM is also given. The slope of each curve gives the estimate of the instantaneous error rate. The OAM causes the slope for both writers to decrease dramatically as the adaptation progresses. Over the last third of the test set, the word error rate for writer "nm" was 0%, while the word error rate for writer "tw" was 5%. These examples show the the OAM can substantially improve the accuracy of a writer-independent neural network. Quantitative results are shown in table 1, where the word error rates obtained with the OAM are compared to the baseline word error rates without the OAM. The right two columns contain the number of stored basis functions and the number of words tested by the OAM. The OAM corrects an average of 45% of the errors in the test set. The accuracy A Constructive RBF Network/or Writer Adaptation 771 rates with the OAM were taken for the entire test run, and therefore count the errors that were made while adaptation was taking place. By the end of the test, the true error rates for these writers would be even lower than those shown in table 1, as can be seen in figure 3. These experiments showed that the OAM adapts very quickly and requires a small amount of additional memory and computation. For most writers, only 2-3 presentations of a writer-dependent variant of a character were sufficient for the OAM to adapt. The maximum number of stored basis functions were 25 in these experiments. The OAM did not substantially affect the recognition speed of the system. 4 Conclusions We have designed a widely applicable Output Adaptation Module (OAM) to place on top of standard neural networks. The OAM takes the output of the network as input, and determines an additional adaptation vector to add to the output. The adaptation vector is computed via a radial basis function network, which is learned with a simplification of the RAN algorithm. The OAM has many nice properties: only a few examples are needed to learn atypical inputs, the number of stored memories grows sub-linearly with the number of errors, the recognition rate of the writer-independent neural network is unaffected by adaptation and the output of the module is a confidence vector suitable for further post-processing. The OAM addresses the difficult problem of creating a high-perplexity adaptive recognizer. We applied the OAM to create a writer-adaptive handwriting recognition system. On a test set of five difficult writers, the adaptation module decreased the error rate by 45%, and only stored between 3 and 25 basis functions per writer. 5 Acknowledgements We wish to thank Steve Nowlan for helpful suggestions during the development of the OAM algorithm and for his work on the writer-independent neural network. We would also like to thank Joe Decker for his work on the writer-independent neural network. References [1] 1. Guyon, D. Henderson, P. Albrecht, Y. Le Cun, and J. Denker. Writer independent and writer adaptive neural network for on-line character recognition. In S. Impedovo, editor, From Pixels to Features III, Amsterdam, 1992. Elsevier. [2] V. Kadirkamanathan and M. Niranjan. A function estimation approach to sequential learning with neural networks. Neural Computation, 5(6):954- 976, 1993. [3] N. Matic, 1. Guyon, J. Denker, and V. Vapnik. Writer adaptation for on-line handwritten character recognition. In ICDAR93, Tokyo, 1993. IEEE Computer Society Press. [4] J. Platt. A resource-allocating network for function interpolation. Neural Computation, 3(2):213-225, 1991. [5] M. Powell. Radial basis functions for multivariate interpolation: A review. In J. C. Mason and M. G. Cox, editors, Algorithms for Approximation, Oxford, 1987. Clarendon Press. [6] D. Wolpert. Stacked generalization. Neural Networks, 5(2):241-260, 1992.
1996
20
1,215
MLP can provably generalise much better than VC-bounds indicate. A. Kowalczyk and H. Ferra Telstra Research Laboratories 770 Blackburn Road, Clayton, Vic. 3168, Australia ({ a.kowalczyk, h.ferra}@trl.oz.au) Abstract Results of a study of the worst case learning curves for a particular class of probability distribution on input space to MLP with hard threshold hidden units are presented. It is shown in particular, that in the thermodynamic limit for scaling by the number of connections to the first hidden layer, although the true learning curve behaves as ~ a-I for a ~ 1, its VC-dimension based bound is trivial (= 1) and its VC-entropy bound is trivial for a ::; 6.2. It is also shown that bounds following the true learning curve can be derived from a formalism based on the density of error patterns. 1 Introduction The VC-formalism and its extensions link the generalisation capabilities of a binary valued neural network with its counting function l , e.g. via upper bounds implied by VC-dimension or VC-entropy on this function [17, 18]. For linear perceptrons the counting function is constant for almost every selection of a fixed number of input samples [2], and essentially equal to its upper bound determined by VC-dimension and Sauer's Lemma. However, in the case for multilayer perceptrons (MLP) the counting function depends essentially on the selected input samples. For instance, it has been shown recently that for MLP with sigmoidal units although the largest number of input samples which can be shattered, Le. VC-dimension, equals O(w2 ) [6], there is always a non-zero probability of finding a (2w + 2)-element input sample which cannot be shattered, where w is the number of weights in the network [16]. In the case of MLP using Heaviside rather than sigmoidal activations (McCullochPitts neurons), a similar claim can be made: VC-dimension is O(wl1og21lt} [13, 15], 1 Known also as the partition function in computational learning theory. MLP Can Provably Generalize Much Better than VC-bounds Indicate 191 where WI is the number of weights to the first hidden layer of 11.1 units, but there is a non-zero probability of finding a sample of size WI + 2 which cannot be shattered [7, 8]. The results on these "hard to shatter samples" for the two MLP types differ significantly in terms of techniques used for derivation. For the sigmoidal case the result is "existential" (based on recent advances in "model theory") while in the Heaviside case the proofs are constructive, defining a class of probability distributions from which "hard to shatter" samples can be drawn randomly; the results in this case are also more explicit in that a form for the counting function may be given [7, 8]. Can the existence of such hard to shatter samples be essential for generalisation capabilities of MLP? Can they be an essential factor for improvement of theoretical models of generalisation? In this paper we show that at least for the McCullochPitts case with specific (continuous) probability distributions on the input space the answer is "yes". We estimate "directly" the real learning curve in this case and show that its bounds based on VC-dimension or VC-entropy are loose at low learning sample regimes (for training samples having less than 12 x WI examples) even for the linear perceptron. We also show that a modification to the VC-formalism given in [9, 10] provides a significantly better bound. This latter part is a more rigorous and formal extension and re-interpretation of some results in [11, 12]. All the results are presented in the thermodynamic limit, i.e. for MLP with WI ~ 00 and training sample size increasing proportionally, which simplifies their mathematical form. 2 Overview of the formalism On a sample space X we consider a class H of binary functions h : X ~ {a, 1} which we shall call a hypothesis space. Further we assume that there are given a probability distribution jJ on X and a target concept t : X ~ {a, 1}. The quadruple C = (X, jJ, H, t) will be called a learning system. In the usual way, with each hypothesis h E H we associate the generalization error fh d~ Ex [It(x) - h(x)l] and the training error fh,x d~ ~ L:~l It(Xi) - h(xdl for any training m-sample x = (Xl, ... ,xm) E xm. Given a learning threshold ° ~ ,X ~ 1, let us introduce an auxiliary random variable f~ax(X) d~ max{fh ; h E H & fh,x ~ ,X} for x E xm, giving the worst generalization error of all hypotheses with training error ~ ,X on the m-sample x E xm. 2 The basic objects of intE'rest in this paper are the learning cUnJe3 defined as tL>C( ) d!l E [ max ( ... )] f). m Xm f). X. 2.1 Thermodynamic limit Now we introduce the thermodynamic limit of the learning curve. The underlying idea of such asymptotic analysis is to capture the essential features of learning 2In this paper max(S), where S C R, denotes the maximal element in the closure of S, or 00 if no such element exists. Similarly, we understand mineS). 3Note that our learning curve is determined by the worst generalisation error of acceptable hypotheses and in this respect differs from "average generalisation error" learning curves considered elsewhere, e.g. [3, 5]. 192 A. Kowalczyk and H. Ferra systems of very large size. Mathematically it turns out that in the thermodynamic limit the functional forms of learning curves simplify significantly and analytic characterizations of these are possible. We are given a sequence of learning systems, or shortly, LN = (XN,/J.N,HN,tN)' N = 1,2, ... and a scaling N f-7 TN E R+, with the property TN ~ 00; the scaling can be thought of as a measure of the size (complexity) of a learning system, e.g. VC-dimension of HN. The thermodynamic limit of scaled learning curves is defined for a > ° as follows 4 we ( ) de! l' we (L J) €AOO a = 1m sup €A,N aTN , N--+oo (1) Here, and below, the additional subscript N refers to the N-th learning system. 2.2 Error pattern density formalism This subsection briefly presents a thermodynamic version of a modified VC formalism discussed previously in [9J; more details and proofs can be found in [1OJ. The main innovation of this approach comes from splitting error patterns into error shells and using estimates on the size of these error shells rather than the total number of error patterns. We shall see on examples discussed in the following section that this improves results significantly. The space {O, l}m of all binary m-vectors naturally splits into m + 1 error pattern shells Ern, i = 0,1, ... , m, with the i-th shell composed of all vectors with exactly i entries equal to 1 . For each h E Hand i = (Xl, ... ,Xm ) E X m , let vh(i) E {O,l}m denote a vector (error pattern) having 1 in the j-th position if and only if h(xj) :j:. t(Xj). As the i-th error shell has en elements, the average error pattern density falling into this error shell is (i = 0,1, ... ,m), (2) where # denotes the cardinality of a set5 . Theorem 1 Given a sequence of learning systems LN = (XN, /J.N, HN, tN), a scaling TN and a function 'P : R+ X (0, 1) ~ R+ such that In (dfN) ~ -TN'P (m ,i) + O(TN), , TN m for all m,N = 1,2, ... , ° ~ i ~ m. Then (3) (4) 4We recall that lxJ denotes the largest integer $ x and limsuPN-+oo XN is defined as limN-+oo of the monotonic sequence N 1--+ max{xl' X2, ••• , XN}' Note that in contrast to the ordinary limit, lim sup always exists. 5Note the difference to the concept of error shells used in [4] which are partitions of the finite hypothesis space H according to the generalisation error values. Both formalisms are related though, and the central result in [4], Theorem 4, can be derived from our Theorem 1 below. MLP Can Provably Generalize Much Better than VC-bounds Indicate 193 for any ° :s A :s 1 and a, /3 > 0, where de! { ( y + /3X ) } f>.,6(a) = max f E (0,1) ; 3 0::;y::;>. a(1i(y) + /31i(x)) - rp a + a/3, 1 /3 ~ ° t::;x::;1 + and 1i(y) d~ -y In y - (1 - y) In(l - y) denotes the entropy function. 3 Main results : applications of the formalism 3.1 VC-bounds We consider a learning sequence L N = (X N , J.L N , H N , t N), t N E H N (realisable case) and the scaling of this sequence by VC-dimension [17], i.e. we assume TN = dvc(HN) -+ 00. The following bounds for the N-th learning system can be derived for A = ° (consistent learning case) [1, 17]: fwe (m) < O,N . 2-mt/2 em ; .1 ( (2) dVc(HN») ° mm 1,2 dvc(HN) df. In the thermodynamic limit, i.e. as N -+ 00, we get for any a > lie fg'~(a) . (1 210g2(2ea)) < mIn , , a Note that this bound is independent of probability distributions J.LN. 3.2 Piecewise constant functions (5) (6) Let PC(d) denote the class of piecewise constant binary functions on the unit segment [0,1) with up to d ~ ° discontinuities and with their values defined as 1 at all these discontinuity points. We consider here the learning sequence LN = ([0,1), J.LN, PC(dN), tN) where J.LN is any continuous probability distributions on [0,1), dN is a monotonic sequence of positive integers diverging to 00 and targets tN E PC(dt N ) are such that the limit c5t d~ limN-+oo .!!.I:Ldd exists. (Without loss of tN generality we can assume that all J.LN are the uniform distribution on [0,1).) For this learning sequence the following can be established. Claim 1. The following function defined for a > 1 and ° :s x :s 1 as rp(a, x) d~ -a(l-x)1i (2;(t~X») -ax1i (~~~) +a1i(x) for2ax(1-x) > 1, and as 0, otherwise, satisfies assumption (3) with respect to the scaling TN d;! dN. Claim 2. The following two sided bound on the learning curve holds: (7) 194 A. Kowalczyk and H. Ferra We outline the main steps of proof of these two claims now. For Claim 1 we start with a combinatorial argument establishing that in the particular case of constant target { ( ':1'-1) -1 "'~N /2 (m-:-i-l) (i~l) for d + dt < min(2i, 2(m - i)), tl,"!lN = ,-1 L...J)=O )-1 ) " 1 otherwise. Next we observe that that the above sum equals This easily gives Claim 1 for constant target (tSt = 0). Now we observe that this particular case gives an upper bound for the general case (of non-constant target) if we use the "effective" number of discontinuities dN + dt N instead of dN. For Claim 2 we start with the estimate [12, 11] derived from the Mauldon result [14] for the constant target tN = canst, m ~ dN. This implies immediately the expression €o~(a) = -..!.. (1 + In(2a)) . 2a (8) for the constant target, which extends to the estimate (7) with a straightforward lower and upper bound on the "effective" number of discontinuities in the case of a non-constant target. 3.3 Link to multilayer perceptron Let MLpn(wd denote the class offunction from R n to {O, I} which can be implemented by a multilayer perceptron (feedforward neural network) with ~ 1 number of hidden layers, with Wt connections to the first hidden layer and the first hidden layer composed entirely of fully connected, linear threshold logic units (i.e. units able to implement any mapping of the form (Xl, .. , Xn) f-t O(ao + L~l aixi) for ai E R). It can be shown from the properties of Vandermonde determinant (c.f. [7, 8]) that if 1 : [0,1) -+ R n is a mapping with coordinates composed of linearly independent polynomials (generic situation) of degree::; n, then (9) This implies immediately that all results for learning the class of PC functions in Section 5.2 are applicable (with obvious modifications) to this class of multilayer perceptrons with probability distribution concentrated on the I-dimensional curves of the form 1([0,1)) with 1 as above. However, we can go a step further. We can extend such a distribution to a continuous distribution on R n with support "sufficiently close" to the curve 1([0,1)), MLP Can Provably Generalize Much Better than VC-bounds Indicate 1.0 ,...... ..!:!. .... 0.8 0 t:: CD 0.6 c: 0 :; 0.4 . ~ Iii .... CD 0.2 c: CD \.. .... \.. .... " .. ..... "-: .... .... ''':2: :'C: 2:'.:: ::CC:.: .. ......... ~. " ........................... ~."".csc . C) 0.0 0.0 10.0 20.0 scaled training sample size (a) 195 VC Entropy EPD,0,=0.2 EPD,o,= O.O TC+, 0,=0.2 TCO,o,=O.O TC-, 0, = 0.2 Figure 1: Plots of different estimates for thermodynamic limit of learning curves for the sequence of multilayer perceptrons as in Claim 3 for consistent learning (A = 0). Estimates on true learning curve from (7) are for 8t = 0 ('TCO') and 8t = 0.2 ('TC+' and 'TC-' for the upper and lower bound, respectively). Two upper bounds of the form (4) from the modified VC-formalism for r.p as in Claim 1 and f3 = 1 are plotted for 8t = 0.0 and 8t = 0.2 (marked EPD). For comparison, we plot also the bound (10) based on the VC-entropy; VC bound (5) being trivial for this scaling, = 1, c.f. Corollary 2, is not shown. with changes to the error pattern densities /)/fN, the learning curves, etc., as small as desired. This observa.tion implies the follo~ing result: Claim 3 For any sequence of multilayer perceptrons, M LpnN (WIN), WIN ~ 00, there exists a sequence of continuous probability distributions J.1.N on R nN with properties as follows. For any sequence of targets tN E M LpnN (WltN)' both Claim 1 and Claim 2 of Section 3.2 hold for the learning sequence (RnN, J.1.N, M LpnN (WIN), tN) with scaling TN d~ nIN and 8t = limN-+oo WltN IWIN. In particular bound (4) on the learning curve holds for r.p as in Claim 1. Corollary 2 If additionally the number of units in first hidden layer 1llN ~ 00, then the thermodynamic limit of VC-bound (5) with respect to the scaling TN = WIN is trivial, i.e. = 1 for all a > O. Proof. The bound (5) is trivial for m ~ 12dN, where dN d~ dvc(M LpnN (WltN )). As dN = O(WIN IOg2(1lIN)) [13, 15] for any continuous probability on the input space, this bound is trivial for any a = ~ < 12..4H... ~ 00 if N ~ 00. 0 tt'lN WIN There is a possibility that VC dimension based bounds are applicable but fail to capture the true behavior because of their independence from the distribution. One option to remedy the situation is to try a distribution-specific estimate such as VC entropy (i.e. the expectation of the logarithm of the counting function IIN(XI, ... , xm) which is the number of dichotomies realised by the perceptron for the m-tuple of input points [18]). However, in our case, lIN (Xl , ... , xm) has the lower bound 2 ",min(wlN/2 m-l) (m) £, • l' . h' h' . all h L."i=O ' i ' or Xl, ... , Xm m genera pOSItIOn, w IC IS VIrtu y t e expression from Sauer's lemma with VC-dimension replaced by WIN 12. Thus using 196 A. Kowalczyk and H. Ferra VC entropy instead of VC dimension (and Sauer's Lemma) we cannot hope for a better result than bounds of the form (5) with WlN 12 replacing VC-dimension resulting in the bound (0 > lie) (10) in the thermodynamic limit with respect to the scaling TN = WlN. (Note that more "optimistic" VC entropy based bounds can be obtained if prior distribution on hypothesis space is given and taken into account [3].) The plots of learning curves are shown in Figure l. Acknowledgement. The permission of Director of Telstra Research Laboratories to publish this paper is gratefully acknowledged. References [1) A. Blumer, A. Ehrenfeucht, D. Haussler, and M.K. Warmuth. Learnability and the Vapnik-Chervonenkis dimensions. Journal of the ACM, 36:929-965, (Oct. 1989). [2) T.M. Cover. Geometrical and statistical properties of linear inequalities with applications to pattern recognition. IEEE Trans. Elec. Comp., EC-14:326-334, 1965. [3) D. Hausler, M. Kearns, and R. Shapire. Bounds on the Sample Complexity of Bayesian Learning Using Information Theory and VC Dimension. Machine Learning, 14:83113, (1994). [4) D. Haussler, M. Kearns, H.S. Seung, and N. Tishby. Rigorous learning curve bounds from statistical mechanics. In Proc. COLT'94, pages 76-87, 1994. [5) S.B. Holden and M. Niranjan. On the Practical Applicability of VC Dimension Bounds. Neural Computation, 1:1265-1288, 1995). [6) P. Koiran and E.D. Sontag. Neural networks with quadratic VC-dimension. In Proc. NIPS 8, pages 197-203, The MIT Press, Cambridge, Ma., 1996 .. [7) A. Kowalczyk. Counting function theorem for multi-layer networks. In Proc. NIPS 6, pages 375-382. Morgan Kaufman Publishers, Inc., 1994. [8) A. Kowalczyk. Estimates of storage capacity of multi-layer perceptron with threshold logic hidden units. Neural networks, to appear. [9) A. Kowalczyk and H. Ferra. Generalisation in feedforward networks. Proc. NIPS 6, pages 215-222, The MIT Press, Cambridge, Ma., 1994. [10) A. Kowalczyk. An asymptotic version of EPD-bounds on generalisation in learning systems. 1996. Preprint. [11) A. Kowalczyk, J. Szymanski, and R.C. Williamson. Learning curves from a modified VC-formalism: a case study. In Proc. of 1CNN'95 , 2939-2943, IEEE, 1995. [12) A. Kowalczyk, J. Szymanski, P.L. Bartlett, and R.C. Williamson. Examples of learning curves from a modified VC-formalism. Proc. NIPS 8, pages 344- 350, The MIT Press, 1996. [13) W. Maas. Neural Nets with superlinear VC-dimesnion. Neural Computation, 6:877884, 1994. [14) J.G. Mauldon. Random division of an interval. Proc. Cambridge Phil. Soc., 41:331336, 1951. [15) A. Sakurai. Tighter bounds of the VC-dimension of three-layer networks. In Proc. of the 1993 World Congress on Neural Networks, 1993. [16) E. Sontag. Shattering all sets of k points in "general position" requires (k - 1)/2 parameters. Report 96-01, Rutgers Center for Systems and Control, 1996. [17) V. Vapnik. Estimation of Dependences Based on Empirical Data. Springer-Verlag, 1982. [18) V. Vapnik. The Nature of Statistical Learning Theory. Springer-Verlag, 1995.
1996
21
1,216
3D Object Recognition: A Model of View-Tuned Neurons Emanuela Bricolo Tomaso Poggio Department of Brain and Cognitive Sciences Massachusetts Institute of Technology Cambridge, MA 02139 {emanuela,tp}Gai.mit.edu Nikos Logothetis Baylor College of Medicine Houston, TX 77030 nikosGbcmvision.bcm.tmc.edu Abstract In 1990 Poggio and Edelman proposed a view-based model of object recognition that accounts for several psychophysical properties of certain recognition tasks. The model predicted the existence of view-tuned and view-invariant units, that were later found by Logothetis et al. (Logothetis et al., 1995) in IT cortex of monkeys trained with views of specific paperclip objects. The model, however, does not specify the inputs to the view-tuned units and their internal organization. In this paper we propose a model of these view-tuned units that is consistent with physiological data from single cell responses. 1 INTRODUCTION Recognition of specific objects, such as recognition of a particular face, can be based on representations that are object centered, such as 3D structural models. Alternatively, a 3D object may be represented for the purpose of recognition in terms of a set of views. This latter class of models is biologically attractive because model acquisition - the learning phase - is simpler and more natural. A simple model for this strategy of object recognition was proposed by Poggio and Edelman (Poggio and Edelman, 1990). They showed that, with few views of an object used as training examples, a classification network, such as a Gaussian radial basis function network, can learn to recognize novel views of that object, in partic42 E. Bricolo, T. Poggio and N. Logothetis (a) (b) View angle Figure 1: (a) Schematic representation of the architecture of the Poggio-Edelman model. The shaded circles correspond to the view-tuned units, each tuned to a view of the object, while the open circle correspond to the view-invariant, object specific output unit. (b) Tuning curves ofthe view-tuned (gray) and view-invariant (black) units. ular views obtained by in depth rotation of the object (translation, rotation in the image plane and scale are probably taken care by object independent mechanisms). The model, sketched in Figure 1, makes several prediction about limited generalization from a single training view and about characteristic generalization patterns between two or more training views (Biilthoff and Edelman, 1992). Psychophysical and neurophysiological results support the main features and predictions of this simple model. For instance, in the case of novel objects, it has been shown that when subjects -both humans and monkeys- are asked to learn an object from a single unoccluded view, their performance decays as they are tested on views farther away from the learned one (Biilthoff and Edelman, 1992; Tarr and Pinker, 1991; Logothetis et al., 1994). Additional work has shown that even when 3D information is provided during training and testing, subjects recognize in a view dependent way and cannot generalize beyond 40 degrees from a single training view (Sinha and Poggio, 1994). Even more significantly, recent recordings in inferotemporal cortex (IT) of monkeys performing a similar recognition task with paperclip and amoeba-like objects, revealed cells tuned to specific views of the learned object (Logothetis et al., 1995). The tuning, an example of which is shown in Figure 3, was presumably acquired as an effect of the training to views of the particular object. Thus an object can be thought as represented by a set of cells tuned to several of its views, consistently with finding of others (Wachsmuth et al., 1994). This simple model can be extended to deal with symmetric objects (Poggio and Vetter, 1992) as well as objects which are members of a nice class (Vetter et al., 1995): in both cases generalization from a single view may be significantly greater than for objects such as the paperclips used in the psychophysical and physiological experiments. The original model of Poggio and Edelman has a major weakness: it does not specify which features are inputs to the view-tuned units and what is stored as a representation of a view in each unit. The simulation data they presented employed features such as the x,y coordinates of the object vertices in the image plane or the angles between successive segments. This representation, however, is biologically implausible and specific for objects that have easily detectable vertices and measurable angles, like paperclips. In this paper, we suggest a view representation which is more biologically plausible and applies to a wider variety of cases. We will also show that this extension of the Poggio-Edelman model leads to properties that are 3D Object Recognition: A Model o/View-Tuned Neurons Filters Bank Feature units (m=4) Filtered Images Normalized Output 43 Figure 2: Model overview: during the training phase the images are first filtered through a bank of steerable filters. Then a number of image locations are chosen by an attentional mechanism and the vector of filtered values at these locations is stored in the feature units. consistent with the cell response to the same objects. 2 A MODEL OF VIEW-TUNED UNITS Our approach consists in representing a view in terms of a few local features, which can be regarded as local configurations of grey-levels. Suppose one point in the image of the object is chosen. A feature vector is computed by filtering the image with a set of filters with small support centered at the chosen location. The vector of filter responses serves as a description of the local pattern in the image. Four such points were chosen, for example, in the image of Figure 3a, where the white squares indicate the support of the bank of filters that were used. Since the support is local but finite, the value of each filter depends on the pattern contained in the support and not only on the center pixel; since there are several filters one expects that the vector of values may uniquely represent the local feature, for instance a corner of the paperclip. We used filters that are somewhat similar to oriented receptive fields in VI (though it is far from being clear whether some VI cells behave as linear filters). The ten filters we used are the same steerable filters (Freeman and Adelson, 1991) suggested by Ballard and Rao (Rao and Ballard, 1995; Leung et al., 1995). The filters were chosen to be a basis of steerable 2-dimensional filters up to the third order. If Gn represents the nth derivative of a Gaussian in the x direction and we define the rotation operator ( ... )9 as the operator that rotates a function through an angle 0 about the origin, the ten basis filters are: G9k n = 0, 1,2,3 n O,,=O, ... ,k'rr/(n+l),k=I, ... n (1) 44 E. Bricolo, T. Poggio and N. Logothetis Therefore for each one of the chosen locations m in the image we have a lO-value array Tm given by the output of the filters bank. The representation of a given view of an object is then the following. First m = 1, ... , M locations are chosen, then for each of these M locations the lO-valued vectors Tm are computed and stored. These M vectors, with M typically between 1 and 4, form the representation of the view which is learned and commited to memory. How are the locations chosen? Precise location is not critical. Feature locations can be chosen almost randomly. Of course each specific choice will influence properties of the unit but precise location does not affect the qualitative properties of the model, as verified in simulation experiments. Intuitively, features should be centered at salient locations in the object where there are large changes in contrast and curvature. We have implemented (Riesenhuber and Bricolo, in preparation) a simple attentional mechanism that chooses locations close to edges with various orientations I . The locations shown in Figures 3 and 4 were obtained with this unsupervised technique. We emphasize however that all the results and conclusions of this paper do not depend on the specific location of the feature or the precise procedure used to choose them. We have described so far the learning phase and how views are represented and stored. When a new image V is presented, recognition takes place in the following way. First the new image is filtered through the bank of filters. Thus at each pixel location i we have the vector of values fi provided by the filters. Now consider the first stored vector T I. The closest ft is found searching over all i locations and the distance DI = IITI - ftll is computed. This process is repeated for the other feature vectors Tm for m = 2, ... , M. Thus for the new image V, M distances Dm are computed; the distance Dm is therefore the distance to the stored feature T m of the closest image vector searched over the whole image. The model uses these M distances as exponents in M Gaussian units. The output of the system is a weighted average of the output of these units with an output non linearity of the sigmoidal type: ( M D:l) YV=h LCme-~ m=l (3) In the simulations presented in this paper we estimated (J' from the distribution of distances over several images; the Cm are Cm = M-I, since we have only one training view; h is hex) = 1/(1 - e- X ). In Figure 3b we see the result obtained by simply combining linearly the output of the four feature detectors followed by the sigmoidal non linearity (Figure 4a shows another example). We have also experimented with a multiplicative combination of the output of the feature units. In this case the system performs an AND of the M features. If the response to the distractors is used to set a threshold for 1 A saliency map is at first constructed as the average of the convolutions of the image with four directional filters (first order steer able filters with 0 = 0, ... , k7r /( 4), k = 1, .. .4). The locations with higher saliency are extracted one at the time. After each selection, a region around the selected position is inhibited to avoid selecting the same feature over again. 3D Object Recognition: A Model of VIeW-Tuned Neurons 45 111111 N C ~ Q) 20 E e20 (a) Cl L c ·c 10 ;;:: 10 L fii O~ Q) ~ 0 0 o 20 40 60 1 M ~0.8 0.8 0 'S 0.6 (b) 00.6 D G) 0.4 E "8 0.4 L ~0 .2 0.2 0 0 Figure 3: Comparison between a model view-tuned unit and cortical neuron tuned to a view of the same object. (a) Mean spike rate of an inferotemporal cortex cell recorded in response to views of a specific paperclip Oeft] and to a set of 60 distractor paperclip objects [right]'(Logothetis and Pauls, personal communication). (b) Model response for the same set of objects. This is representative for other cells we have simulated, thought there is considerable variability in the cells (and the model) tuning. classification, then the two versions of the system behave in a similar way. Similar results (not shown) were obtained using other kinds of objects. 3 RESULTS 3.1 COMPARISON BETWEEN VIEW-TUNED UNITS AND CORTICAL NEURONS Electrophysiological investigations in alert monkeys, trained to recognize wire-like objects presented from any view show that the discharge rate of many IT neurons is a bell-shaped function of orientation centered on a preferred view (Logothetis et aI., 1995). The properties of the units here described are comparable to those of the cortical neurons (see Figure 3). The model was tested with the exactly the same objects used in the physiological experiments. As training view for the model we used the view preferred by the cell (the cell became tuned presumably as an effect of training during which the monkey was shown in this particular case several views of this object). 3.2 OCCLUSION EXPERIMENTS What physiological experiments could lend additional support to our model? A natural question concerns the behavior of the cells when various parts of the object are occluded. The predictions of our model are given in Figure 4 for a specific object and a specific choice of feature units (m = 4) and locations. 46 (a) 60 120 180 240 300 360 Rotation from zero view :t!l ... l..lLi..M dAM o 20 40 60 80 100 Distractor indices (b) 'S 0.8 a. 'S o 0.6 Q) 0.4 "'C o ~ 0.2 o E. Bricolo, T. Poggio and N. Logothetis II (i) (ii) (iii) (iv) (v) Figure 4: (a) Model behavior in response to a learned object in full view (highlighted on the learned image are the positions of the four features) at different rotations and to 120 other paperclip objects (distractors), (b) Response dependence on occluder characteristics: (i) object in full view at learned location, (ii) object occluded with a small occluder, (iii) occluded region in (ii) presented in isolation, (iv-v) same as (ii-iii) but with a larger occluder. The simulations show that the behavior depends on the position of key features with respect to the occluder itself. Occluding a part of the object can drastically reduce the response to that specific view (Figure 4b(ii-iv» because of interference with more than one feature. But since the occluded region does not completely overlap with the occluded featUres (considering the support of the filters) , the presentation of this region alone does not always evoke a significant response (Figure 4b(iii-v». 4 Discussion Poggio and Edelman model was designed specifically for paperclip objects and did not explicitly specify how to compute the response for any object and image. In this paper we fill this gap and propose a model of these IT cells that become view tuned as an effect of training. The key aspect of the model is that it relies on a few local features (1-4) that are computed and stored during the training phase. Each feature is represented as the set of responses of oriented filters at one location in the image. During recognition the system computes a robust conjunction of the best matches to the stored features. Clearly, the version of the model described here does not exploit information about the geometric configuration of the features. This information is available once the features are detected and can be critical to perform more robust recognition. We have devised a model of how to use the relative position of the features ft in the image. The model can be made translation and scale invariant in a biologically plausible way by using a network of cells with linear receptive fields, similar in spirit to a model proposed for spatial representation in the parietal cortex (Pouget and Sejnowski, 1996). Interestingly enough, this additional information is not needed to account for the selectivity and the generalization properties of the IT cells we 3D Object Recognition: A Model o/VIeW-Tuned Neurons 47 have considered so far. The implication is that IT cells may not be sensitive to the overall configuration of the stimulus but to the presence of moderately complex local features (according to our simulations, the number of necessary local features is greater than one for the most selective neurons, such as the one of Figure 3a). Scrambling the image of the object should therefore preserve the selectivity of the neurons, provided this can be done without affecting the filtering stage. In practice this may be very difficult. Though our model is still far from being a reasonable neuronal model, it can already be used to make useful predictions for physiological experiments which are presently underway. References Biilthoff, H. and Edelman, S. (1992). Psychophisical support for a two-dimensional view interpolation theory of object recognition. Proceedings of the National Academy of Science. USA, 89:60-64. . Freeman, W. and Adelson, E. (1991). The design and use of steerable filters. IEEE transactions on Pattern Analysis and Machine Intelligence, 13(9):891-906. Leung, T., Burl, M., and Perona, P. (1995). Finding faces in cluttered scenes using random labelled graph matching. In Proceedings of the 5th Internatinal Conference on Computer Vision, Cambridge, Ma. Logothetis, N., Pauls, J., Biilthoff, H., and Poggio, T. (1994). View dependent object recognition by monkeys. Current Biology, 4(5):401-414. Logothetis, N., Pauls, J., and Poggio, T . (1995). Shape representation in the inferior temporal cortex of monkeys. Current Biology, 5(5):552- 563. Poggio, T . and Edelman, S. (1990). A network that learns to recognize threedimensional objects. Nature, 343:263-266. Poggio, T. and Vetter, T. (1992). Recognition and structure from one 2d model view: observations on prototypes, object classes and symmetries. Technical Report A.1. Memo No.1347, Massachusetts Institute of Technnology, Cambridge, Ma. Pouget, A. and Sejnowski, T. (1996). Spatial representations in the parietal cortex may use basis functions. In Tesauro, G., Touretzky, D., and Leen, T., editors, Advances in Neural Information Processing Systems, volume 7, pages 157-164. MIT Press. Rao, R. and Ballard, D. (1995). An active vision architecture based on iconic representations. Artificial Intelligence Journal, 78:461-505. Sinha, P. and Poggio, T. (1994). View-based strategies for 3d object recognition. Technical Report A.1. Memo No.1518, Massachusetts Institute of Technnology, Cambridge, Ma. Tarr, M. and Pinker, S. (1991). Orientation-dependent mechanisms in shape recognition: further issues. Psychological Science, 2(3):207- 209. Vetter, T., Hurlbert, A., and Poggio, T. (1995). View-based models of 3d object recognition: Invariance to imaging transformations. Cerebral Cortex, 3(261269). Wachsmuth, E., Oram, M., and Perrett, D. (1994). Recognition of objects and their component parts: Responses of single units in the temporal cortex of the macaque. Cerebral Cortex, 4(5):509-522.
1996
22
1,217
Learning Bayesian belief networks with neural network estimators Stefano Monti* *Intelligent Systems Program University of Pittsburgh 901M CL, Pittsburgh, PA - 15260 smonti~isp.pitt.edu Gregory F. Cooper*'''' "Center for Biomedical Informatics University of Pittsburgh 8084 Forbes Tower, Pittsburgh, PA - 15261 gfc~cbmi.upmc.edu Abstract In this paper we propose a method for learning Bayesian belief networks from data. The method uses artificial neural networks as probability estimators, thus avoiding the need for making prior assumptions on the nature of the probability distributions governing the relationships among the participating variables. This new method has the potential for being applied to domains containing both discrete and continuous variables arbitrarily distributed. We compare the learning performance of this new method with the performance of the method proposed by Cooper and Herskovits in [7]. The experimental results show that, although the learning scheme based on the use of ANN estimators is slower, the learning accuracy of the two methods is comparable. Category: Algorithms and Architectures. 1 Introduction Bayesian belief networks (BBN) are a powerful formalism for representing and reasoning under uncertainty. This representation has a solid theoretical foundation [13], and its practical value is suggested by the rapidly growing number of areas to which it is being applied. BBNs concisely represent the joint probability distribution over a set of random variables, by explicitly identifying the probabilistic dependencies and independencies between these variables. Their clear semantics make BBNs particularly. suitable for being used in tasks such as diagnosis, planning, and control. The task of learning a BBN from data can usually be formulated as a search over the space of network structures, and as the subsequent search for an optimal parametrization of the discovered structure or structures. The task can be further complicated by extending the search to account for hidden variables and for Learning Bayesian Belief Networks with Neural Network Estimators 579 the presence of data points with missing values. Different approaches have been successfully applied to the task of learning probabilistic networks from data [5]. In all these approaches, simplifying assumptions are made to circumvent practical problems in the implementation of the theory. One common assumption that is made is that all variables are discrete, or that all variables are continuous and normally distributed. In this paper, we propose a novel method for learning BBNs from data that makes use of artificial neural networks (ANN) as probability distribution estimators, thus avoiding the need for making prior assumptions on the nature of the probability distribution governing the relationships among the participating variables. The use of ANNs as probability distribution estimators is not new [3], and its application to the task of learning Bayesian belief networks from data has been recently explored in [11]. However, in [11] the ANN estimators were used in the parametrization of the BBN structure only, and cross validation was the method of choice for comparing different network structures. In our approach, the ANN estimators are an essential component of the scoring metric used to search over the BBN structure space. We ran several simulations to compare the performance of this new method with the learning method described in [7]. The results show that, although the learning scheme based on the use of ANN estimators is slower, the learning accuracy of the two methods is comparable. The rest of the paper is organized as follows. In Section 2 we briefly introduce the Bayesian belief network formalism and some basics of hbw to learn BBNs from data. In Section 3, we describe our learning method, and detail the use of artificial neural networks as probability distribution estimators. In Section 4 we present some experimental results comparing the performance of this new method with the one proposed in [7]. We conclude the paper with some suggestions for further research. 2 Background A Bayesian belief network is defined by a triple (G,n,p), where G = (X,E) is a directed acyclic graph with a set of nodes X = {Xl"'" xn} representing domain variables, and with a set of arcs E representing probabilistic dependencies among domain variables; n is the space of possible instantiations of the domain variablesl ; and P is a probability distribution over the instantiations in n. Given a node X EX, we use trx to denote the set of parents of X in X. The essential property of BBNs is summarized by the Markov condition, which asserts that each variable is independent of its non-descendants given its parents. This property allows for the representation of the multivariate joint probability distribution over X in terms of the univariate conditional distributions P( Xi l7ri, 8i ) of each variable Xi given its parents 7ri, with 8i the set of parameters needed to fully characterize the conditional probability. Application of the chain rule, together with the Markov condition, yields the following factorization of the joint probability of any particular instantiation of all n variables: n P(x~, ... , x~) = II P(x~ 17r~., 8i ) . (1) i=l 1 An instantiation w of all n variables in X is an n-uple of values {x~, ... , x~} such that Xi = X: for i = 1 ... n. 580 S. Monti and G. F. Cooper 2.1 Learning Bayesian belief networks The task of learning BBNs involves learning the network structure and learning the parameters of the conditional probability distributions. A well established set of learning methods is based on the definition of a scoring metric measuring the fitness of a network structure to the data, and on the search for high-scoring network structures based on the defined scoring metric [7, 10]. We focus on these methods, and in particular on the definition of Bayesian scoring metrics. In a Bayesian framework, ideally classification and prediction would be performed by taking a weighted average over the inferences of every possible belief network containing the domain variables. Since this approach is in general computationally infeasible, often an attempt has been made to use a high scoring belief network for classification. We will assume this approach in the remainder of this paper. The basic idea ofthe Bayesian approach is to maximize the probability P(Bs I V) = P(Bs, V)j P(V) of a network structure Bs given a database of cases V. Because for all network structures the term P(V) is the same, for the purpose of model selection it suffices to calculate PCBs, V) for all Bs. The Bayesian metrics developed so far all rely on the following assumptions: 1) given a BBN structure, all cases in V are drawn independently from the same distribution (multinomial sample); 2) there are no cases with missing values (complete database); 3) the parameters of the conditional probability distribution of each variable are independent (global parameter independence); and 4) the parameters associated with each instantiation of the parents of a variable are independent (local parameter independence). The application of these assumptions allows for the following factorization of the probability PCBs, V) n PCBs, V) = P(Bs)P(V I Bs) = PCBs) II S(Xi, 71'i, V) , (2) i=l where n is the number of nodes in the network, and each s( Xi, 71'i, V) is a term measuring the contribution of Xi and its parents 71'i to the overall score of the network Bs. The exact form of the terms s( Xi 71'i, V) slightly differs in the Bayesian scoring metrics defined so far, and for lack of space we refer the interested reader to the relevant literature [7, 10]. By looking at Equation (2), it is clear that if we assume a uniform prior distribution over all network structures, the scoring metric is decomposable, in that it is just the product of the S(Xi, 71'i, V) over all Xi times a constant P(Bs). Suppose that two network structures Bs and BSI differ only for the presence or absence of a given arc into Xi. To compare their metrics, it suffices to compute s( Xi, 71'i, V) for both structures, since the other terms are the same. Likewise, if we assume a decomposable prior distribution over network structures, of the form P(Bs) = 11 Pi, as suggested in [10], the scoring metric is still decomposable, since we can include each Pi into the corresponding s( Xi, 71'i, V). Once a scoring metric is defined, a search for a high-scoring network structure can be carried out. This search task (in several forms) has been shown to be NP-hard [4,6]. Various heuristics have been proposed to find network structures with a high score. One such heuristic is known as K2 [7], and it implements a greedy search over the space of network structures. The algorithm assumes a given ordering on the variables. For simplicity, it also assumes that no prior information on the network is available, so the prior probability distribution over the network structures is uniform and can be ignored in comparing network structures. Learning Bayesian Belief Networks with Neural Network Estimators 581 The Bayesian scoring metrics developed so far either assume discrete variables [7, 10], or continuous variables normally distributed [9]. In the next section, we propose a possible generalization which allows for the inclusion of both discrete and continuous variables with arbitrary probability distributions. 3 An ANN-based scoring metric The main idea of this work is to use 'artificial neural networks as probability estimators, to define a decomposable scoring metric for which no informative priors on the class, or classes, of the probability distributions of the participating variables are needed. The first three of the four assumptions described in the previous section are still needed, namely, the assumption of a multinomial sample, the assumption of a complete database, and the assumption of global parameter independence. However, the use of ANN estimators allows for the elimination of the assumption of local parameter independence. In fact, the conditional probabilities corresponding to the different instantiations of the parents of a variable are represented by the same ANN, and they share the same network weights and the same training data. Let us denote with VI == {C1 , .. . , CI- 1} the set of the first I cases in the database, and with x~l) and 7rr) the instantiations of Xi and 7ri in the l-th case respectively. The joint probability P( Bs, V) can be written as: m P(Bs)P(VIBs) = P(Bs) IIp(CIIVI,Bs) 1=1 m n II II (1) (I) P(Bs) P(xi l7ri , VI, Bs). (3) 1=1 i=l If we assume uninformative priors, or decomposable priors on network structures, of the form P(Bs) = rt Pi, the probability PCBs, V) is decomposable. In fact, we can interchange the two products in Equation 3, so as to obtain n m n PCBs, V) = II [Pi II p(x~l) l7rr), VI , Bs)] = II S(Xi, 7ri, V), (4) 1=1 i=l where S(Xi, 7rj, V) is the term between square brackets, and it is only a function of Xi and its parents in the network structure Bs (Pi can be neglected if we assume a uniform prior over the network structures). The computation of Equation 4 corresponds to the application of the prequential method discussed by Dawid [8]. The estimation of each term P( Xi l7ri , VI, Bs) can be done by means of neural network. Several schemes are available for training a neural network to approximate a given probability distribution, or density. Notice that the calculation of each term S(Xi, 7ri, V) can be computationally very expensive. For each node Xi, computing S( Xi, 7ri, V) requires the training of mANNs, where m is the size of the database. To reduce this computational cost, we use the following approximation, which we call the t-invariance approximation: for any I E {I, . .. , m-l}, given the probability P(Xi l7ri, VI, Bs), at least t (1 s t S m -I) new cases are needed in order to alter such probability. That is, for each positive integer h, such that h < t, we assume P(Xi l7rj, VI+h, Bs) = P(Xi l7ri, VI , Bs). Intuitively, this approximation implies the assumption that, given our present belief about the value of each P(Xi l7rj, VI, Bs), at least t new cases are needed to revise this belief. By making this approximation, we achieve a t-fold reduction in the computation needed, since we now need to build and train only mit ANNs for each Xi , instead of the original m. In fact, application 582 s. Monti and G. F. Cooper of the t-invariance approximatioin to the computation of a given S(Xi, 7ri, 'D) yields: Rather than selecting a constant value for t, we can choose to increment t as the size of the training database 'DI increases. This approach seems preferable. When estimating P(Xi l7ri, 'DI, Bs), this estimate will be very sensitive to the addition of new cases when 1 is small, but will become increasingly insensitive to the addition of new cases as 1 grows. A scheme for the incremental updating oft can be summarized in the equation t = rAil, where 1 is the number of cases already seen (i.e., the cardinality of'D/), and 0 < A ~ 1. For example, given a data set of 50 cases, the updating scheme t = rO.511 would require the training of the ANN estimators P(Xi I 7ri,'DI, Bs) for 1= 1,2,3,5,8,12,18,27,41. 4 Evaluation In this section, we describe the experimental evaluation we conducted to test the feasibility of use of the ANN-based scoring metric developed in the previous section. All the experiments are performed on the belief network Alarm, a multiplyconnected network originally developed to model anesthesiology problems that may occur during surgery [2]. It contains 37 nodes/variables and 46 arcs. The variables are all discrete, and take between 2 and 4 distinct values. The database used in the experiments was generated from Alarm, and it is the same database used in [7]. In the experiments, we use a modification of the algorithm K2 [7]. The modified algorithm, which we call ANN-K2, replaces the closed-form scoring metric developed in [7] with the ANN-based scoring metric of Equation (5). The performance of ANNK2 is measured in terms of accuracy of the recovered network structure, by counting the number of edges added and omitted with respect to the Alarm network; and in terms of the accuracy of the learned joint probability distribution, by computing its cross entropy with respect to the joint probability distribution of Alarm. The learning performance of ANN-K2 is also compared with the performance of K2. To train the ANNs, we used the conjugate-gradient search algorithm [12]. Since all the variables in the Alarm network are discrete, the ANN estimators are defined based on the softmax model,with normalized exponential output units, and with cross entropy as cost function. As a regularization technique, we augment the training set so as to induce a uniform conditional probability over the unseen instantiantions of the ANN input. Given the probability P(Xi l7ri, 'DI) to be estimated, and assuming Xi is a k-valued variable, for each instantiation 7r~ that does not appear in the database D I , we generate k new cases, with 7ri instantiated to 7ri, and Xi taking each of its k values. As a result, the neural network estimates P(Xi 17r~, 'DI) to be uniform, with P(Xi I7rL 'D/) = l/k for each of Xi'S values Xn,.··, Xlk. We ran simulations where we varied the size of the training data set (100, 500, 1000, 2000, and 3000 cases), and the value of A in the updating scheme t = rAil described in Section 3. We used the settings A = 0.35, and A = 0.5. For each run, we measured the number of arcs added, the number of arcs omitted, the cross entropy, and the computation time, for each variable in the network. That is, we considered each node, together with its parents, as a simple BBN, and collected the measures of interest for each of these BBNs. Table 1 reports mean and standard deviation of each measure over the 37 variables of Alarm, for both ANN-K2 and K2. The results for ANN-K2 shown in Table 1 correspond to the setting A = 0.5, Learning Bayesian Belief Networks with Neural Network Estimators 583 Uata Algo. arcs + arcs cross entropy time (sees) set m s.d. m s.d. m med s.d. m med s.d. 100 ANN-K2 0.19 0 .40 0.62 0 .86 0.23 .051 0.52 130 88 159 K2 0 .75 1.28 0.22 0.48 0.08 .070 0 .10 0.44 .06 1.48 500 ANN-K2 0.19 0.40 0.22 0 .48 0 .04 .010 0 .11 1077 480 1312 K2 0.22 0.42 0.11 0.31 0 .02 .010 0.02 0.13 .06 0 .22 1000 ANN·K2 0 .24 0.49 0.22 0 .48 0 .05 .005 0 .15 6909 4866 6718 K2 0.11 0.31 0.03 0.16 0.01 .006 0 .01 0.34 .23 0 .46 2000 ANN·K2 0.19 0.40 0.11 0 .31 0.02 .002 0 .06 6458 4155 7864 K2 0 .05 0.23 0 .03 0.16 0.005 .002 0.007 0 .46 .44 0 .65 3000 ANN-K2 0.16 0.37 0.05 0 .23 0 .01 .001 0.017 11155 4672 2136 K2 0 .00 0 .00 0.03 0 .16 0.004 .001 0 .005 1.02 .84 1.11 Table 1: Comparison ofthe performance of ANN-K2 and of K2 in terms of number of arcs wrongly added (+), number of arcs wrongly omitted (-), cross entropy, and computation time. Each column reports the mean m , the median med, and the standard deviation s.d. of the corresponding measure over the 37 nodes/variables of Alarm. The median for the number of arcs added and omitted is always 0, and is not reported. since their difference from the results corresponding to the setting A = 0.35 was not statistically significant. Standard t-tests were performed to assess the significance of the difference between the measures for K2 and the measures for ANN-K2, for each data set cardinality. No technique to correct for multiple-testing was applied. Most measures show no statistically significant difference, either at the 0.05 level or at the 0.01 level (most p values are well above 0.2). In the simulation with 100 cases, both the difference between the mean number of arcs added and the difference between the mean number of arcs omitted are statistically significant (p ~ 0.01). However, these differences cancel out, in that ANN-K2 adds fewer extra arcs than K2, and K2 omits fewer arcs than ANN-K2. This is reflected in the corresponding cross entropies, whose difference is not statistically significant (p = 0.08). In the simulation with 1000 cases, only the difference in the number of arcs omitted is statistically significant (p ~ .03). Finally, in the simulation with 3000 cases, only the difference in the number of arcs added is statistically significant (p ~ .02). K2 misses a single are, and does not add any extra are, and this is the best result to date. By comparison, ANN-K2 omits 2 arcs, and adds 5 extra arcs. For the simulation with 3000 cases, we also computed Wilcoxon rank sum tests. The results were consistent with the t-test results, showing a statistically significant difference only in the number of arcs added. Finally, as it can be noted in Table 1, the difference in computation time is of several order of magnitude, thus making a statistical analysis superfluous. A natural question to ask is how sensitive is the learning procedure to the order of the cases in the training set. Ideally, the procedure would be insensitive to this order. Since we are using ANN estimators, however, which perform a greedy search in the solution space, particular permutations of the training cases might cause the ANN estimators to be more susceptible to getting stuck in local maxima. We performed some preliminary experiments to test the sensitivity of the learning procedure to the order of the cases in the training set. We ran few simulations in which we randomly changed the order of the cases. The recovered structure was identical in all simulations. Morevoer, the difference in cross entropy for different orderings of the cases in the training set showed not to be statistically significant. 5 Conclusions In this paper we presented a novel method for learning BBN s from data based on the use of artificial neural networks as probability distribution estimators. As a prelim584 s. Monti and G. F. Cooper inary evaluation, we have compared the performance of the new algorithm with the performance of K2, a well established learning algorithm for discrete domains, for which extensive empirical evaluation is available [1,7]. With regard to the learning accuracy of the new method, the results are encouraging, being comparable to stateof-the-art results for the chosen domain. The next step is the application of this method to domains where current techniques for learning BBNs from data are not applicable, namely domains with continuous variables not normally distributed, and domains with mixtures of continuous and discrete variables. The main drawback of the new algorithm is its time requirements. However, in this preliminary evaluation, our main concern was the learning accuracy of the algorithm, and little effort was spent in trying to optimize its time requirements. We believe there is ample room for improvement in the time performance of the algorithm. More importantly, the scoring metric of Section 3 provides a general framework for experimenting with different classes of probability estimators. In this paper we used ANN estimators, but more efficient estimators can easily be adopted, especially if we assume the availability of prior information on the class of probability distributions to be used. Acknowledgments This work was funded by grant IRI-9509792 from the National Science Foundation. References [1] C. Aliferis and G. F. Cooper. An evaluation of an algorithm for inductive learning of Bayesian belief networks using simulated data sets. In Proceedings of the 10th Conference of Uncertainty in AI, pages 8-14, San Francisco, California, 1994. [2] I. Beinlich, H. Suermondt, H. Chavez, and G. Cooper. The ALARM monitoring system: A case study with two probabilistic inference techniques for belief networks. In 2nd Conference of AI in Medicine Europe, pages 247- 256, London, England, 1989. [3] C. Bishop. Neural Networks for Pattern Recognition. Oxford University Press, 1995. [4] R. Bouckaert. Properties of learning algorithms for Bayesian belief networks. In Proceedings of the 10th Conference of Uncertainty in AI, pages 102-109, 1994. [5] W. Buntine. A guide to the literature on learning probabilistic networks from data. IEEE Transactions on Knowledge and Data Engineering, 1996. To appear. [6] D. Chickering, D. Geiger, and D. Heckerman. Learning Bayesian networks: search methods and experimental results. Proc. 5th Workshop on AI and Statistics, 1995. [7] G. Cooper and E. Herskovits. A Bayesian method for the induction of probabilistic networks from data. Machine Learning, 9:309-347, 1992. [8] A. Dawid. Present position and potential developments: Some personal views. Statistical theory. The prequential approach. Journal of Royal Statistical Society A, 147:278-292, 1984. [9] D. Geiger and D. Heckerman. Learning Gaussian networks. Technical Report MSRTR-94-10, Microsoft Research, One Microsoft Way, Redmond, WA 98052, 1994. [10] D. Heckerman, D. Geiger, and D. Chickering. Learning Bayesian networks: the combination of knowledge and statistical data. Machine Learning, 1995. [11] R. Hofmann and V. Tresp. Discovering structure in continuous variables using Bayesian networks. In Advances in NIPS 8. MIT Press, 1995. [12] M. Moller. A scaled conjugate gradient algorithm for fast supervised learning. Neural Networks, 6:525-533, 1993. [13] J. Pearl. Probabilistic Reasoning in Intelligent Systems: networks of plausible inference. Morgan Kaufman Publishers, Inc., 1988.
1996
23
1,218
Approximate Solutions to Optimal Stopping Problems John N. Tsitsiklis and Benjamin Van Roy Laboratory for Information and Decision Systems Massachusetts Institute of Technology Cambridge, MA 02139 e-mail: jnt@mit.edu, bvr@mit.edu Abstract We propose and analyze an algorithm that approximates solutions to the problem of optimal stopping in a discounted irreducible aperiodic Markov chain. The scheme involves the use of linear combinations of fixed basis functions to approximate a Q-function. The weights of the linear combination are incrementally updated through an iterative process similar to Q-Iearning, involving simulation of the underlying Markov chain. Due to space limitations, we only provide an overview of a proof of convergence (with probability 1) and bounds on the approximation error. This is the first theoretical result that establishes the soundness of a Q-Iearninglike algorithm when combined with arbitrary linear function approximators to solve a sequential decision problem. Though this paper focuses on the case of finite state spaces, the results extend naturally to continuous and unbounded state spaces, which are addressed in a forthcoming full-length paper. 1 INTRODUCTION Problems of sequential decision-making under uncertainty have been studied extensively using the methodology of dynamic programming [Bertsekas, 1995]. The hallmark of dynamic programming is the use of a value junction, which evaluates expected future reward, as a function of the current state. Serving as a tool for predicting long-term consequences of available options, the value function can be used to generate optimal decisions. A number of algorithms for computing value functions can be found in the dynamic programming literature. These methods compute and store one value per state in a state space. Due to the curse of dimensionality, however, states spaces are typically Approximate Solutions to Optimal Stopping Problems 1083 intractable, and the practical applications of dynamic programming are severely limited. The use of function approximators to "fit" value functions has been a central theme in the field of reinforcement learning. The idea here is to choose a function approximator that has a tractable number of parameters, and to tune the parameters to approximate the value function. The resulting function can then be used to approximate optimal decisions. There are two preconditions to the development an effective approximation. First, we need to choose a function approximator that provides a "good fit" to the value function for some setting of parameter values. In this respect, the choice requires practical experience or theoretical analysis that provides some rough information on the shape of the function to be approximated. Second, we need effective algorithms for tuning the parameters of the function approximator. Watkins (1989) has proposed the Q-Iearning algorithm as a possibility. The original analyses of Watkins (1989) and Watkins and Dayan (1992), the formal analysis of Tsitsiklis (1994), and the related work of Jaakkola, Jordan, and Singh (1994), establish that the algorithm is sound when used in conjunction with exhaustive lookup table representations (i.e., without function approximation). Jaakkola, Singh, and Jordan (1995), Tsitsiklis and Van Roy (1996a), and Gordon (1995), provide a foundation for the use of a rather restrictive class of function approximators with variants of Q-Iearning. Unfortunately, there is no prior theoretical support for the use of Q- Iearning-like algorithms when broader classes of function approximators are employed. In this paper, we propose a variant of Q-Iearning for approximating solutions to optimal stopping problems, and we provide a convergence result that established its soundness. The algorithm approximates a Q-function using a linear combination of arbitrary fixed basis functions. The weights of these basis functions are iteratively updated during the simulation of a Markov chain. Our result serves as a starting point for the analysis of Q-learning-Iike methods when used in conjunction with classes of function approximators that are more general than piecewise constant. In addition, the algorithm we propose is significant in its own right. Optimal stopping problems appear in practical contexts such as financial decision making and sequential analysis in statistics. Like other problems of sequential decision making, optimal stopping problems suffer from the curse of dimensionality, and classical dynamic programming methods are of limited use. The method we propose presents a sound approach to addressing such problems. 2 OPTIMAL STOPPING PROBLEMS We consider a discrete-time, infinite-horizon, Markov chain with a finite state space S = {I, ... , n} and a transition probability matrix P. The Markov chain follows a trajectory Xo, Xl, X2, .•• where the probability that the next state is y given that the current state is X is given by the (x, y)th element of P, and is denoted by Pxy. At each time t E {O, 1,2, ... } the trajectory can be stopped with a terminal reward of G(Xt). If the trajectory is not stopped, a reward of g(Xt) is obtained. The objective is to maximize the expected infinite-horizon discounted reward, given by E [~"tg(Xt} +<>'G(X,}] , where Q: E (0, 1) is a discount factor and T is the time at which the process is stopped. The variable T is defined by a stopping policy, which is given by a sequence 1084 J. N. Tsitsiklis and B. Van Roy of mappings I-'t : st+l I-t {stop, continue}. Each I-'t determines whether or not to terminate, based on xo, . .. , Xt. If the decision is to terminate, then T = t. We define the value function to be a mapping from states to the expected discounted future reward, given that an optimal policy is followed starting at a given state. In particular, the value function J* : S I-t ~ is given by where T is the stopping time given by the policy {I-'d. It is well known that the value function is the unique solution to Bellman's equation: J*(x) = max [G(X)' g(x) + Q L pxyJ*(Y)]. yES Furthermore, there is always an optimal policy that is stationary (Le., of the form {I-'t = 1-'*, "It}) and defined by * () {stop, I-' x = continue, if G(x) ;::: V*(x), otherwise. Following Watkins (1989), we define the Q- function as the function Q* : S I-t ~ given by Q*(x) = g(x) + Q LpxyV*(y). yES It is easy to show that the Q-function uniquely satisfies Q* (x) = g(x) + Q L Pxy max [G(y), Q* (y)] , yES Furthermore, an optimal policy can be defined by *() {stop, I-' x = continue, if G(x) ;::: Q*(x), otherwise. 3 APPROXIMATING THE Q-FUNCTION "Ix E S. (1) Classical computational approaches to solving optimal stopping problems involve computing and storing a value function in a tabular form. The most common way for doing this is through use of an iterative algorithm of the form Jk+l (x) = max [G(X),g(X) + Q LPxyJk(Y)]. yES When the state space is extremely large, as is the typical case, two difficulties arise. The first is that computing and storing one value per state becomes intractable, and the second is that computing the summation on the right hand side becomes intractable. We will present an algorithm, motivated by Watkins' Q-Iearning, that addresses both these issues, allowingJor approximate solution to optimal stopping problems with large state spaces. Approximate Solutions to Optimal Stopping Problems 3.1 LINEAR FUNCTION APPROXIMATORS We consider approximations of Q* using a function of the form K Q(x, r) = L r(k)¢k (x). k=l 1085 Here, r = (r(I), ... ,r(K)) is a parameter vector and each ¢k is a fixed scalar function defined on the state space S. The functions ¢k can be viewed as basis functions (or as vectors of dimension n), while each r(k) can be viewed as the associated weight. To approximate the Q-function, one usually tries to choose the parameter vector r so as to minimize some error metric between the functions Q(., r) and Q*(.). It is convenient to define a vector-valued function ¢ : S t-+ lRK , by letting ¢(x) = (¢l(X), ... ,¢K(X)). With this notation, the approximation can also be written in the form Q(x,r) = (4)r)(x) , where 4> is viewed as a lSI x K matrix whose ith row is equal to ¢(x). 3.2 THE APPROXIMATION ALGORITHM In the approximation scheme we propose, the Markov chain underlying the stopping problem is simulated to produce a single endless trajectory {xtlt = 0,1,2, ... }. The algorithm is initialized with a parameter vector ro, and after each time step, the parameter vector is updated according to rt+l = rt + ,t¢(Xt) (g(Xt) + a max [¢'(xt+drt,G(xt+d] - ¢'(xdrt) , where It is a scalar stepsize. 3.3 CONVERGENCE THEOREM Before stating the convergence theorem, we introduce some notation that will make the exposition more concise. Let 7r(I), ... , 7r(n) denote the steady-state probabilities for the Markov chain. We assume that 7r(x) > 0 for all xES. Let D be an n x n diagonal matrix with diagonal entries 7r(I), . . . , 7r(n). We define a weighted norm II·IID by IIJIID = L 7r(X)P(x). xES We define a "projection matrix" Il that induces a weighted projection onto the subspace X = {4>r IrE lRK} with projection weights equal to the steady-state probablilities. In particular, IlJ = arg!p.in IIJ - JIID. JEX It is easy to show that Il is given by Il = 4>(4)' D4»-l4>' D. We define an operator F : lRn t-+ lRn by F J = 9 + aPmax [4>rt, a] , where the max denotes a componentwise maximization. We have the following theorem that ensures soundness of the algorithm: 1086 J N. Tsitsiklis and B. Van Roy Theorem 1 Let the following conditions hold: (a) The Markov chain has a unique invariant distribution 7r that satisfies 7r' P = 7r', with 7r(x) > 0 for all xES. (b) The matrix «P has full column rank; that is, the ((basis functions" {¢k 1 k = 1, . .. ,K} are linearly independent. (c) The step sizes 'Yt are nonnegative, nonincreasing, and predetermined. Furthermore, they satisfy L~o 'Yt = 00, and L~o 'Yl < 00. We then have: (a) The algorithm converges with probability 1. (b) The limit of convergence r* is the unique solution of the equation IlF(<<pr*) = «Pr*. (c) FUrthermore, r* satisfies lI«pr* _ Q*IID ~ IIIlQ; = ~"IID. 3.4 OVERVIEW OF PROOF The proof of Theorem 1 involves an analysis in a Euclidean space where the operator F and projection II serve as tools for interpreting the algorithm's dynamics. The ideas for this type of analysis can be traced back to Van Roy and Tsitsiklis (1996) and have since been used to analyze Sutton's temporal-difference learning algorithm (Tsitsiklis and Van Roy, 1996b). Due to space limitations, we only provide an overview of the proof. We begin by establishing that, with respect to the norm II ·11 D, P is a nonexpansion and F is a contraction. In the first case, we apply Jensen's inequality to obtain IIPJII1 = L7r(x) (LpxyJ(Y») 2 xES yES < L 7r(x) LPXyJ2(y) xES yES L L 7r(X)PXy J 2(y) yES xES yES = IIJI11· The fact that F is a contraction now follows from the fact that I(F J)(x) - (F J)(x)1 ~ al(P J)(x) - (P J)(x)l, for any J, J E ~n and any state XES. Let s : ~m H ~m denote the "steady-state" expectation of the steps taken by the algorithm: s(r) = Eo [¢(Xt) (g(Xt) + a max [¢' (xt+l)r, G(Xt+dJ - ¢' (xt)r) ], where Eo['] denotes the expectation with respect to steady-state probabilities. Some simple algebra gives s(r) = «P'D(F(4>r) - «pr) . Approximate Solutions to Optimal Stopping Problems 1087 We focus on analyzing a deterministic algorithm of the form ft+l = ft + 'Yts(ft). The convergence of the stochastic algorithm we have proposed can be deduced from that of this deterministic algorithm through use of a theorem on stochastic approximation, contained in (Benveniste, et al., 1990). Note that the composition IIFO is a contraction with respect to 1/ . liD with contraction coefficient a since projection is nonexpansive and F is a contraction. It follows that IIF (.) has a fixed point of the form <Pr" for some r" E Rm that uniquely satisfies <Pr" = IIF(<Pr"). To establish convergence, we consider the potential function U(r) = ~ Ilr - r"lIb· We have (V'U(r)y s(r) (r - r" ) I <pI D (F( <pr) - <pr) (r - r" r<PID(IIF(<Pr) - (I - II)F(<Pr) - <pr) ( <Pr - <Pr" ) I D ( IIF( <pr) - <pr) , where the last equality follows because <pI DII = <pI D. Using the contraction property of F and the nonexpansion property of projection, we have UIIF(<pr) - <pr"UD = I/IIF(<pr) - IIF(<pr")IID ~ al/<Pr - <pr"I/D' and it follows from the Cauchy-Schwartz inequality that (V'U(r)y s(r) = (<pr - <Pr" r D(IIF(<pr) - <Pr" + <Pr" - <pr) < II<pr - <pr"IIDI/IIF(<pr) - <pr"I/D -1I<pr - <pr"lIb < (a - 1 )1I<p/r - <Pr"l/b. Since <P has full column rank, it follows that (V'U(r))' s(r) ~ -EU(r), for some fixed E > 0, and f t converges to r" . We can further establish the desired error bound: II<Pr" - Q"UD < II<pr" - IIQ"UD + UIIQ" - Q"IID IIIIF(<pr") - IIQ"UD + IIIIQ" - Q"UD < all<Pr" - Q"IID + I/IIQ" - Q"IID, and it follows that I/<pr" - Q"I/D ~ I/IIQ" - Q"IID. I-a 4 CONCLUSION We have proposed an algorithm for approximating Q-functions of optimal stopping problems using linear combinations of fixed basis functions. We have also presented a convergence theorem and overviewed the associated analysis. This paper has served a dual purpose of establishing a new methodology for solving difficult optimal stopping problems and providing a starting point for analyses of Q-learning-Iike algorithms when used in conjunction with function approximators. 1088 J. N. Tsitsiklis and B. Van Roy The line of analysis presented in this paper easily generalizes in several directions. First, it extends to unbounded continuous state spaces. Second, it can be used to analyze certain variants of Q-Iearning that can be used for optimal stopping problems where the underlying Markov processes are not irreducible and/or aperiodic. Rigorous analyses of some extensions, as well as the case that was discussed in this paper, are presented in a forthcoming full-length paper. Acknowledgments This research was supported by the NSF under grant DMI-9625489 and the ARO under grant DAAL-03-92-G-Ol15. References Benveniste, A., Metivier, M., & Priouret, P. (1990) Adaptive Algorithms and Stochastic Approximations, Springer-Verlag, Berlin. Bertsekas, D. P. (1995) Dynamic Programming and Optimal Control. Athena Scientific, Belmont, MA. Gordon, G. J. (1995) Stable Function Approximation in Dynamic Programming. Technical Report: CMU-CS-95-103, Carnegie Mellon University. Jaakkola, T., Jordan M. 1., & Singh, S. P. (1994) "On the Convergence of Stochastic Iterative Dynamic Programming Algorithms," Neural Computation, Vol. 6, No.6. Jaakkola T., Singh, S. P., & Jordan, M. 1. (1995) "Reinforcement Learning Algorithms for Partially Observable Markovian Decision Processes," in Advances in Neural Information Processing Systems 7, J. D. Cowan, G. Tesauro, and D. Touretzky, editors, Morgan Kaufmann. Sutton, R. S. (1988) Learning to Predict by the Method of Temporal Differences. Machine Learning, 3:9-44. Tsitsiklis, J. N. (1994) "Asynchronous Stochastic Approximation and Q-Learning," Machine Learning, vol. 16, pp. 185-202. Tsitsiklis, J. N. & Van Roy, B. (1996a) "Feature-Based Methods for Large Scale Dynamic Programming," Machine Learning, Vol. 22, pp. 59-94. Tsitsiklis, J. N. & Van Roy, B. (1996b) An Analysis of Temporal-Difference Learning with Function Approximation. Technical Report: LIDS-P-2322, Laboratory for Information and Decision Systems, Massachusetts Institute of Technology. Van Roy, B. & Tsitsiklis, J. N. (1996) "Stable Linear Approximations to Dynamic Programming for Stochastic Control Problems with Local Transitions," in Advances in Neural Information Processing Systems 8, D. S. Touretzky, M. C. Mozer, and M. E. Hasselmo, editors, MIT Press. Watkins, C. J. C. H. (1989) Learning from Delayed Rewards. Doctoral dissertation, University of Cambridge, Cambridge, United Kingdom. Watkins, C. J. C. H. & Dayan, P. (1992) "Q-Iearning," Machine Learning, vol. 8, pp. 279-292.
1996
24
1,219
An Orientation Selective Neural Network for Pattern Identification in Particle Detectors Halina Abramowicz, David Horn, Ury Naftaly, Carmit Sahar- Pikielny School of Physics and Astronomy, Tel Aviv University Tel Aviv 69978, Israel halinaOpost.tau.ac.il, horn~neuron.tau.ac.il ury~ost.tau.ac.il, carmitOpost.tau.ac.il Abstract We present an algorithm for identifying linear patterns on a twodimensional lattice based on the concept of an orientation selective cell, a concept borrowed from neurobiology of vision. Constructing a multi-layered neural network with fixed architecture which implements orientation selectivity, we define output elements corresponding to different orientations, which allow us to make a selection decision. The algorithm takes into account the granularity of the lattice as well as the presence of noise and inefficiencies. The method is applied to a sample of data collected with the ZEUS detector at HERA in order to identify cosmic muons that leave a linear pattern of signals in the segmented calorimeter. A two dimensional representation of the relevant part of the detector is used. The algorithm performs very well. Given its architecture, this system becomes a good candidate for fast pattern recognition in parallel processing devices. I Introduction A typical problem in experiments performed at high energy accelerators aimed at studying novel effects in the field of Elementary Particle Physics is that of preselecting interesting interactions at as early a stage as possible, in order to keep the data volume manageable. One class of events that have to be eliminated is due to cosmic muons that pass all trigger conditions. 926 H. Abramowicz. D. Hom. U. Naftaly and C. Sahar-Pikielny The most characteristic feature of cosmic muons is that they leave in the detector a path of signals aligned along a straight line. The efficiency of pattern recognition algorithms depends strongly on the granularity with which such a line is probed, on the level of noise and the response efficiency of a given detector. Yet the efficiency of a visual scan is fairly independent of those features [1]. This lead us to look for a new approach through application of ideas from the field of vision. The main tool that we borrow from the neuronal circuitry of the visual cortex is the orientation selective simple cell [2]. It is incorporated in the hidden layers of a feed forward neural network, possessing a predefined receptive field with excitatory and inhibitory connections. Using these elements we have developed [3] a method for identifying straight lines of varying slopes and lengths on a grid with limited resolution. This method is then applied to the problem of identifying cosmic muons in accelerator data, and compared with other tools. By using a network with a fixed architecture we deviate from conventional approaches of neural networks in particle physics [4]. One advantage of this approach is that the number of free parameters is small, and it can, therefore, be determined using a small data set. The second advantage is the fact that it opens up the possibility of a relatively simple implementation in hardware. This is an important feature for particle detectors, since high energy physics experiments are expected to produce in the next decade a flux of data that is higher than present analysis methods can cope with. II Description of the Task In a two-dimensional representation, the granularity of the rear part of the ZEUS calorimeter [6] can be emulated roughly by a 23 x 23 lattice of 20 x 20 cm2 squares. While such a representation does not use the full information available in the detector, it is sufficient for our study. In our language each cell of this lattice will be denoted as a pixel. A pixel is activated if the corresponding calorimeter cell is above a threshold level predetermined by the properties of the detector. Figure 1: Example of patterns corresponding to a cosmic muon (left), a typical accelerator event (middle), and an accelerator event that looks like a muon (right), as seen in a two dimensional projection. A cosmic muon, depending on its angle of incidence, activates along its linear path typically from 3 to 25 neighboring pixels anywhere on the 23 x 23 grid. The pattern of signals generated by accelerator events consists on average of 3 to 8 clusters, of typically 4 adjacent activated pixels, separated by empty pixels. The clusters Orientation Selective Neural Network 927 tend to populate the center of the 23 X 23 lattice. Due to inherent dynamics of the interactions under study, the distribution of clusters is not isotropic. Examples of events, as seen in the two-dimensional projection in the rear part of the ZEUS calorimeter, are shown in figure 1. The lattice discretizes the data and distorts it. Adding conventional noise levels, the decision of classification of the data into accelerator events and cosmic muon events is difficult to obtain through automatic means. Yet, it is the feeling of experimentalists dealing with these problems, that any expert can distinguish between the two cases with high efficiency (identifying a muon as such) and purity (not misidentifying an accelerator event). We define our task as developing automatic means of doing the same. III The Orientation Selective Neural Network Our analysis is based on a network of orientation selective neurons (OSNN) that will be described in this chapter. We start out with an input layer of pixels on a two dimensional grid with discrete labeling i = (x, y) of the neuron (pixel) that can get the values Sj = 1 or 0, depending on whether the pixel is activated or not. Figure 2: Connectivity patterns for orientation selective cells on the second layer of the OSNN. From left to right are examples of orientations of 0, 7r/4 and 57r/8. Non-zero weights are defined only within a 5 X 5 grid. The dark pixels have weights of +1, and the grey ones have weights of -1. White pixels have zero weights. The input is being fed into a second layer that is composed of orientation selective neurons V;·a at location i with orientation Oa where a belongs to a discrete set of 16 labels, i.e. Oa = a7r/16. The neuron V;·a is the analog of a simple cell in the visual cortex. Its receptive field consists of an array of dimension 5 X 5 centered at pixel i. Examples of the connectivity, for three different choices of a, are shown in Fig. 2. The weights take the values of 1,0 and -1. The second layer consists then of 23 x 23 x 16 neurons, each of which may be thought of as one of 16 orientation elements at some (x, y) location of the input layer. Next we employ a modified Winner Take All (WTA) algorithm, selecting the leading orientation amax(i) for which the largest V;·a is obtained at the given location i. If we find that several V;·a at the same location i are close in value to the maximal one, we allow up to five different V;·a neurons to remain active at this stage of the processing, provided they all lie within a sector of amax ± 2, or Omax ± 7r /8. All other V;·a are reset to zero. If, however, at a given location i we obtain several 928 H. Abramowicz, D. Hom, U. Naftaly and C. Sahar-Pikielny large values of V;·a that correspond to non-neighboring orientations, all are being discarded. The third layer also consists of orientation selective cells. They are constructed with a receptive field of size 7 x 7, and receive inputs from neurons with the same orientation on the second layer. The weights on this layer are defined in a similar fashion to the previous ones, but here negative weights are assigned the value of -3, not -1. For linear patterns, the purpose of this layer is to fill in the holes due to fluctuations in the pixel activation, i.e. complete the lines of same orientation of the second layer. As before, we keep also here up to five highest values at each location, following the same WTA procedure as on the second layer. The fourth layer of the OSNN consists of only 16 components, D a , each corresponding to one of the discrete orientations Q' . For each orientation we calculate the convolution of the first and third layers, Da = l:i vi·a Si . The elements D a carry the information about the number of the input pixels that contribute to a given orientation (}a . Cosmic muons are characterized by high values of D a whereas accelerator events possess low values, as shown in figure 3 below. The computational complexity of this algorithm is O( n) where n is the number of pixels, since a constant number of operations is performed on each pixel. There are basically four free parameters in the algorithm. These are the sizes of the receptive fields on the second and third layer and the corresponding activation thresholds. Their values can be tuned for the best performance, however they are well constrained by the spatial resolution, the noise level in the system and the activation properties of the input pixels. The size of the receptive field determines to a large extent the number of orientations allowed to survive in the modified WTA algorithm. IV OSNN and a Selection Criterion on the Training Set The details of the design of the OSNN and the tuning of its parameters were fixed while training it on a sample of 250 cosmic muons and a similar amount of accelerator events. The sample was obtained by preselection with existing algorithms and a visual scan as a cross-check. For cosmic muon events the highest value of Da , D max , determines the orientation of the straight line. In figure 3 we present the correlation between Dmax and the number np of activated input pixels for cosmic muon and accelerator events. As expected one observes a linear correlation between Dmax and np for the muons while almost no correlation is observed for accelerator events. This allows us to set a selection criterion defined by the separator in this figure. We quantify the quality of our selection by quoting the efficiency of properly identifying a cosmic muon for 100% purity, corresponding to no accelerator event misidentified as a muon. In OSNN-D, which we define according to the separator shown in Fig 3, we obtain 93.0% efficiency on the training set. On the right hand side of Fig 3 we present results of a conventional method for detecting lines on a grid, the Hough transform [7, 8, 9]. This is based on the analysis of a parameter space describing locations and slopes of straight lines. The cells of this space with the largest occupation number, N max , are the analogs of our Dmax. In the figure we show the correlation of N max with np which allows us to draw a separator between cosmic muons and accelerator events, leading to an efficiency of 88% for 100% purity. Although this number is not much lower than the Orientation Selective Neural Network 0_ . . 10 11 20 21 3G • 40 30 Hough . : . ...... . . . .-'" 10 15 10 25 30 35 40 929 Figure 3: Left: Correlation between the maximum value of DOt , D max , and the number np of input pixels for cosmic muon (dots) and accelerator events (open circles). The dashed line defines a separator such that all events above it correspond to cosmic muons (100% purity). This selection criterion has 93% efficiency. Right: Using the Hough Transform method, we compare the values of the largest accumulation cell N max with np and find that the two types of events have different slopes, thus allowing also the definition of a separator. In this case, the efficiency is 88%. efficiency of OSNN-D, we note that the difference between the two types of event distributions is not as significant as in OSNN-D. In the test set, to be discussed in the next chapter, we will consider 40,000 accelerator events contaminated by less than 100 cosmic muons. Clearly the expected generalization quality of OSNN-D will be higher than that of the Hough transform. It should of course be noted that the OSNN is a multi-layer network, whereas the Hough transform method that we have described is a single-layer operation, i.e. it calculates global characteristics. If one wishes to employ some quasi-local Hough transform one is naturally led back to a network that has to resemble our OSNN. V Training and Testing of OSNN-S If instead of applying a simple cut we employ an auxiliary neural network to search for the best classification of events using the OSNN outputs, we obtain still better results. The auxiliary network has 6 inputs, one hidden layer with 5 nodes and one output unit. The input consists of a set of five consecutive DOt centered around Dmax and the total number of activated input pixels, np ' The cosmic muons are assigned an output value s = 1 and the accelerator events s = O. The net is trained on our sample with error back-propagation. This results in an improved separation of cosmic muon events from the rest. Whereas in OSNN-D we find a continuum of cosmic muons throughout the range of Dmax , here we obtain a clear bimodal distribution, as seen in Figure 4. For s ~ 0.1 no accelerator events are found and the muons are selected with an efficiency of 94.7%. This selection procedure will be denoted as OSNN-S. As a test of our method we apply OSNN-S to a sample of 38,606 data events 930 H. Abramowicz, D. Hom, U. Naftaly and C. Sahar-Pikielny that passed the standard physics cuts [5]. The distribution of the neural network output s is presented in Figure 4. It looks very different from the one obtained with the training sample. Whereas the former consisted of approximately 500 events distributed equally among accelerator events and cosmic muons, this one contains mostly accelerator events, with a fraction of a percent of muons. This proportion is characteristic of physics samples. The vast majority of accelerator events are found in the first bin, but a long tail extends throughout s. The last bin in s is indeed dominated by cosmic muons. We performed a visual scan of all 181 events with s ~ 0.1 using the full information from the detector. This allowed us to identify the cosmic-muon events represented by shaded areas in figure 4. For s ~ 0.1 we find 55 cosmic-muon events and 123 accelerator events, 55 of which resemble muons on the rear segment of the calorimeter. The latter, together with the genuine cosmic muons, populate mainly the region of large s values. We conclude that our method picked out the cosmic muons from the very large sample of data, in spite of the fact that it relied just on two-dimensional information from the rear part of the detector. This fact is, however, responsible for the contamination of the high s region by accelerator events that resemble cosmic muons. Even with all its limitations, our method reduces the problem of rejecting cosmic-muon events down to scanning less than one percent of all the events. We conclude that we have achieved the goal that we set for ourselves, that of replacing a laborious visual scan by a computer algorithm with similar reliability. N N OSNN-S (Train) Test 10 10 0.4 0 .• o.a s I Figure 4: Left: Number of events as a function of the output s of an auxiliary neural net. Choosing the separator to be s = 0.1 we obtain an efficiency of 94.7% on our training set. This bimodal distribution holds the promise of better generalization than the OSNN-D method depicted in Figure 3. Muons are represented by shaded areas. Right: Distribution of the auxiliary neural network output s obtained with the OSNN-S selector for the test sample of 38,606 events. The tail of the distribution of accelerator events leads to 123 accelerator events with s > 0.1, including 55 that resemble straight lines on the input layer. 55 genuine cosmic muons were identified in the high s region. Orientation Selective Neural Network 931 VI Summary We have presented an algorithm for identifying linear patterns on a two-dimensional lattice based on the concept of an orientation selective cell, a concept borrowed from neurobiology of vision. Constructing a multi-layered neural network with fixed architecture that implements orientation selectivity, we define output elements corresponding to different orientations, that allow us to make a selection decision. The algorithm takes into account the granularity of the lattice as well as the presence of noise and inefficiencies. Our feed-forward network has a fixed set of synaptic weights. Hence, although the number of neurons is very high, the complexity of the system, as determined by the number of free parameters, is low. This allows us to train our system on a small data set. We are gratified to see that, nontheless, it generalizes well and performs excellently on a test sample that is larger by two orders of magnitude. One may regard our method as a refinement of the Hough transform, since each of our orientation selective cells acts as a filter of straight lines on a limited grid. The major difference from conventional Hough transforms is that we perform semi-local calculations, and proceed in several stages, reflected by the different layers of our network, before evaluating global parameters. The task that we have set to ourselves in the application described here is only one example of problems of pattern recognition that are encountered in the analysis of particle detectors. Given the large flux of data in these experiments, one is faced by two requirements: correct identification and fast performance. Using a structure like our OSNN for data classification, one can naturally meet the speed requirement through its realization in hardware, taking advantage of the basic features of distributed parallel computation. Acknow ledgements We are indebted to the ZEUS Collaboration for allowing us to use the sample of data for this analysis. This work was partly supported by a grant from the Israel Science Foundation. References [1] ZEUS Collab., The ZEUS Detector, Status Report 1993, DESY 1993; M. Derrick et al., Phys. Lett. B 293 (1992) 465. [2] D. H. Hubel and T. N. Wiesel, J. Physiol. 195 (1968) 215. [3] H. Abramowicz, D. Horn, U. Naftaly and C. Sahar-Pikielny, Nuclear Instrum. and Methods in Phys. Res. A378 (1996) 305. [4] B. Denby, Neural Computation, 5 (1993) 505. [5] ZEUS Calorimeter Group, A. Andresen et al., Nucl. Inst. Meth. A 309 (1991) lOI. [6] P. V. Hough, "Methods and means to recognize complex patterns", U.S. patent 3.069.654. [7] D. H. Ballard, Pattern Recognition 3 (1981) II. [8] R. O. Duda and P. E. Hart, Commun. ACM. 15 (1972) I. [9] ZEUS collab., M. Derrick et al., Phys. Lett. B 316 (1993) 412; ZEUS collab., M. Derrick et al., Zeitschrift f. Physik C 69 (1996) 607-620
1996
25
1,220
Early Brain Damage Volker Tresp, Ralph Neuneier and Hans Georg Zimmermann* Siemens AG, Corporate Technologies Otto-Hahn-Ring 6 81730 Miinchen, Germany Abstract Optimal Brain Damage (OBD) is a method for reducing the number of weights in a neural network. OBD estimates the increase in cost function if weights are pruned and is a valid approximation if the learning algorithm has converged into a local minimum. On the other hand it is often desirable to terminate the learning process before a local minimum is reached (early stopping). In this paper we show that OBD estimates the increase in cost function incorrectly if the network is not in a local minimum. We also show how OBD can be extended such that it can be used in connection with early stopping. We call this new approach Early Brain Damage, EBD. EBD also allows to revive already pruned weights. We demonstrate the improvements achieved by EBD using three publicly available data sets. 1 Introduction Optimal Brain Damage (OBD) was introduced by Le Cun et al. (1990) as a method to significantly reduce the number of weights in a neural network. By reducing the number of free parameters, the variance in the prediction of the network is often reduced considerably which -in some cases- leads to an improvement in generalization performance of the neural network. OBD might be considered a realization of the principle of Occam's razor which states that the simplest explanation (of the training data) should be preferred to more complex explanations (requiring more weights). If E is the cost function which is minimized during training, OBD calculates the {Volker. Tresp, Ralph. Neuneier, Georg.Zimmermann}@mchp.siemens.de. 670 V. Tresp, R. Neuneierand H. G. Zimmermann saliency of each parameter Wi defined as 11)2E OBD(wd = A(Wi) = 2 I)w~ wl· s Weights with a small OBD(wd are candidates for removal. OBD(wd has the intuitive meaning of being the increase in cost function if weight Wi is set to zero under the assumptions • that the cost function is quadratic, • that the cost function is "diagonal" which means it can be written as E = Bias + 1/2l:i hi(Wi - wi)2 where where {wn~l are the weights in a (local) optimum of the cost function (Figure 1) and the hi and BIAS are parameters which are dependent on the training data set. • and that Wi ~ wi. In practice, all three assumptions are often violated but experiments have demonstrated that OBD is a useful method for weight removal. In this paper we want to take a closer look at the third assumption, i. e. the assumption that weights are close to optimum. The motivation is that theory and practice have shown that it is often advantageous to perform early stopping which means that training is terminated before convergence. Early stopping can be thought of as a form of regularization: since training typically starts with small weights, with early stopping weights are biased towards small weights analogously to other regularization methods such as ridge regression and weight decay. According to the assumptions in OBD we might be able to apply OBD only in heavily overtrained networks where we loose the benefits of early stopping. In this paper we show that OBD can be extended such that it can work together with early stopping. We call the new criterion Early Brain Damage (EBD). As in OBD, EBD contains a number of simplifying assumptions which are typically invalid in practice. Therefore, experimental results have to demonstrate that EBD has benefits. We validate EBD using three publicly available data sets. 2 Theory As in OBD we approximate the cost function locally by a quadratic function and assume a "diagonal" form. Figure 1 illustrates that OBD(Wi) for Wi = wi calculates the increase in cost function if Wi is set to zero. In early stopping where Wi =f. wi, o B D( Wi) calculates the quantity denoted as Ai in Figure 1. Consider I)E Bi = ---Wi' I)wi The saliency of weights Wi in Early Stopping Pruning ESP(wd = Ai + Bi is an estimate of how much the cost function increases if the current Wi (i. e. Wi m early stopping) is set to zero. Finally, consider Early Brain Damage \ \ \ \ \ \ \ _.l __ _ \ \ \ \ B. , 1 c· _____ 1 E / / / w* 1 I I I I I I I I I I 671 Figure 1: The figure shows the cost function E as a function of one weight Wi in the network. wi is the optimal weight. Wi is the weight at an early stopping point. If OBD is applied at Wi, it estimates the quantity Ai- ESP(Wi) = Ai + Bi = E(Wi) - E(Wi = 0) estimate the increase in cost function if Wi is pruned. EBD{Wi) = Ai + Bi + Ci = E{wi) - E(Wi = 0) is the difference in cost function if we would train to convergence and if we would set Wi = O. In other words EBD{wd = OBD(wi). 672 V. Tresp, R. Neuneierand H. G. Zimmermann The saliency of weight Wi in EBD is EBD(wd = OBD(w;) = Ai + B j + Cj which estimates the increase in cost function if Wj is pruned after convergence (i.e. EBD(wd = OBD(wt)) but based on local information around the current value of Wi. In this sense EBD evaluates the "potential" of Wi. Weights with a small EBD(wd are candidates for pruning. Note, that all terms required for EBD are easily calculated. With a quadratic cost function E = l:f=I(J!' - N N(xk))2 OBD approximates (OBD-approximation) 82E ~ 2 K (8NN(Xk))2 8w~ 2: 8w' I k=1 I where (xk, J!'){f=1 are the training data and N N(xk) is the network response. 3 Extensions 3.1 Revival of Weights (1) In some cases, it is beneficial to revive weights which are already pruned. Note, that Ci exactly estimate the decrease in cost function if weight Wi is "revived". Weights with a large Ci(Wi = 0) are candidates for revival. 3.2 Early Brain Surgeon (EBS) After OBD or EBD is performed, the network needs to be retrained since the "diagonal" approximation is typically violated and there are dependencies between weights. Optimal Brain Surgeon (OBS, Hassibi and Storck, 1993) does not use the "diagonal" approximation and recalculates the new weights without explicit retraining. OBS still assumes a quadratic approximation of the cost function. The saliency in OBS is W~ Li = 2[H~I]ii where [H- 1]ii is i-th diagonal element of the inverse of the Hessian. Li estimates the increase in cost if the i-th weight is set to zero and all other weights are retrained. To recalculate all weights after weight Wi is removed apply Wi -1 Wnew = Wold [H-l]ii H ei where ei is the unit vector in the i-th direction. Analogously to OBS, Early Brain Surgeon EBS would first calculate the optimal weight vector using a second order approximation of the cost function w*=w_H-18E 8w and then apply OBS using w*. We did not pursue this idea any further since our initial experiments indicated that W* was not estimated very accurately in praxis. Hassibi et al. (1994) achieved good performance with OBS even when weights were far from optimal. Early Brain Damage 673 3.3 Approximations to the Hessian and the Gradient Finnoff et al. (1993) have introduced the interesting idea that the relevant quantities for OBD can be estimated from the statistics of the weight changes. Consider the update in pattern by pattern gradient descent learning and a quadratic cost function aWi = -'1 OEk = 2'1(yk _ N N(xk)) oN N(xk) ow OWi with Ek = (Yk - N N(Xk))2 where '1 is the learning rate. We assume that xk and yk are drawn online from a fixed distribution (which is strictly not true since in pattern by pattern learning we draw samples from a fixed training data set). Then, using the quadratic and "diagonal" approximation of the cost function and assuming that the noise f in the model is additive uncorrelated with variance (J'2 1 (2) and v AR(aWi) = V AR (2'1(yk - N N(xk)) oN :w:xk)) = 4~'V AR ((v" -N N'(zk)) aN:w:xk)) +4~'V AR ((W; _ Wi) (aN :w:zk)) ') = 4,Nc (aN:w:zk))' +4~'(w; - wo)'VAR ( eN :w:zk)) ') where N N* (xk) is the network output with optimal weights {wi} ~1. Note, that in the OBD approximation (Equation 1) and If we make the further assumption that oN N(xk)j OWi is Gaussian distributed with zero mean2 1 e stands for the expected value. With Wi kept at a fixed value. 2The zero mean assumption is typically violated but might be enforced by renormalization. 674 V. Tresp, R. Neuneier and H. G. Zimmermann we obtain (3) The first term in Equation 3 is a result of the residual error which is translated into weight fluctuations. But note, that weights with a small variance with a large /)2 E / /)wl fluctuate the most. The first term is only active when there is a residual error, i.e. (72 > O. The second term is non-zero independent of (72 and is due to the fact that in sample-by-sample learning, weight updates have a random component. From Equation 2 and Equation 3 all terms needed in EBD (i. e. /)E//)w, and /)2 E / /)wl) are easily estimated. 4 Experimental Results In our experiments we studied the performance of OBD, ESP and EBD in connection with early stopping. Although theory tells us that EBD should provide the best estimate of the the increase in cost function by the removal of weight Wi, it is not obvious how reliable that estimate is when the assumptions ("diagonal" quadratic cost function) are violated. Also we are not really interested in the correct estimate of the increase in cost function but in a ranking of the weights. Since the assumptions which go into OBD, EBD, ESP (and also OBS and EBS) are questionable, the usefulness of the new methods have to be demonstrated using practical experiments. We used three different data sets: Breast Cancer Data, Diabetes Data, and Boston Housing Data. All three data sets can be obtained from the DCI repository (ftp:ics.uci.edu/pub/machine-Iearning-databases). The Breast Cancer Data contains 699 samples with 9 input variables consisting of cellular characteristics and one binary output with 458 benign and 241 malignant cases. The Diabetes Data contains 768 samples with 8 input variables and one binary output. The Boston Housing Data consist of 506 samples with 13 input variables which potentially influence the housing price (output variable) in a Boston neighborhood (Harrison & Rubinfeld, 1978). Our procedure is as follows. The data set is divided into training data, validation data and test data. A neural network (MLP) is trained until the error on the validation data set starts to increase. At this point OBD, ESP and EBD are employed and 50% of all weights are removed. After pruning the networks are retrained until again the error on the validation set starts to increase. At this point the results are compared. Each experiment was repeated 5-times with different divisions of the data into training data, validation data and test data and we report averages over those 5 experiments. Table 1 sums up the results. The first row shows the number of data in training set, validation set and test set. The second row displays the test set error at the (first) early stopping point. Rows 3 to 5 show test set performance of OBD, ESP and EBD at the stopping point after pruning and retraining (absolute / relative to early stopping). In all three experiments, EBD performed best and OBD was second best in two experiments (Breast Cancer Data and Diabetes Data). In two experiments (Breast Cancer Data and Boston Housing Data) the performance after pruning improved. Early Brain Damage 675 Table 1: Comparing OBD, ESP, and EBD. Breast Cancer Diabetes Boston Housing Data Train/V /Test 233/233/233 256/256/256 168/169/169 Hidden units 10 5 3 MSE (Stopp) 0.0340 0.1625 0.2283 OBD 0.0328/ 0.965 0.1652/1.017 0.2275 /0.997 ESP 0.0331 / 0.973 0.1657 /1.020 0.2178 / 0.954 EBD 0.0326/ 0.959 0.1647 /1.014 0.2160 / 0.946 5 Conclusions In our experiments, EBD showed better performance than OBD if used in conjunction with early stopping. The improvement in performance is not dramatic which indicates that the rankings of the weights in OBD are reasonable as well. References Finnoff, W., Hergert, F., and Zimmermann, H. (1993). Improving model selection by nonconvergent methods, Neural Networks, Vol. 6, No.6. Hassibi, B. and Storck, D. G. (1993). Second order derivatives for network pruning: Optimal Brain Surgeon. In: Hanson, S. J., Cowan, J. D., and Giles, C. L. (Eds.). Advances in Neural Information Processing Systems 5, San Mateo, CA, Morgan Kaufman. Hassibi, B., Storck, D. G., and Wolff, G. (1994). Optimal Brain Surgeon: Extensions and performance comparisons. In: Cowan, J. D., Tesauro, G., and Alspector, J. (Eds.). Advances in Neural Information Processing Systems 6, San Mateo, CA, Morgan Kaufman. Le Cun, Y., Denker, J. S., and SolI a, S. A. (1990). Optimal brain damage. In: D. S. Touretzky (Ed.). Advances in Neural Information Processing Systems 2, San Mateo, CA, Morgan Kaufman.
1996
26
1,221
Text-Based Information Retrieval Using Exponentiated Gradient Descent Ron Papka, James P. Callan, and Andrew G. Barto * Department of Computer Science University of Massachusetts Amherst, MA 01003 papka@cs.umass.edu, callan@cs.umass.edu, barto@cs.umass.edu Abstract The following investigates the use of single-neuron learning algorithms to improve the performance of text-retrieval systems that accept natural-language queries. A retrieval process is explained that transforms the natural-language query into the query syntax of a real retrieval system: the initial query is expanded using statistical and learning techniques and is then used for document ranking and binary classification. The results of experiments suggest that Kivinen and Warmuth's Exponentiated Gradient Descent learning algorithm works significantly better than previous approaches. 1 Introduction The following work explores two learning algorithms - Least Mean Squared (LMS) [1] and Exponentiated Gradient Descent (EG) [2] - in the context of text-based Information Retrieval (IR) systems. The experiments presented in [3] use connectionist learning models to improve the retrieval of relevant documents from a large collection of text. Here, we present further analysis of those experiments. Previous work in the area employs various techniques for improving retrieval [6, 7, 14]. The experiments presented here show that EG works significantly better than widely used ad hoc methods for finding a good set of query term weights. The retrieval processes being considered operate on a collection of documents, a natural-language query, and a training set of documents judged relevant or nonrelevant to the query. The query may be, for example, the information request submitted through a web-search engine, or through the interface of a system with This material is based on work supported by the National Science Foundation, Library of Congress, and Department of Commerce under cooperative agreement number EEC9209623. Any opinions, findings and conclusions or recommendations expressed in this material are those of the author and do not necessarily reflect those of the sponsor. 4 R. Papka, J. P. Callan and A. G. Barto domain-specific information such as legal, governmental, or news data maintained as a collection of text. The query, expressed as complete or incomplete sentences, is modified through a learning process that incorporates the terms in the test collection that are important for improving retrieval performance. The resulting query can then be used against collections similar in domain to the training collection. Natural language query: An insider-trading case. IR system query using default weights: #WSUM( 1.0 An 1.0 insider 1.0 trading 1.0 case ); After stop word and stemming process: #WSUM( 1.0 insid 1.0 trade 1.0 case )j After Expansion and learning new weights: #WSUM( 0.181284 insid 0.045721 trade 0.016127 case 0.088143 boesk 0.000001 ivan 0.026762 sec 0.052081 guilt 0 .074493 drexel 0.000001 plead 0.003834 fraud 0.091436 takeov 0.018636 lavyer 0.000000 crimin 0.137799 alleg 0.057393 attorney 0.155781 charg 0.024237 scandal 0.000000 burnham 0.000000 lambert 0.026270 investig 0.000000 vall 0.000000 firm 0.000000 illeg 0.000000 indict 0.000000 prosecutor 0.000000 profit 0.000000 ); Figure 1: Query Transformation Process. The query transformation process is illustrated in Figure 1. First, the naturallanguage query is transformed into one which can be used by the query-parsing mechanism of the IR system. The weights associated with each term are assigned a default value of 1.0, implying that each term is equally important in discriminating relevant documents. The query then undergoes a stopping and stemming process, by which morphological stemming and the elimination of very common words, called stopwords, increases both the effectiveness and efficiency of a system [9]. The query is subsequently expanded using a statistical term-expansion process producing terms from the training set of documents. Finally, a learning algorithm is invoked to produce new weights for the expanded query. 2 Retrieval Process Text-based information retrieval systems allow the user to pose a query to a collection or a stream of documents. When a query q is presented to a collection c, each document dEc is examined and assigned a value relative to how well d satisfies the semantics of the request posed by q. For any instance of the triple < q, d,c >, the system determines an evaluation value attributed to d using the function eval(q, d, c) . The evaluation function eval(q, d, c) = L:t~:i;idi was used for this work, and is based on an implementation of INQUERY [8]. It is assumed that q and d are vectors of real numbers, and that c contains precomputed collection statistics in addition to the current set of documents. Since the collection may change over time, it may be necessary to change the query representation over time; however, in what follows the training collection is assumed to be static, and successful learning implies that the resulting query generalizes to similar collections. An IR system can perform several kinds of retrieval tasks. This work is specifically concerned with two retrieval processes: document ranking and document classification. A ranking of documents based on query q is achieved by sorting all documents in a collection by eval1,1ation value. Binary classification is achieved by determining a threshold () such that for class R, eval (q, d, c) ~ () -+ d E R, and Text-Based Information Retrieval Using Exponentiated Gradient Descent 5 eval (q, d, c) < () --+ d E R, so that R is the set of documents from the collection that are classified as relevant to the query, and R is the set classified as non-relevant. Central to any IR system is a parsing process used for documents and queries, which produces tokens called terms. The terms derived from a document are used to build an inverted list structure which serves as an index to the collection. The natural-language query is also parsed into a set of terms. Research-based IR systems such as INQUERY, OKAPI [111, and SMART [5], assume that the cooccurrence of a term in a query and a document indicates that the document is relevant to the query to some degree, and that a query with multiple terms requires a mechanism by which to combine the evidence each co-occurrence contributes to the document's degree of relevance to the query. The document representation for such systems is a vector, each element of which is associated with a unique term in the document. The values in the vector are produced by a term-evaluation function comprised of a t.erm frequency component, tf, and an inverse document frequency component, idj, which are described in [8, 11]. The tf component causes the termevaluation value to increase as a query-term's occurrence in the document increases, and the idj component causes the term-evaluation value to decrease as the number of documents in the collection in which the term occurs increases. 3 Query Expansion Though it is possible to learn weights for terms in the original query, better results are obtained by first expanding the query with additional terms that can contribute to identifying relevant documents, and then learning the weights for the expanded query. The optimal number of terms by which to expand a query is domain-dependent, and query expansion can be performed using several techniques, including thesaurus expansion and statistical methods [12]. The query expansion process performed in this work is a two-step process: term selection followed by weight assignment. The term selection process ranks all terms found in relevant documents by an information metric described in [8]. The top n terms are used in the expanded query. The experiments in this work used values of 50 and 1000 for n. The most common technique for weight assi~ment is derived from a closed-form function originally presented by Rocchio in l6], but our experiments show that a single-neuron learning approach is more effective. 3.1 Rocchio Weights We assume that the terms of the original query are stored in a vector t, and that their associated weights are stored in q. Assuming that the new terms in the expanded query are stored t', the weights for q' can be determined using a method originally developed by Rocchio that has been improved upon in [7, 8]. Using the notation presented above, the weight assignment can be represented in the linear form: q' = Ci* j(t) + /hr(t', R q , c) +"I*nr(t', Rq , c), where j is a function operating on the terms in the original query, r is a function operating on the term statistics available from the training set of relevant documents (Rq), and nr is a function operating on the statistics from the non-relevant documents (Rq). The values for Ci, (3, and "I have been the focus of many IR experiments, and 1.0, 2.0, and 0.5, have been found to work well with various implementations of the functions j, r, and nr [7]. 3.2 LMS and EG In the experiments that follow, LMS and EG were used to learn query term weights. Both algorithms were used in a training process attempting to learn the association between the set of training instances t documents) and their corresponding binary classifications (relevant or non-relevant). A set of weights tV is updated given an input instance x and a target binary classification value y. The algorithms learn the association between x and y perfectly if tV· x = y, otherwise the value (y - tV· x) is the error or loss incurred. The task of the learning algorithm is to learn the values of tV for more than one instance of X. The update rule for LMS is tVt+l = tVt + Tt, where it = -21Jt(tVt' Xt - Yt)Xt, where . 1 Th d I £ EG' -. uh -eFt,; h the step-SIze 1Jt = x .x . e up ate ru e or IS Wt+l,i = "N " ;: _, were t t ~j=l Wt ,; e t" 6 R. Papka, J. P. Callan and A. G. Barto r . - -2" (w . X - y)x . and" 2 t,t 'It t t t t,t, 'It 3(maxi(Xt,i)-mini(Xt,i»' There are several fundamental differences between LMS and EG; the most salient is that EG has a multiplicative exponential update rule, while LMS is additive. A less obvious difference is the derivation of these two update rules. Kivinen and Warmuth [2] show that both rules are approximately derivable from an optimization task that minimizes the linear combination of a distance and a loss function: distance (Wt+1 ,Wt) + 1Jtloss(Yt, Wt . Xt). But the distance component for the derivation leading to the LMS update rule uses the squared Euclidean distance Ilwt+1 wtll~, while the derivation leading to the EG update rule uses relative entropy or l:~1 Wt+1,i In W~:,l:i . Entropy metrics had previously been used as the loss component [4]. One purpose of Kivinen and Warmuth's work was to describe loss bounds for these algorithms; however, they also observed that EG suffers significantly less from irrelevant attributes than does LMS. This hypothesis was tested in the experiments conducted for this work. 4 Experiments Experiments were conducted on 100 natural-language queries. The queries were manually transformed into INQUERY syntax, expanded using a statistical technique described in [8], and then given a weight assignment as a result of a learning process, One set of experiments expanded each query by 50 terms and another set of experiments expanded each query by 1000 terms. The purpose of the latter was to test the ability of each algorithm to learn in the presence of many irrelevant attributes. 4.1 Data The queries used are the description fields of information requests developed for Text Retrieval Conferences (TREC) [10]. The first set of queries was taken from TREC topics 51-100 and the second set from topics 101-150, for a total of 100 queries. After stopping and stemming, the average number of terms remaining before expansion was 8.34 terms. Training and testing for all queries was conducted on subsets of the Tipster collection, which currently contains 3.4 gigabytes of text, including 206,201 documents whose relevance to the TREC topics has been evaluated. The collection is partitioned into 3 volumes. The judged documents from volumes 1 and 2 were used for training, while the documents from volume 3 were used for testing. Volumes 1 and 2 contain 741,856 documents from the Associated Press(1988-9), Department of Energy abstract, Federal Register(1988-9), Wall Street Journal(1987 -91), and Ziff-Davis Computer-select articles. Volume 3 contains 336,310 documents from Associated Press(1990), San Jose Mercury News(1991), and Ziff-Davis articles. Only a subset of the data for the TREC-Tipster environment has been judged. Binary judgments are assessed by humans for the top few thousand documents that were retrieved for each query by participating systems from various commercial and research institutions. Based on the judged documents available for volumes 1 and 2, on average 280 relevant documents and 1236 non-relevant documents were used to train each query. 4.2 Training Parameters Rocchio weights were assigned based on coefficients described in Section 3.1. LMS and EG update rules were applied using 100,000 random presentations of training instances. It was empirically determined that this number of presentations was sufficient to allow both learning algorithms to produce better query weights than the Rocchio assignment based on performance metrics calculated using the training instances. In reality, of course, the number of documents that will be relevant to a particular query is much smaller than the number of documents that are non-relevant. This property gives rise to the question of what is an appropriate sampling bias Text-Based Information Retrieval Using Exponentiated Gradient Descent 7 of training instances, considering that the ratio of relevant to non-relevant documents approaches 0 in the limit. In the following experiments, LMS benefitted from uniform random sampling from the set of training instances, while EG benefitted from a balanced sampling, that is uniform random sampling from relevant training instances on even iterations and from non-relevant instances on odd iterations. A pocketing technique was applied to the learning algorithms [131. The purpose of this technique is to find a set of weights that optimize a specilic user's utility function. In the following experiments, weights were tested every 1000 iterations using a recall and precision performance metric. If a set of weights produced a new performance-metric maximum, it was saved. The last set saved was assumed to be the result of the algorithm, and was used for testing. A binary classification value pair (A, B) is supplied as the target for training, where A is the classification value for relevant documents, and B is the classification value for non-relevant documents. Using the standard classification value pair (1, 0), INQUERY's document representation inhibits learning due to the large error caused by these unattainable values. Therefore, testing was done and resulted in the observation that .4 was the lowest attainable evaluation value for a document, and .47 appeared to be a good classification value for relevant documents. The classification value pair used for both the LMS and EG algorithms was thus (.47, .40). 4.3 Evaluation In the experiments that follow, R-Precision (RP) was used to evaluate ranking performance, and a new metric, Lower Bound Accuracy (LBA) was used to evaluate classification. Both metrics make use of recall and precision, which are defined as follows: Assume there exists a set of documents sorted by evaluation value and a process that has performed classification, and that a = number of relevant documents classified as relevant, b = number of non-relevant documents classified as relevant, c = number of relevant documents classified as non-relevant, and d = number of non-relevant documents classified as non-relevant; then, Recall = a~c' and Precision = a~b [3]. Precision and recall can be calculated at any cut-off point in the sorted list of documents. R-Precision is calculated using the top n documents, where n is the number of relevant training documents available for a query. Lower Bound Accuracy (LBA) is a metric that assumes the minimum of a classifier's accuracy with respect to relevant documents and its accuracy with respect to nonrelevant documents. It is defined as LBA = min(a~c' btd)' An LBA value can be interpreted as the lower bound of the percent of instances a classifier will correctly classify, regardless of an imbalance between the actual number of relevant and nonrelevant documents. This metric requires a threshold e. The threshold is taken to be the evaluation value of the document at a cut-off point in the sorted list of training documents where LBA is maximized. Hence, e = maXi (LBA(di , Rq, Rq», where di is the ith document in the sorted list. 4.4 Results Query type RP LBA NL 22.0 88.6 EXP 28.7 92.0 ROC 33.4 94.0 LMS 32.5 89.8 EG 40.3 95.1 Table 1: Query expansion by 50 terms 8 R. Papka, J. P. Callan and A. G. Barto The following results show the ability of a query weight assignment to generalize. The weights are derived from a subset of the training collection, and the values reported are based on performance on the test collection. The results of the 50term-expansion experiments are listed in Table 1 1. They indicate that the expanded query has an advantage over the original query, and that the EG-trained query generalized better than the other algorithms, while Rocchio appears to be the next best. In terms of ranking, EG gives rise to a 20% improvement over the Rocchio assignment, and realizes 1.2% improvement in terms of classification. This apparently slight improvement in classification in fact implies that EG is correctly classifying at least 3000 documents more than the other approaches. Table 2 shows a cross-algorithm analysis in which any two algorithms can be compared. The analysis is calculated using both RP and LBA over all queries. An entry for row i column j indicates the number of queries for which the performance of algorithm i was better than algorithm j. Based on sign tests with 0: = .01, the results confirm that EG significantly generalized better than the other algorithms.2 Query counts: RP-LBA Query type NL EXP ROC LMS EG NL 30 -37 18 - 13 24 - 53 12 - 13 EXP 60 - 62 9 - 17 35 - 66 11 - 19 ROC 71- 86 72 -79 53 - 73 17 - 37 LMS 66 - 46 54 -34 38 - 26 13 - 15 EG 79 - 85 80 -80 70 - 62 74 - 84 Table 2: Cross Algorithm Analysis over 100 queries expanded by 50 terms. As explained in Section 4.3, the thresholds used to calculate the LBA performance metric are determined by obtaining an evaluation value in the training data corresponding to the cut-off point where LBA was maximized. The threshold analysis in Table 3 shows the best attainable classification performance against performance actually achieved. The results indicate that there is still room for improvement; however, they also indicate that this methodology is acceptable. The results for queries expanded by 1000 terms are listed in Table 4. Since the average document length in the Tipster collection is 806 terms (non-unique), at least 20% of the terms in the expanded query are generally irrelevant to a particular document. The results indicate that irrelevant attributes prevent all but EG from generalizing well. Comparing the performance of EG and LMS adds evidence to the Kivinen-Warmuth hypothesis that EG yields a smaller loss than LMS, given many irrelevant attributes. Juxtaposing the results of the 50-term and 1000-termexpansion experiments suggests that using a statistical filter for selecting the top few terms is better than expanding the query by many terms and having the learning algorithm perform term selection. 5 Conclusion The experiment results presented here provide evidence that single-neuron learning algorithms can be used to improve retrieval performance in IR systems. Based on performance metrics that test the quality of a classification process and a document ranking process, the weights produced by EG were consistently better than previously available methods. lR-Precision (RP) and Lower Bound Accuracy (LBA) performance values are normalized to a 0-100 scale. Values are reported for: NL = original natural language query; EXP = expanded query with weights set to 1.0; ROC = expanded query with weights based on Rocchio assignment; LMS = expanded query with weights based on LMS learning; and EG= expanded query with weights based on EG learning. 2Recent experiments using the optimization algorithm DFO (presented in [7]) suggest that certain parameter settings make it competitive with EG. Text-Based Information Retrieval Using Exponentiated Gradient Descent 9 I Query type I Potential LBA I Actual LBA I NL 91.9 88.6 EXP 95.5 92.0 ROC 96.7 94.0 LMS 92.6 89.8 EG 97.1 95.1 Table 3: Threshold Analysis: Query expansion by 50 terms. I Query type I RP I LBA I NL 22.0 88.6 EXP 14.4 76.5 ROC 19.7 82.5 LMS 20.4 86.7 EG 35.0 93.2 Table 4: Query expansion by 1000 terms. References [1] B. Widrow and M. Hoff, "Adaptive switching circuits", In 1960 IRE WESCON Convention Record, pp. 96-104, New York, 1960. [2] J. Kivinen, Manfred Wartmuth, "Exponentiated Gradient Versus Gradient Descent for Linear Predictors", UCSC Tech report: UCSC-CRL-94-16, June 21, 1994. [3] D. Lewis, R. Schapire, J. Callan, and R. Papka, "Thaining Algorithms for Linear Text Classifiers", Proceeding of SIGIR 1996. [4] B.S. Wittner and J.S. Denker, "Strategies for Teaching Layered Networks Classification Tasks", NIPS proceedings, 1987. [5] G. Salton, "Relevance Feedback and optimization of retrieval effectiveness. In The Smart system - experiments in automatic document processing" , 324-336. Englewood Cliffs, NJ: Prentice Hall Inc., 1971. [6] J.J. Rocchio, "Relevance Feedback in Information Retrieval in The Smart System - Experiments in Automatic document processing", 313-323. Englewood Cliffs, NJ: Prentice Hall Inc., 1971. [7] C. Buckley and G. Salton, "Optimization of Relevance Feedback Weights", Proceeding of SIGIR 95 Seattle WA, 1995. [8] J. Allan, L. Ballesteros, J. Callan, W.B. Croft, and Z. Lu, "Recent Experiments with Inquery", TREC-4 Proceedings, 1995. [9] M. Porter, "An Algorithm for Suffix Stripping", Program, Vol 14(3), pp. 130137,1980. [10] D. Harman, Proceedings of Text REtrievl Conferences (TREC), 1993-5. [11] S.E. Robertson, W. Walker, S. Jones, M.M. Hancock-Beaulieu, and M.Gatford, "Okapi at TREC-3" , TREC-3 Proceedings, 1994. [12] G. Salton, Automatic Text Processing, Addison-Wesley Publishing Co, Massachusetts, 1989. [13] S.I. Gallant, "Optimal Linear Discrimants", Proceedings ofInternational Conference on Pattern Recognition, 1986. [14] B.T. Bartell, "Optimizing Ranking Functions: A Connectionist Approach to Adaptive Information Retrieval" , Ph.D. Theis, UCSD 1994.
1996
27
1,222
An Analog Implementation of the Constant Statistics Constraint For Sensor Calibration John G. Harris and Yu-Ming Chiang Computational Neuro-Engineering Laboratory Department of Computer and Electrical Engineering University of Florida Gainesville, FL 32611 Abstract We use the constant statistics constraint to calibrate an array of sensors that contains gain and offset variations. This algorithm has been mapped to analog hardware and designed and fabricated with a 2um CMOS technology. Measured results from the chip show that the system achieves invariance to gain and offset variations of the input signal. 1 Introduction Transistor mismatches and parameter variations cause unavoidable nonuniformities from sensor to sensor. A one-time calibration procedure is normally used to counteract the effect of these fixed variations between components. Unfortunately, many of these variations fluctuate with time--either with operating point (such as datadependent variations) or with external conditions (such as temperature). Calibrating these sensors one-time only at the "factory" is not suitable-much more frequent calibration is required. The sensor calibration problem becomes more challenging as an increasing number of different types of sensors are integrated onto VLSI chips at higher and higher integration densities. Ullman and Schechtman studied a simple gain adjustment algorithm but their method provides no mechanism for canceling additive offsets [10]. Scribner has addressed this nonuniformity correction problem in software using a neural network technique but it will be difficult to integrate this complex solution into analog hardware [9]. A number of researchers have studied sensors that output the time-derivative of the signal[9][4]. A simple time derivative 700 1. G. Harris and Y. Chiang cancels any additive offset in the signal but also loses all of the DC and most of the low frequency temporal information present. The offset-correction method proposed by this paper, in effect, uses a time-derivative with an extremely long time constant thereby preserving much of the low-frequency information present in the signal. However, even if an ideal time-derivative approximation is used to cancel out additive offsets, the standard deviation process described in this paper can be used to factor out gain variations. We hope to obtain some clues for sensory adaptation from neurobiological systems which possess a tremendous ability to adapt to the surrounding environment at multiple time-scales and at multiple stages of processing. Consider the following experiments: • After staring at a single curved line ten minutes, human subjects report that the amount of curvature perceived appears to decrease. Immediately after training, the subjects then were shown a straight line and perceived it as slightly curved in the opposite direction[5]. • After staring long enough at an object in continuous motion, the motion seems to decrease with time. Immediately after adaptation, subjects perceive motion in the opposite direction when looking at stationary objects. This experiment is called the waterfall effect[2]. • Colors tend to look less saturated over time. Color after-images are perceived containing exactly the opponent colors of the original scene[1]. Though the purpose of these biological adaptation mechanisms is not clear, some theories suggest that these methods allow for fine-tuning the visual system through long-term averaging of measured visual parameters[lO]. We will apply such continuous-calibration procedures to VLSI sensor calibration. The real-world variable x(t) is transduced by a nonlinear response curve into a measured variable y(t). For a single operating point, the linear approximation can be written as: y(t) = ax(t) + b (1) with a and b being the multiplicative gain and additive offset respectively. The gain and offset values vary from pixel to pixel and may vary slowly over time. Current infra-red focal point arrays (IRFPAs) are limited by their inability to calibrate out component variations [3]. Typically, off-board digital calibration is used to correct nonuniformities in these detector arrays; Special calibration images are used to calibrate the system at startup. One-time calibration procedures such as these do not take into account other operating points and will fail to recalibrate for any drift in the parameters. 2 Implementing Natural Constraints A continuous calibration system must take advantage of natural constraints available during the normal operation of the sensors. One theory holds that biological systems adapt to the long-term average of the stimulus. For example, the constraints for the three psychophysical examples mentioned above (curvature, motion and color adaptation) may rely on the following constraints: The Constant Average Statistics Constraint 701 • The average line is straight. • The average motion is zero. • The average color is gray. The system adapts over time in the direction of this average, where the average must be taken over a very long time: from minutes to hours. We use two additional constraints for offset/gain normalization, namely: • The average pixel intensities are identical. • The variances of the input for each pixel are all identical. Each of these constraints assumes that the photoarray is periodically moving in the real-world and that the average statistics each pixels sees should be constant when averaged over a very long time. In pathological situations where humans or machines are forced to stare at a single static scene for a long time, we violate this assumption. We estimate the time-varying mean and variance by using an exponentially shaped window into the past. The equations for mean and variance are: m(t) = y(t - b.)e-tJ,./T db. 11 00 T 0 (2) and s(t) = Iy(t - b.) - m(t - b.)le-tJ,./T db. 11 00 T 0 (3) The m(t) and s(t) in Equation 2 and 3 can be expressed as low-pass filters with inputs y(t) and Iy(t) - m(t)1 respectively. To simplify the hardware implementation further, we chose the Ll (absolute value) definition of variance instead of the more usual L2 definition. The Ll definition is an equally acceptable definition of signal variation in terms of the complete calibration system. Using this definition, no squares or square roots need be calculated. An added benefit of the Ll norm is that it provides robustness to outliers in the estimation. A zero-mean, unity variance 1 signal can then be produced with the following shift/ normalization formula: x(t) = y(t) - m(t) s(t) (4) Equation 2, Equation 3 and Equation 4 constitute a new algorithm for continuously calibrating systems with gain and offset variations. Note that without additional apriori knowledge about the values of the gains and offsets, it is impossible to recover the true value of the signal x(t) given an infinite history of y(t). This is an ill-posed problem even with fixed but unknown gain and offset parameters for each sensor. All that can be done is to calibrate each sensor output to have zero offset and unity variance. After calibration, each sensor would therefore all have the same offset and variance when averaged over a long time. The fundamental assumption embedded 1 For simplicity the signal s(t) will be called the variance estimate throughout the rest of this paper even though technically s(t) is neither the variance nor the standard deviation. 702 J. G. Harris and Y. Chiang y(t} Variance m(t} estimation Mean -----.:S(I) estimation -------1 y(t} 1 R 1 1 A~A 1 :v'V,-L 1 , CI' s(t} 1 1 1 1 m(t) 1 1 x(t 1 _I 1 _____ -_.1 y(t} Divider m(t) x(t) v .. y(t) } I Figure 1: Left: block diagram of continuous-time calibration system, Right: schematic of the divider circuit. in this algorithm is that each sensor measures real-world quantities with the same statistical properties (e.g., mean and variance). For example, this would mean that all pixels in a camera should eventually see the same average intensity when integrated over a long enough time. This assumption leads to other system-level constraints-in this case, the camera must be periodically moving. We have successfully demonstrated this algorithm in software for the case of nonuniformity (gain and offset) correction of images [6]. Since there may be potentially thousands of sensors per chip, it is desirable to build calibration circuitry using subthreshold analog MOS technology to achieve ultra-low power consumption[8]. The next section describes the analog VLSI implementation of this algorithm. 3 Continuous-time calibration circuit The block diagram of the continuous-time gain and offset calibration circuit is shown in Figure 1a. This system includes three building blocks: a mean estimation circuit, a variance estimation circuit and a divider circuit. As is shown, the mean of the signal can be easily extracted by a RC low-pass filter circuit. Since there may be potentially thousands of sensors per chip, it is desirable to build calibration circuitry using subthreshold analog MOS technology to achieve ultra-low power consumption[8] . Figure 2 shows the schematic of the variance estimation circuit. A full-wave rectifier [8] operating in the sub-threshold region is used to obtain the absolute value of the difference between the input and its mean. In the linear region, the current lout is proportional to Iy(t) - m(t)l. As indicated in Equation 3, lout has to be low-pass filtered to obtain s(t). In Figure 2, transconductance amplifiers A3 , A4 and capacitor C2 are used to form a current mode low-pass filter. For signals in the linear region, we can derive the Laplace transform of VI as: R Vt (s) = RC2 S + 1Iout(s) (5) which is a first-order low-pass filter for lout. The value of R is a function of several The Constant Averoge Statistics Constraint 703 yet) 11 1001 V1 met) C2 I V'1 Vb2 Figure 2: Variance estimation circuit. The triangle symbols represent 5-transistor transconductance amplifiers that output a current proportional to the difference of their inputs. fabrication constants and an adjustable bias current. Figure 3(a) shows the expected linear relationship between the measured variance s(t) and the peak-to-peak amplitude of the sine-wave input. The third building block in the calibration system is the divider circuit shown in Figure lb. A fed-back multiplier is used to enforce the constraint that y(t) - m(t) is proportional to x(t)s(t) which results in a scaled version of Equation 4. The characteristics of the divider have been measured and shown in Figure 3(b). With a fixed Vb6 and m(t), we sweep y(t) from m(t) - 0.3V to m(t) + 0.3V and measure the the change of output. A family of input/output characteristics with s(t) =20, 25, 30, 40, 50, 60 and 70nA is shown in Figure 3(b). The divider circuit has been tested up to frequencies of 45kHz. The first version of the calibration circuit has been designed and fabricated in a 2-um CMOS technology. The chip includes the major parts of this calibration circuit: the variance estimation circuit and divider circuit. In our initial design, the mean estimation circuit, which is simply a RC low-pass filter, was built off-chip. However, it can be easily integrated on-chip using a transconductance amplifier and a capacitor. The calibration results for a signal with gain and offset variations are shown in Figure 4. The input signal is a sine wave with a severe gain and offset jump as shown at the top of Figure 4. At the middle of Figure 4, the convergence of the variance estimation is illustrated. It takes a short time for the circuit to converge after any change of the mean or variance or of the input signal. At the bottom of Figure 4, we show the calibrated signal produced by the chip. The output eventually converges to a zero-mean, constant-height sine wave independent of the values of the DC offset and amplitude of the input sine wave. Additional experiments have shown that with the input amplitude changing from 20mV to 90mV, the measured output amplitude varies by less than 3mV. Similarly, when the DC offset is varied from l.5V to 3.5V, the amplitude of the output varies by less than 5mV. These 704 253 .52 x 1"51 >( . ' 25 >( L~ >( .... 2.4 20 30 40 50 eo 70 eo 90 100 110 120 .... poaI<_mV) .... M ' .8 .... jf .5 2.4' .. .~. -<> .• -<>1 J. G. Harris and Y. Chiang 0 0.1 y(1f.<n(Q _I(Q 0.' 0.3 M Figure 3: Left (a) shows the characteristics of measured variance s{t) vs. peakto-peak input voltage. Right (b) shows the characteristics of divider with different s{t). results demonstrate that system is invariant to gain and offset variations of the input. 4 Conclusions The calibration circuit has been demonstrated with the time-constants on the order of lOOms. In many applications, much longer time constants will be necessarily and these cannot be reached with on-chip capacitors even with subthreshold CMOS operation. We expect to use floating-gate techniques where essentially arbitrarily long time-constants can be achieved. Mead has demonstrated a novel adaptive adaptive silicon retina that requires UV light for adaptation to occur [7]. The adaptive silicon retina implemented the constant average brightness constraint. The unoptimized layout area of one of our calibration circuits is about 250x300 um2 in 2um CMOS technology. A future challenge will be to reduce this area and replace the large on-chip capacitors with floating gates. Acknowledgments The authors would like to acknowledge an NSF CAREER Award and Office of Naval Research contract #N00014-94-1-0858. References [1] M. Akita, C. Graham, and Y. Hsia. Maintaining an absolute hue in the presence of different background colors. Vision Research, 4:539-556, 1964. [2] V.R. Carlson. Adaptation in the perception of visual velocity. J. Exp. Psychol., 64(2): 192-197, 1962. [3] M.R. Kruer D.A. Scribner and J.M. Killiany. Infrared focal plane array technology. Proc. IEEE, 79(1):66-85, 1991. The Constant Average Statistics Constraint ~2.52 CD () c::: .~ 2.5 ~ 0.1 0.2 0.3 0.4 705 0.5 0.6 0.7 0.8 0.9 1 2.48 L-_....l.._ _ ___L.......:::::==t==~=~~_'___....l.._ _ ___L.. __ '___...J.......I ...... 1.8 ~ '5 a. '5 1.6 o o 0.1 0.1 0.2 0.3 0.2 0.3 0.4 0.5 0.6 0.4 0.5 0.6 time(sec) 0.7 0.7 0.8 0.9 1 0.8 0.9 1 Figure 4: Calibrating signal with offset and gain 1Jariations. Top: the input signal, y(t). Middle: the computed signal1Jariance s(t). Bottom: the output signal, x(t). [4] T. Delbriick. An electronic photoreceptor sensitive to small changes. In D. Touretzky, editor, Ad1Jance in Neural Information Processing Systems, Volume 1, pages 720-727. Morgan Kaufmann, Palo Alto, CA, 1989. [5] J. Gibson. Adaptation, aftereffect and contrast in the perception of curved lines. J. Exp. Psychol., 16:1-16, 1933. [6] J. G. Harris. Continuous-time calibration of VLSI sensors for gain and offset variations. In SPIE Internatiional Symposium on Aerospace Sensing and DualUse Photonices:Smart Focal Plane Arrays and Focal Plane Array Testing, pages 23-33, Orlando, FL, April 1995. [7] C. Mead. Adaptive retina. In C. Mead and M. Ismail, editors, Analog VLSI Implementation of Neural Systems, pages 239-246. Kluwer Academic Publishers, 1989. [8] C. Mead. Analog VLSI and Neural Systems. Addison-Wesley, 1989. [9] D.A. Scribner, K.A. Sarkady, M.R. Kruer, J.T. Calufield, J.D. Hunt, M. Colbert, and M. Descour. Adaptive retina-like preprocessing for imaging detector arrays. In Proc. of the IEEE International Conference on Neural Networks, pages 1955-1960, San Francisco, CA, Feb. 1993. [10] S. Ullman and G. Schechtman. Adaptation and gain normalization. Pmc. R. Soc. Lond. B, 216:299-313, 1982.
1996
28
1,223
A neural model of visual contour integration Zhaoping Li Computer Science, Hong Kong University of Science and Technology Clear Water Bay, Hong Kong zhaoping~uxmail.ust.hkl Abstract We introduce a neurobiologically plausible model of contour integration from visual inputs of individual oriented edges. The model is composed of interacting excitatory neurons and inhibitory interneurons, receives visual inputs via oriented receptive fields (RFs) like those in VI. The RF centers are distributed in space. At each location, a finite number of cells tuned to orientations spanning 1800 compose a model hypercolumn. Cortical interactions modify neural activities produced by visual inputs, selectively amplifying activities for edge elements belonging to smooth input contours. Elements within one contour produce synchronized neural activities. We show analytically and empirically that contour enhancement and neural synchrony increase with contour length, smoothness and closure, as observed experimentally. This model gives testable predictions, and in addition, introduces a feedback mechanism allowing higher visual centers to enhance, suppress, and segment contours. 1. Introduction The visual system must group local elements in its input into meaningful global features to infer the visual objects in the scene. Sometimes local features group into regions, as in texture segmentation; at other times they group into contours which may represent object boundaries. Although much is known about the processing steps that extract local features such as oriented input edges, it is still unclear how local features are grouped into global ones more meaningful for objects. In this 1r would very much like to thank Jochen Braun for introducing me to the topic, and Peter Dayan for many helpful conversations and comments on the drafts. This work was supported by the Hong Kong Research Grant Council. 70 Z. Li study, we model the neural mechanisms underlying the grouping of edge elements into contours contour integration. Recent psychophysical and physiological observations[14, 8] demonstrate a decrease in detection threshold of an edge element, by human observers or a primary cortical cell, if there are aligned neighboring edge elements. Changes in neural responses by visual stimuli presented outside their RFs have been observed physiologically[9, 8]. Human observers easily identify a smooth curve composed of individual, even disconnected, Gabor "edge" elements distributed among many similar elements scattered in the background[4]. Horizontal neural connections observed in the primary visual cortex[5], and the finding that these connections preferably link cells tuned to similar orienations[5], provide a likely neural basis underlying the primitive visual grouping phenomena such as contour integration. These findings suggest that simple and local neural interactions even in VI could contribute to grouping. However, it has been difficult to model contour integration using only VI elements and operation. Most existing models[15, 18] of contour integration lack well-founded biological bases. More neurally based models, e.g., the one by Grossberg and Mingolla[7], require operations beyond VI or biologically questionable. It is thus desirable to find out whether contour enhancement can indeed occur within VI or has to be attributed to top-down feedback. We introduce a VI model of contour integration, using orientation selective cells, local cortical circuits, and horizontal connections. This model captures the essentials of the contour integration behavior. More details of the model can be found in a longer paper[12]. 2. The Model 2.1 Model outline K neuron pairs at each spatial location i model a hypercolumn in VI (figure 1). Each neuron has a receptive field center i and an optimal orientation fJ = k1r / K for k = 1,2, ... K. A neuron pair consist of a connected excitatory neuron and inhibitory neuron which are denoted by indice (ifJ) for their receptive field center and preferred orientation, and are referred to as an edge segment. An edge segment receives the visual input via the excitatory cell, whose output quantifies the saliency of the edge segment and projects to higher visual centers. The inhibitory cells are treated as interneurons. When an input image contains an edge at i oriented at fJo, the edge segment ifJ receives input Ii9 (X </J(fJ fJo ), where </J(fJ) = e- 191/(1I"/8) is a cell's orientation tuning curve. Visual space, bypen:oluDlDS and neural edge segmenos ~~~ ~ ~--~ ~ ~:~j .. ~ ~~~~ A hype<oolumn A oeuraI edge atlocalion i segment i8 excitatory An intercon neuron pair (ClI' edge segment ill Edge outputs to higher visual areas Figure 1: Model visual space, hypercolumn, edge segments, neural elements, visual inputs, and neural connections. The input space is a discrete hexagonal or Manhatten grid. A Neural Model o/Visual Contour Integration Model visual input Model output - , , .... ,...... " 1\1'\-\' '~ 1\, r./,,,I\ \', , II.... " ............... , \ I , ... ,I A (I - ,1 ... \ "'....'.,!' ... ,' '..{... I, I, 71 Output after removing ed,ges of activities lower than 1/"1. of the most active edge I "\ . ............. . \.. I Figure 2: Model neural connections and performance for contour enhancement and noise reduction. The top graph depicts connections J;(J,j(J1 and W;(J,j(J1 respectively between the center (white) horizontal edge and other edges. The whiteness and blackness of each edge is proportional to the connection sthength Ji(J,j(J1 and W;(J,j(J1 respectively. The bottom row plots visual input and the maximum model outputs. The edge thickness in this row is proportional to the value of edge input or activity. The same format applies to the other figures in this paper. Let Xifl and YiO be the cell membrane potentials for the excitatory and inhibitory cells respectively in the edge segment, then YiO -axXiO 9y(YiO) + J 09x(XiO) + L J iOj OI9x(XjOI) + Iio + 10 jO'#iO -ayYio + 9x(XiO) + L WiOjOI9x(XjO') + Ie jO'#iO (1) (2) where 9x(XiO) and 9y(YiO) are the firing rates from the excitatory and inhibitory cells respectively, l/ax and l/ay are the membrane constants, Jo is the self-excitatory connection weight, JiO ,jO' and WiO,jOI are synaptic weights between neurons, and 10 and Ie are the background inputs to the excitatory and inhibitory cells. Without loss of generality, we take ax = a y = 1, 9x(X) as threshold linear with saturation and an unit gain 9~(X) = 1 in the linear range. The synaptic connections JiO,jO' and Wi(},jOI are local and translation and rotation invariant (Figure (2)). JiO,jO' increases with the smoothness (small curvature) of the curve that best connects (if}) and (jf}') , and edge elements inhibit each other via WiO,jOI when they are alternative choices in a smooth curve route. Given an input pattern lUI, the network approaches a dynamic state after several membrane time constants. As in Figure (2), the neurons with relatively higher final activities are those belonging to smooth curves in the input. 2.2 Model analysis Ignoring neural connections between edge segments, the neuron in edge segment if} 72 has input sensitivity 1 + g~(Yo)g~(xo) Jog~(xo) g~(Yo)g~(xo) Z. Li (3) (4) where gx(xo) and gy(yo) are roughly the average neural activities (omitting iO for simplicity). Thus the edge activity increases with Ii9 and decreases with Ie (in cases that interest us, g~(Yo)g~(xo) > Jog~(xo) -1). The resulting input-output function given Ie, gx(xo) vs. Ii9 , corresponds well with physiological data. By effectively increasing IiO or Ie, the edge element (jO') can excite or inhibit the element (iO) with excitatory-to-excitatory input Ji9,j9,gx(Xj9 1 ) and excitatory-toinhibitory input WiOj9' gx(Xj9') respectively. Contour enhancement is so (Fig. 2) achieved. In the simplest example when the visual input has equally spaced equal strength edges from a line and all other edge segments are silent, we can treat the line system as one dimensional, omit 0 and take i as locations along the line. A lack of inhibition between line segments gives: -axxi - gy(Yi) + Jogx(Xi) + L Jijgx(Xj) + 10 + Iline-input #i (5) (6) If line is infinite, by symmetry, each edge segment has the same average activity gx(Xi) rv gx(xo) for all i. This system can then be seen[12] either as a giant edge with self-excitatory connection (Jo + 2::#i Jij ), or a single edge with extra external input t1I = (2::#i Jij)gx(xo). Either way, activities gx(xo) are enhanced for each edge element in the line (figure 3 and 2). This analysis is also applicable to constant curvature curves[12]. It can be shown that, in the linear range of gx 0, the response ratio between a curve segment and an isolated segment is (g~(yo) + 1- Jo)/(g~(yo) + 1 - Jo - 2::i#j Jij ). Since 2::ih Jij decreases with increasing curvature, so does the response enhancement. Translation invariance along the curve breaks down in a finite length curve near its two ends, where activity enhancement decays by a decreased excitation from fewer neighboring segments. This suggests that a closed or longer curve has higher saliency than an open or shorter one (figure (3)). This prediction is expected to hold also for curves of non-constant curvature, and should playa significant role in the psychophysical observation[lO] showing a decreased detection threshold for closed curves from that of the open ones. Further analysis[12] shows that the edge segments in a curve normally exhibit neural oscillations around their mean activity levels with near-zero phase delays from each other. The model predicts that, like the contour enhancement, the oscillation is stronger for longer, smoother, and closed curves than open and shorter ones, and tapers off near curve endings where oscillation synchrony also deteriorates (figure (3)). 2.3 Central feedback control for contour enhancement, snpression, filling in, and segmentation A Neural Model o/VlSual Contour Integration -I \ ,I \1-/ .......... # Model visual inputs / " I \ -I , 1 \ I ,-I Model outputs /#- .... \ -I , \ . -I # / 1 Neural signals in time -I , I, I,' I,' I, " I' I, " i l it i. i' I" \: I:: I" I.. !, U~fi9IllP5lll1 j" I .. ( \ \ ) , , " \ ( \ \ ) i:, .' i:, i:, i:, 1"f--L---;-1-..-_.--;;--.J I,ol--' .L.-:-=----_---,,--;r-----J I'ol--' -'---,,..-_----;-----;;r-----J I,ol--' .L.-,---o-----;;--;---' 73 Figure 3: Model performance for input curves and noises, Each column is dedicated to one input condition. The top row is the visual input; the middle row is the maximum neural responses, and the bottom row the segment outputs as a function of time. The neural signals in the bottom row are shown superposed. The solid curves plot outputs for the segments away from the curve endings, the dash-dotted curves for segments near the curve endings, and the dashed curve in each plot, usually the lowest lying one, is an example from a noise segment. Note the decrease in neural signal synchrony between the curve segments, in the oscillation amplitudes, and average neural activities for the curve segments, as the curve becomes open, shorter, and more curled or when the segment is near the curve endings. Because the model employs discrete a grid input space, the figure '8' curve in the right column is almost not a smooth curve, hence the contour is only very weakly enhanced. The enhancement for this curve could be increased by a denser input grid or a multiscale input space. The model assumes that higher visual centers send inputs Ie to the inhibitory cells, to influence neural activities by equation (4). Take Ie = Ie,baekground+Ie,eontrol ~ 0, where Ie,baekground is the same for all edge segments and is useful for modulating the overall visual alertness level. By setting Ie,eontrol differently for different edge segments, higher centers can selectively suppress or enhance activities for some visual objects (contours) (compare figure 4D,G), and even effectively achieve contour segmentation (figure (4H)) by silencing segments from a given curve. A similar feedback control mechanism was used in an olfactory model for odor sensitivity control and odor segmentation[ll]. Nevertheless, the feedback cannot completely substitute for the visual input Iii}. By equation (4), Ie is effective only when g~(Yo)g~(xo) "# 0. With insufficient background input 10 , some visual input or excitation from other segments is needed to have g~(xo) > 0, even when all the inhibition from Ie is removed by central control. However, a segment with weak visual input or excitation from aligned neighboring segments can increase its activity or become active from subthreshold. Therefore, this model central feedback can enhance a weak contour or fill in an incomplete one (compare figure 4F ,I), but can not enhance or "hallucinate" any contour not existing at least partially in the visual input I (figure 4E), 3. Summary and Discussion We have presented a model which accounts for phenomena of contour enhance74 A: Input B: Input ;:\ - ~ '\ , , \ I / c: Input Z.Li D: Outpl,lt for A no central control G.: Output for A, line suppresion circle enhancement I:' ':x I~-:-·,\ .... , \ -..-+-a-------a-'\-',) " \. '.1 , .. :-.E: Qutout for B, enh~cement H: Same.as G, except with fOr clrcre and non-exlstmg line stronger line suppreslOn I~-:-·,\ \ ~ .,' . '. '.1 F: Output f~r C no central control I~'-·,\ \ , • I • '. '.1 I; Outp~t.for-C, en):lajlcement lOr both hne and circle " ''I. I~'-·~ ,I ... \ :\ I " ... , ' ... , ... \ \ \ 7~\ -'--.- -7-,-'.-; ~·\·~T ... -... -... . .... I .,. , .. I .. Figure 4: Central feedback control. The visual inputs are in A, B and C. The rest are the model responses under different feedback conditions. D: model response for input A without central control. G: model response for input A with line suppression by Ic,control = O.25Ic ,background on the line segments, and circle enhancement by Ic,control = -O.29Ic ,background on the circle elements. H: Same as G except the line suppression signal Ic,control is doubled. E: Model response to input B with enhancement for the line and the circle by central feedback Ic,control = -O.29Ic ,background' F: Response to input C with no central control. I: Response to input C with line and circle enhancement by Ic ,control = -O.29Ic ,background' Note how the input gaps are partiallly filled in F and almost completely filled in I. Note that the apparent gaps in the circle are caused by the underlying discrete grid in the model input space, and hence no gap actually exists, and no filling in is needed, for this circle. Also, with the wrap around boundary condition, the line is actually a closed or infinitely long line, and is thus naturally more salient in this model without feedback. ment using neurally plausible elements in VI. It is shown analytically and empirically that both the contour enhancement and the neural oscillation amplitudes are stronger for longer, closed, and smaller curvature curves, agreeing with experimental observations [8, 4, 10, 6, 2]. The model predicts that horizontal connections target preferentially excitatory or inhibitory post-synaptic cells when the linked edges are aligned or less aligned (Fig. 2). In addition, we introduce a possible feedback mechanism by which higher visual centers could selectively enhance or suppress contour activities, and achieve contour segmentation. This feedback mechanism has the desirable property that while the higher centers can enhance or complete an existing weak and/or fragmented input contour, they cannot enhance a non-exist ant contour in the input, thus preventing "hallucination". This property could be exploited by higher visual centers for hypothesis testing and object reconstruction by cooperating with lower visual centers. Analogous computational mechanisms have been used in an olfactory model to achieve odor segmentation and sensitivity modulation[ll]. It will be interesting to explore the universality of such computational mechanisms across sensory modalities. The organization of the model is based on various experimental finding[17, 5, 1, 8, 14]: recurrent excitatory-inhibitory interactions; excitatory and inhibitory linking of edge elements with similar orientation preferences; and neural connection A Neural Model of Visual Contour Integration 75 patterns. At the cost of analytical tractability without essential changes in model performance, one can relax the model's idealization of a 1:1 ratio in the excitatory and inhibitory cell numbers, the lack of connections between the inhibitory cells, and the excitatory cells as the exclusive recipients of visual input. While abundant feedback connections are observed from higher visual centers to the primary visual cortex, there is as yet no clear indication of cell types of their targets [3, 16]. It is desirable to find out whether the feedback is indeed directed to the inhibitory cells as predicted. This model can be extended to stereo, temporal, and chromatic dimensions, by linking edge segments aligned in orientation, depth, motion direction and color. V1 cells have receptive field tuning in all these dimensions, and cortical connections are indeed observed to link cells of similar receptive field properties[5]. This model does not model many other apparently non-contour related visual phenomena such as receptive field adaptations[5]. It is also beyond this model to explain how the higher visual centers decide which segments belong to one contour in order to achieve feedback control, although it has been hypothesized that phase locked neural oscillations and neural correlations can play such a role[13]. References [1] Douglas R.J. and Martin K. A. "Neocortex" in Synaptic Organization of the Brain 3rd Edition, Ed. G. M. Shepherd, Oxford University Press 1990. [2] Eckhorn R, Bauer, R. Jordan W. Brosch, M. Kruse W., Munk M. and Reitboeck, H. J. 1988. Bioi. Cybern. 60:121-130. [3] van Essen D. Peters and E G Jones, Plenum Press, New York, 1985. p. 259-329 [4] Field DJ; Hayes A; Hess RF 1993. Vision Res. 1993 Jan; 33(2): 173-93 [5] Gilbert CD Neuron. 1992 Julj 9(1): 1-13 [6] Gray C.M. and Singer W. 1989 Proc. Natl. Acad. Sci. USA 86: 1698-1702. [7] Grossberg S; Mingolla E Percept Psychophys. 1985 Aug; 38(2): 141-71 [8] Kapadia MKj Ito M; Gilbert CD; Westheimer G, Neuron. 1995 Oct; 15(4): 843-56 [9] Knierim J. J. and van Essen D. C. 1992, J. Neurophysiol. 67, 961-980. [10] Kovacs I; Julesz B Proc Natl Acad Sci USA . 1993 Aug 15; 90(16): 7495-7 [11] Li Zhaoping, Biological Cybernetics, 62/4 (1990), P. 349-361 [12] Li Zhaoping, "A neural model contour integration in primary visual cortex" manuscript submitted for publication. [13] von der Malsburg C. 1981 "The correlation theory of brain function." Internal report, Max-Planck-Institute for Biophysical Chemistry, Gottingen, West Germany. [14] Polat U; Sagi D, 1994 Vision Res. 1994 Jan; 34(1): 73-8 [15] Shashua A. and Ullman S. 1988 Proceedings of the International Conference on Computer Vision. Tempa, Florida, 482-488. [16] Valverde F. in Cerebral Cortex Eds. A Peters and E G Jones, Plenum Press, New York, 1985. p. 207-258. [17] White E. L. Cortical circuits 46-82, Birkhauser, Boston, 1989 [18] Zucker S. W., David C., Dobbins A, and Iverson L. in Second international conference on computer vision pp. 568-577, IEEE computer society press, 1988.
1996
29
1,224
ARTEX: A Self-Organizing Architecture for Classifying Image Regions Stephen Grossberg and James R. Williamson {steve, jrw}@cns.bu.edu Center for Adaptive Systems and Department of Cognitive and Neural Systems Boston University 677 Beacon Street, Boston, MA 02215 Abstract A self-organizing architecture is developed for image region classification. The system consists of a preprocessor that utilizes multiscale filtering, competition, cooperation, and diffusion to compute a vector of image boundary and surface properties, notably texture and brightness properties. This vector inputs to a system that incrementally learns noisy multidimensional mappings and their probabilities. The architecture is applied to difficult real-world image classification problems, including classification of synthetic aperture radar and natural texture images, and outperforms a recent state-of-the-art system at classifying natural textures. 1 INTRODUCTION Automatic processing of visual scenes often begins by detecting regions of an image with common values of simple local features, such as texture, and mapping the pattern offeature activation into a predicted region label. We develop a self-organizing neural architecture, called the ARTEX algorithm, for automatically extracting a novel and effective array of such features and mapping them to output region labels. ARTEX is made up of biologically motivated networks, the Boundary Contour System and Feature Contour System (BCS/FCS) networks for visual feature extraction (Cohen & Grossberg, 1984; Grossberg & Mingolla, 1985a, 1985b; Grossberg & Todorovic, 1988; Grossberg, Mingolla, & Williamson, 1995), and the Gaussian ARTMAP (GAM) network for classification (Williamson, 1996). ARTEX is first evaluated on a difficult real-world task, classifying regions of synthetic aperture radar (SAR) images, where it reliably achieves high resolution (single 874 S. Grossberg and 1. R. Williamson pixel) classification results, and creates accurate probability maps for its class predictions. ARTEX is then evaluated on classification of natural textures, where it outperforms the texture classification system in Greenspan, Goodman, Chellappa, & Anderson (1994) using comparable preprocessing and training conditions. 2 FEATURE EXTRACTION NETWORKS Filled-in surface brightness. Regions of interest in an image can often be segmented based on first-order differences in pixel intensity. An improvement over raw pixel intensities can be obtained by compensating for variable illumination of the image to yield a local brightness feature. A further improvement over local brightness features can be obtained with a surface brightness feature, which is obtained by smoothing local brightness values when they belong to the same region, while maintaining differences when they belong to different regions. Such a procedure tends to maximize the separability of different regions in brightness space by minimizing within-region variance while maximizing between-region variance. In Grossberg et al. (1995) a multiple-scale BCS/FCS network was used to process noisy SAR images for use by human operators by normalizing and segmenting the SAR intensity distributions and using these transformed data to fill-in surface representations that smooth over noise while maintaining informative structures. The single-scale BCS/FCS used here employs the middle-scale BCS/FCS used in that study. The BCS/FCS equations and parameters are fully described in Grossberg et al. (1995). The BCS/FCS is herein applied to SAR images that are spatially consolidated to half the size (in each dimension) of the images used in that study, and so is comparable to the large-scale BCS/FCS used there. Multiple-scale oriented contrast. In addition to surface brightness, another image property that is useful for region segmentation is texture. One popular approach for analyzing texture, for which there is a great deal of supporting biological and computational evidence, decomposes an image, at each image location, into a set of energy measures at different oriented spatial frequencies. This may be done by applying a bank of orientation-selective bandpass filters followed by simple nonlinearities and spatial pooling, to extract multiple-scale oriented contrast features. The early stages of the BCS, which define a Static Oriented Constrast (or SOC) filtering network, carry out these operations, and variants of them have been used in many texture segregation algorithms (Bergen, 1991; Greenspan et al., 1994). Here, the SOC network produces K = 4 oriented contrast features at each of four spatial scales. The first stage of the SOC network is a shunting on-center off-surround network that compensates for variable illumination, normalizes, and computes ratio contrasts in the image. Given an input image, I, the output at pixel (i,i) and scale 9 in the first stage of the SOC network is 9 _ Iij - (Gg * I)ij - DE aij D + Iij + (Gg * I)ij , ( 1) where E = 0.5, and Gg is a Gaussian kernel defined by Gfj(p, q) = -2 1 2 exp[-«i - p)2 + (j - q)2)/20";], (2) 'frO" 9 with O"g = 2g , for the spatial scales 9 = 0,1,2,3. The value of D is determined by the range of pixel intensities in the input image. We use D=2000 for SAR images and D = 255 for natural texture images. The next stage obtains a local measure of orientational contrast by convolving the output of (1) with Gabor filters, H!, which ARTEX: A Self-organizing Architecture/or Classifying Inwge Regions 875 are defined at four orientations, and then full-wave rectifying the result: bfjk = I(HZ * ag)ij I· (3) The horizontal Gabor filter (k = 0) is defined by: HfJo(p, q) = Grj(p, q) . sin[0.757r(j - q)/ug] . (4) Orientational contrast responses may exhibit high spatial variability. A smooth, reliable measure of orientational contrast is obtained by spatially pooling the responses within the same orientation: e!jk = (Gg * bt)ij. (5) Equation (5) yields an orientationaliy variant, or OV, representation of oriented contrast. A further optional stage yields an orientationaliy invariant, or 01, representation by shifting the oriented responses at each scale into a canonical ordering, to yield a common representation for rotated versions of the same texture: d!jk = erjkl where k' = [k + arg ~~ (cfjk ll )] mod K. (6) 3 CLASSIFICATION NETWORK GAM is a constructive, incremental-learning network which self-organizes internal category nodes that learn a Gaussian mixture model of the M-dimensional input space, as well as mappings to output class labels. Here, mappings are learned from 17 -dimensional input vectors (composed of a filled-in brightness feature and 16 oriented contrast features) to a class label representing a shadow, road, grass, or tree region. The ph category's receptive field is parametrized by two M-dimensional vectors: its mean, Pj, and standard deviation, ifj . A scalar, nj, also represents the node's cumulative credit. Category j is activated only if its match, Gj , satisfies the match criterion, which is determined by a vigilance parameter, p. Match is a measure, obtained from the category's unit-height Gaussian distribution, of how close an input, X, is to the category's mean, relative to its standard deviation: ( 1~(Xi-l'"i)2) Gj = exp -- LJ • 2 i=l Uji (7) The match criterion is a threshold: the category is activated only if Gj > Pi otherwise, the category is reset. The input strength, gj, is determined by n' gj = M J Gj if Gj > Pi gj = 0 otherwise. (8) TIi=l Uji The category's activation, Yj, which represents P(jlx), is obtained by g' Y. _ J J N' D + Ll=l g, (9) where N is the number of categories and D is a shunting decay term that maintains sensitivity to the input magnitude in the activation level (D = 0.01 here). When category j is first chosen, it learns a permanent mapping to the output class, k, associated with the current training sample. All categories that map to the same class prediction belong to the same ensemble: j E E(k). Each time an input is presented, the categories in each ensemble sum their activations to generate a net probability estimate, Zk, of the class prediction k that they share: Zk = L Yj· jEE(k) (10) 876 S. Grossberg and 1. R. Williamson The system prediction, }(, is determined by the maximum probability estimate, K = argmax(zk), (11) k which determines the chosen ensemble. Once the class prediction K is chosen, we obtain the category's "chosen-ensemble" activation, Y;, which represents P(jlx, K): Y; = L: Yj if j E E(K); Y; = 0 otherwise. (12) lEE(K) Yl If K is the correct prediction, then the network resonates and learns; otherwise, match tracking is invoked: p is raised to the average match of the chosen ensemble. p = exp (-~ L Y; t (Xi ; .:ji) 2) . jEE(K) i=l l' (13) In addition, all categories in the chosen ensemble are reset. Equations (8)-(11) are then re-evaluated. Based on the remaining non-reset categories, a new prediction K in (11), and its corresponding ensemble, are chosen. This automatic search cycle continues until the correct prediction is made, or until all committed categories are reset and an uncommitted category is chosen. Upon presentation of the next training sample, p is reassigned its baseline value: p = p. Here, p ~ O. When category j learns, nj is updated to represent the amount of training data the node has been assigned credit for: nj := nj + Y; • (14) The vectors flj and iij are then updated to learn the input statistics: I'ji (1 • -1) +.-1 Yj nj I'ji Yj nj Xi, (15) (16) GAM is initialized with N = O. When a category is first chosen, N is incremented, and the new category, indexed by J =N, is initialized with nJ = 1, fl = X, G'ji =;, and with a permanent mapping to the correct output class. Initializing G'ji = ; is necessary to make (7) and (8) well-defined. Varying; has a marked effect on learning: as ; is raised, learning becomes slower, but fewer categories are created. The input vectors are normalized to have the same standard deviation in each dimension so that; has the same meaning in each dimension. 4 SIMULATION RESULTS Classifying SAR image regions. Figure 1 illustrates the classification results obtained on one SAR image after training on the other eight images in the data set. The final classification result (bottom, right) closely resembles the hand-labeled regions (middle, left). The caption summarizes the average results obtained on all nine images. ARTEX learns this problem very quickly, using a small number of self-organized categories, as shown in Figure 2 (left). The best classification result of 84.2% correct is obtained by filling-in the probability estimates from equation (10) within the BCS boundaries, using an FCS diffusion equation as described in Grossberg et al. (1995). These filled-in probability estimates predict the actual classification rates with remarkable accuracy (Figure 2, right). Classifying natural textures. ARTEX performance is now compared to that of a texture analysis system described in Greenspan et al. (1994), which we refer to as the "hybrid system" because it is a hybrid architecture made up of a ART EX: A Self-organizing Architecture/or Classifying Image Regions 877 Figure 1: Results are shown on a 180x180 pixel SAR image, which is one of nine images in data set. Top row: Center/surround, first stage output (left); BCS boundaries to FCS filling-in (middle); final BCS/FCS filled-in output (right). Note that BCS accurately localizes region boundaries, and that FCS improves appearance by smoothing intensities within regions while maintaining sharp differences between regions. Middle row: Hand-labeled regions corresponding to shadow, road, grass, trees (left); Gaussian classifier results based on center/surround feature (middle, 59.6% correct), and based on filled-in feature (right, 70.7%). Note that fillingin greatly improves classification by reducing brightness variability within regions. However, the lack of textural information results in errors, such as the misclassification of the vertical road as a shadow region. Bottom row: GAM results (1' = 4) based on 16 SOC features in addition to the filled-in brightness feature: using the OV representation (left, 81.9%), using the 01 representation (middle, 83.2%), and using filled-in 01 prediction probabilities (right, 84.2%). With the OV representation (bottom, left), the thin vertical road is misclassified as shadows because there are no thin vertical roads in the training set. With the 01 representation, however (bottom, middle), the road is classified correctly because the training set includes thin roads at other orientations. Finally, the classification results are improved by filling-in the prediction probabilities from equation (10) within the BCS boundaries, thereby taking advantage of spatial and structural context (bottom, right). 878 s. Grossberg and J. R. Williamson '00 .. .. 50 50 '00 '50 ... '"'" 03 0 .4 0.5 05 0.7 0.8 0.1 Number of Categories Filled-in Probability Figure 2: Left: classification rate is plotted as a function of the number of categories after training on different sized subsets of the SAR training data: (left-to-right) 0.01 %, 0.1%, 1%, 10%, and 100% of the training set. Right: classification rate is plotted as a function of filled-in probability estimates. log-Gabor pyramid representation, followed by unsupervised k-means clustering in the feature space, followed by batch learning of mappings from clusters to output classes using a rule-based classifier. The hybrid system uses three pyramid levels and four orientations at each level. Each level of the pyramid is produced via three blurring/decimation steps, resulting in an 8x8 pixel resolution. For a fair comparison, sufficient blurring/decimation was added as a postprocessing step to ARTEX features to yield the same net amount of blurring. Both ARTEX and the hybrid system use an OV representation for these problems because the textures are not rotated. The first task is classification of a library of ten separate structured and unstructured textures after training on different example images. ARTEX obtains better performance, achieving 96.3% correct after 40 training epochs (with i = 1, 34 categories) versus 94.3% for the hybrid system. Even after only one training epoch, ARTEX achieves better results (94.9%, 23 categories). The second task (Figure 3) is classification of a five-texture mosaic, which requires discriminating texture boundaries, after training on examples of the five textures, plus an additional texture (sand). ARTEX achieves 93.6% correct after 40 training epochs (33 categories), and produces results which appear to be better than those produced by the hybrid system on a similar problem (see Greenspan et al., 1994, Figure 5). In summary, the ARTEX system demonstrates the utility of combining BCS texture and FCS brightness measures for image preprocessing. These features may be effectively classified by the GAM network, whose self-calibrating matching and search operations enable it to carry out fast, incremental, distributed learning of recognition categories and their probabilities. BCS boundaries may be further used to constrain the diffusion of these probabilities according to FCS rules to improve prediction probability. Acknowledgements Stephen Grossberg was supported by the Office of Naval Research (ONR NOOOl495-1-0409 and ONR NOOOl4-95-1-0657). James Williamson was supported by the Advanced Research Projects Agency (ONR NOOOl4-92-J-4015), the Air Force Office of Scientific Research (AFOSR F49620-92-J-0225 and AFOSR F49620-92-J-0334), the National Science Foundation (NSF IRI-90-00530 and NSF IRI-90-24877), and ARTEX: A Self-organizing Architecturejor Classifying Image Regions 879 Figure 3: Left: mosaic of five natural textures. Right: ARTEX classification (93.6% correct) after training on examples of five textures and an additional texture (sand). the Office of Naval Research (ONR NOOOI4-91-J-4100 and ONR NOOOI4-95-1-0409). References Bergen, J.R. ''Theories of visual texture perception," in Spatial Vision, D. M. Regan Ed. New York: Macmillan, 1991, pp. 114-134. Cohen, M. & Grossberg, S., (1984). Neural dynamics of brightness perception: Features, boundaries, diffusion, and resonance. Perception £3 Psychophysics, 36, 428-456. Greenspan, H., Goodman, R., Chellappa, R., & Anderson, C.H. (1994). Learning texture discrimination rules in a multiresolution system. IEEE Trans. PAMI, 16, 894-90l. Grossberg, S. & Mingolla, E. (1985a). Neural dynamicsofform perception: Boundary completion, illusory figures, and neon color spreading. Psychological Review, 92, 173-211. Grossberg, S. & Mingolla, E. (1985b). Neural dynamics of perceptual grouping: Textures, boundaries, and emergent segmentations. Perception £3 Psychophysics, 38, 141-17l. Grossberg, S., Mingolla, E., & Williamson, J. (1995). Synthetic aperture radar processing by a multiple scale neural system for boundary and surface representation. Neural Networks, 8, 1005-1028. Grossberg, S. & Todorovic, D. (1988). Neural dynamics of I-D and 2-D brightness perception: A unified model of classical and recent phenomena. Perception £3 Psychophysics, 43,241-277. Grossberg, S., & Williamson, J.R. (1996). A self-organizing system for classifying complex images: Natural textures and synthetic aperture radar. Technical Report CAS/CNS TR-96-002, Boston, MA: Boston University. Williamson, J .R. (1996). Gaussian ARTMAP: A neural network for fast incremental learning of noisy multidimensional maps. Neural Networks, 9, 881-897.
1996
3
1,225
Unsupervised Learning by Convex and Conic Coding D. D. Lee and H. S. Seung Bell Laboratories, Lucent Technologies Murray Hill, NJ 07974 {ddlee I seung}Obell-labs. com Abstract Unsupervised learning algorithms based on convex and conic encoders are proposed. The encoders find the closest convex or conic combination of basis vectors to the input. The learning algorithms produce basis vectors that minimize the reconstruction error of the encoders. The convex algorithm develops locally linear models of the input, while the conic algorithm discovers features. Both algorithms are used to model handwritten digits and compared with vector quantization and principal component analysis. The neural network implementations involve feedback connections that project a reconstruction back to the input layer. 1 Introduction Vector quantization (VQ) and principal component analysis (peA) are two widely used unsupervised learning algorithms, based on two fundamentally different ways of encoding data. In VQ, the input is encoded as the index of the closest prototype stored in memory. In peA, the input is encoded as the coefficients of a linear superposition of a set of basis vectors. VQ can capture nonlinear structure in input data, but is weak because of its highly localized or "grandmother neuron" representation. Many prototypes are typically required to adequately represent the input data when the number of dimensions is large. On the other hand, peA uses a distributed representation so it needs only a small number of basis vectors to model the input. Unfortunately, it can only model linear structures. Learning algorithms based on convex and conic encoders are introduced here. These encoders are less constrained than VQ but more constrained than peA. As a result, 516 D. D. Lee and H. S. Seung Figure 1: The affine, convex, and conic hulls for two basis vectors. they are able to produce sparse distributed representations that are efficient to compute. The resulting learning algorithms can be understood as approximate matrix factorizations and can also be implemented as neural networks with feedforward and feedback connections between neurons. 2 Affine, convex, conic, and point encoding Given a set of basis vectors {wa}, the linear combination L::=1 vawa is called affine } con~ex comc if { Va:::?: 0 , Va:::?: 0 . L:a Va = 1 , L:a Va = 1 , The complete set of affine, convex, and conic combinations are called respectively the affine, convex, and conic hulls of the basis. These hulls are geometrically depicted in Figure 1. The convex hull contains only interpolations of the basis vectors, whereas the affine hull contains not only the convex hull but also linear extrapolations. The conic hull also contains the convex hull but is not constrained to stay within the set La Va = 1. It extends to any nonnegative combination of the basis vectors and forms a cone in the vector space. Four encoders are considered in this paper. The convex and conic encoders are novel, and find the nearest point to the input x in the convex and conic hull of the basis vectors. These encoders are compared with the well-known affine and point encoders. The affine encoder finds the nearest point to x in the affine hull and is equivalent to the encoding in peA, while the point encoder or VQ finds the nearest basis vector to the input. All of these encoders minimize the reconstruction error: 2 (1) The constraints on Va for the convex, conic, and affine encoders were described above. Point encoding can be thought of as a heavily constrained optimization of Eq. (1): a single Va must equal unity while all the rest vanish. Efficient algorithms exist for computing all of these encodings. The affine and point encoders are the fastest. Affine encoding is simply a linear transformation of the input vector. Point encoding is a nonlinear operation, but is computationally simple Unsupervised Learning by Convex and Conic Coding 517 since it involves only a minimum distance computation. The convex and conic encoders require solving a quadratic programming problem. These encodings are more computationally demanding than the affine and point encodingsj nevertheless, polynomial time algorithms do exist. The tractability of these problems is related to the fact that the cost function in Eq. (1) has no local minima on the convex domains in question. These encodings should be contrasted with computationally inefficient ones. A natural modification of the point encoder with combinatorial expressiveness can be obtained by allowing v to be any vector of zeros and ones [1, 2]. Unfortunately, with this constraint the optimization of Eq. (1) becomes an integer programming problem and is quite inefficient to solve. The convex and conic encodings of an input generally contain coefficients Va that vanish, due to the nonnegativity constraints in the optimization of Eq. (1). This method of obtaining sparse encodings is distinct from the method of simply truncating a linear combination by discarding small coefficients [3]. 3 Learning There correspond learning algorithms for each of the encoders described above that minimize the average reconstruction error over an ensemble of inputs. If a training set of m examples is arranged as the columns of a N x m matrix X, then the learning and encoding minimization can be expressed as: min liX - WVI1 2 W,V (2) where IIXII 2 is the summed squares of the elements of X. Learning and encoding can thus be described as the approximate factorization of the data matrix X into a N x r matrix W of r basis vectors and a r X m matrix V of m code vectors. Assuming that the input vectors in X have been scaled to the range [0,1], the constraints on the optimizations in Eq. (2) are given by: Affine: o ~ Wia ~ 1, Convex: o ~ Wia ~ 1, Conic: o ~ Wia ~ 1, The nonnegativity constraints on W and V prevent cancellations from occurring in the linear combinations, and their importance will be seen shortly. The upper bound on W is chosen such that the basis vectors are normalized in the same range as the inputs X. We noted earlier that the computation for encoding is tractable since the cost function Eq. (2) is a quadratic function of V. However, when considered as a function of both W and V, the cost function is quartic and finding its global minimum for learning can be very difficult. The issue of local minima is discussed in the following example. 4 Example: modeling handwritten digits We applied Affine, Convex, Conic, and VQ learning to the USPS database [4], which consists of examples of haI!dwritten digits segmented from actual zip codes. Each of the 7291 training and 2007 test images were normalized to a 16 x 16 grid 518 D. D. Lee and H. S. Seung VQ Affine (PeA) Convex Conic Figure 2: Basis vectors for "2" found by VQ, Affine, Convex, and Conic learning. with pixel intensities in the range [0,1]. There were noticeable segmentation errors resulting in unrecognizable digits, but these images were left in both the training and test sets. The training examples were segregated by digit class and separate basis vectors were trained for each of the classes using the four encodings. Figure 2 shows our results for the digit class "2" with r = 25 basis vectors. The k-means algorithm was used to find the VQ basis vectors shown in Figure 2. Because the encoding is over a discontinuous and highly constrained space, there exist many local minima to Eq. (2). In order to deal with this problem, the algorithm was restarted with various initial conditions and the best solution was chosen. The resulting basis vectors look like "2" templates and are blurry because each basis vector is the mean of a large number of input images. Affine determines the affine space that best models the input data. As can be seen in the figure, the individual basis vectors have no obvious interpretation. Although the space found by Affine is unique, its representation by basis vectors is degenerate. Any set of r linearly independent vectors drawn from the affine space can be used to represent it. This is due to the fact that the product WV is invariant under the transformation W ~ W 8 and V ~ 8-1 V. 1 Convex finds the r basis vectors whose convex hull best fits the input data. The optimization was performed by alternating between projected gradient steps of W and V. The constraint that the column sums of V equal unity was implemented 1 Affine is essentially equivalent to peA, except that they represent the affine space in different ways. Affine represents it with r points chosen from the space. peA represents the affine space with a single point from the space and r - 1 orthonormal directions. This is still a degenerate representation, but peA fixes it by taking the point to be the sample mean and the r - 1 directions to be the eigenvectors of the covariance matrix of X with the largest eigenvalues. Unsupervised Learning by Convex and Conic Coding Input Image Convex Coding [J ... . . ' . . . . . . . . ¥::' .... . ' . ..... . ·i~~.: ... ,', .. ,\",: ;,._ .. . . . . . . . . . . . .. ,' .. Convex Reconstructions Conic Coding [}§ :""''''''' . . .. .. . ••• • • : lo" ." • • . ~ .. .. : .. ,.'.: .. :, .. : "....: . .... : v ......... ........... . .. . .. . . . -. . Conic Reconstructions 519 Figure 3: Activities and reconstructions of a "2" using conic and convex coding. by adding a quadratic penalty term. In contrast to Affine, the basis vectors are interpretable as templates and are less blurred than those found by VQ. Many of the elements of Wand also of V are zero at the minimum. This eliminates many invariant transformations S, because they would violate the nonnegativity constraints on Wand V. From our simulations, it appears that most of the degeneracy seen in Affine is lifted by the nonnegativity constraints. Conic finds basis vectors whose conic hull best models the input images. The learning algorithm is similar to Convex except there is no penalty term on the sum of the activities. The Conic representation allows combinations of basis vectors, not just interpolations between them. As a result, the basis vectors found are features rather than templates, as seen in Figure 2. In contrast to Affine, the nonnegativity constraint leads to features that are interpretable as correlated strokes. As the number of basis vectors r increases, these features decrease in size until they become individual pixels. These models were used to classify novel test images. Recognition was accomplished by separately reconstructing the test images with the different digit models and associating the image with the model having the smallest reconstruction error. Figure 3 illustrates an example of classifying a "2" using the conic and convex encodings. The basis vectors are displayed weighted by their activites Va and the sparsity in the representations can be clearly seen. The bottom part of the figure shows the different reconstructions generated by the various digit models. With r = 25 patterns per digit class, Convex incorrectly classified 113 digits out of the 2007 test examples for an overall error rate of 5.6%. This is virtually identical to the performance of k = 1 nearest neighbor (112 errors) and linear r = 25 peA models (111 errors). However, scaling up the convex models to r = 100 patterns results in an error rate of 4.4% (89 errors). This improvement arises because the larger convex hulls can better represent the overall nonlinear nature of the input 520 D. D. Lee and H. S. Seung distributions. This is good performance relative to other methods that do not use prior knowledge of invariances, such as the support vector machine (4.0% [5]). However, it is not as good as methods that do use prior knowledge, such as nearest neighbor with tangent distance (2.6% [6]). On the other hand, Conic coding with r = 25 results in an error rate of 6.8% (138 errors). With larger basis sets r > 50, Conic shows worse performance as the features shrink to small spots. These results indicate that by itself, Conic does not yield good models; non-trivial correlations still remain in the Va and also need to be taken into account. For instance, while the conic basis for "9" can fit some "7" 's quite well with little reconstruction error, the codes Va are distinct from when it fits "9" 'so 5 Neural network implementation Conic and Convex were described above as matrix factorizations. Alternatively, the encoding can be performed by a neural network dynamics [7] and the learning by a synaptic update rule. We describe here the implementation for the Conic network; the Convex network is similar. The Conic network has a layer of N error neurons ei and a layer of r encoding neurons Va. The fixed point of the encoding dynamics dVa [i:e;Wi. +v.f (3) dt +Va t=l dei r -+e · Xi ~WiaVa, (4) dt t a=l optimizes Eq. (1), finding the best convex encoding of the input Xi. The rectification nonlinearity [x]+ = max(x,O) enforces the nonnegativity constraint. The error neurons subtract the reconstruction from the input Xi. The excitatory connection from ei to Va is equal and opposite to the inhibitory connection from Va back to ei. The Hebbian synaptic weight update (5) is made following convergence of the encoding dynamics for each input, while respecting the bound constraints on Wia . This performs stochastic gradient descent on the ensemble reconstruction error with learning rate "1. 6 Discussion Convex coding is similar to other locally linear models [8, 9, 10, 11]. Distance to a convex hull was previously used in nearest neighbor classification [12], though no learning algorithm was proposed. Conic coding is similar to the noisy OR [13, 14] and harmonium [15] models. The main difference is that these previous models contain discrete binary variables, whereas Conic uses continuous ones. The use of analog rather than binary variables makes the encoding computationally tractable and allows for interpolation between basis vectors. Unsupervised Learning by Convex and Conic Coding 521 Here we have emphasized the geometrical interpretation of Convex and Conic coding. They can also be viewed as probabilistic hidden variable models. The inputs Xi are visible while the Va are hidden variables, and the reconstruction error in Eq. (1) is related to the log likelihood, log P{xilva). No explicit model P(va) for the hidden variables was used, which limited the quality of the Conic models in particular. The feature discovery capabilities of Conic, however, make it a promising tool for building hierarchical representations. We are currently working on extending these new coding schemes and learning algorithms to multilayer networks. We acknowledge the support of Bell Laboratories. We thank C. Burges, C. Cortes, and Y. LeCun for providing us with the USPS database. We are also grateful to K. Clarkson, R. Freund, L. Kaufman, L. Saul, and M. Wright for helpful discussions. References [1] Hinton, GE & Zemel, RS (1994) . Autoencoders, minimum description length and Helmholtz free energy. Advances in Neural Information Processing Systems 6, 3- 10. [2] Ghahramani, Z (1995). Factorial learning and the EM algorithm. Advances in Neural Information Processing Systems 7, 617-624. [3] Olshausen, BA & Field, DJ (1996). Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature 381, 607-609. [4] Le Cun, Yet al. (1989). Backpropagation applied to handwritten zip code recognition. Neural Comput. 1, 541-55l. [5] Scholkopf, B, Burges, C, & Vapnik, V (1995). Extracting support data for a given task. KDD-95 Proceedings, 252-257. [6] Simard, P, Le Cun Y & Denker J (1993). Efficient pattern recognition using a new transformation distance. Advances in Neural Information Processing Systems 5, 5058. [7] Tank, DW & Hopfield, JJ (1986). Simple neural optimization networks; an AjD converter, signal decision circuit, and a linear programming circuit. IEEE 7rans. Circ. Syst. CAS-33, 533- 54l. [8] Bezdek, JC, Coray, C, Gunderson, R & Watson J (1981). Detection and characterization of cluster substructure. SIAM J. Appl. Math. 40, 339- 357; 358- 372. [9] Bregler, C & Omohundro, SM (1995). Nonlinear image interpolation using manifold learning. Advances in Neural Information Processing Systems 7, 973-980. [10] Hinton, GE, Dayan, P & Revow M (1996). Modeling the manifolds of images of handwritten digits. IEEE 7rans. Neural Networks, submitted. [11] Hastie, T, Simard, P & Sackinger E (1995). Learning prototype models for tangent distance. Advances in Neural Information Processing Systems 7, 999-1006. [12] Haas, HPA, Backer, E & Boxma, I (1980). Convex hull nearest neighbor rule. Fifth Intl. Conf. on Pattern Recognition Proceedings, 87-90. [13] Dayan, P & Zemel, RS (1995). Competition and multiple cause models. Neural Comput. 7, 565- 579. [14] Saund, E (1995). A multiple cause mixture model for unsupervised learning. Neural Comput. 7, 51-7l. [15] Freund, Y & Haussler, D (1992). Unsupervised learning of distributions on binary vectors using two layer networks. Advances in Neural Information Processing Systems 4,912-919.
1996
30
1,226
Reconstructing Stimulus Velocity from Neuronal Responses in Area MT Wyeth Bair, James R. Cavanaugh, J. Anthony Movshon Howard Hughes Medical Institute, and Center for Neural Science New York University 4 Washington Place, Room 809 New York, NY 10003 wyeth@cns.nyu.edu, jamesc@cns.nyu.edu, tony@cns.nyu.edu Abstract We employed a white-noise velocity signal to study the dynamics of the response of single neurons in the cortical area MT to visual motion. Responses were quantified using reverse correlation, optimal linear reconstruction filters, and reconstruction signal-to-noise ratio (SNR). The SNR and lower bound estimates of information rate were lower than we expected. Ninety percent of the information was transmitted below 18 Hz, and the highest lower bound on bit rate was 12 bits/so A simulated opponent motion energy subunit with Poisson spike statistics was able to out-perform the MT neurons. The temporal integration window, measured from the reverse correlation half-width, ranged from 30-90 ms. The window was narrower when a stimulus moved faster, but did not change when temporal frequency was held constant. 1 INTRODUCTION Area MT neurons can show precise and rapid modulation in response to dynamic noise stimuli (Bair and Koch, 1996); however, computational models of these neurons and their inputs (Adelson and Bergen, 1985; Heeger, 1987; Grzywacz and Yuille, 1990; Emerson et al., 1992; Qian et al., 1994; Nowlan and Sejnowski, 1995) have primarily been compared to electrophysiological results based on time and ensemble averaged responses to deterministic stimuli, e.g., drifting sinusoidal gratings. Reconstructing Stimulus Velocity from Neuronal Responses in Area MT 35 Using methods introduced by Bialek et al. (1991) and further analyzed by Gabbiani and Koch (1996) for the estimation of information transmission by a neuron about a white-noise stimulus, we set out to compare the responses of MT neurons for white-noise velocity signals to those of a model based on opponent motion energy sub-units. The results of two analyses are summarized here. In the first, we compute a lower bound on information transmission using optimal linear reconstruction filters and examine the SNR as a function of temporal frequency. The second analysis examines changes in the reverse correlation (the cross-correlation between the stimulus and the resulting spike trains) as a function of spatial frequency and temporal frequency of the moving stimulus pattern. 2 EXPERIMENTAL METHODS Spike trains were recorded extracellularly from 26 well-isolated single neurons in area MT of four anesthetized, paralyzed macaque monkeys using methods described in detail elsewhere (Levitt et al., 1994). The size of the receptive fields and the spatial and temporal frequency preferences of the neurons were assessed quantitatively using drifting sinusoidal gratings, after which a white-noise velocity signal, s(t), was used to modulate the position (within a fixed square aperture) of a low-pass filtered 1D Gaussian white-noise (GWN) pattern. The frame rate of the display was 54 Hz or 81 Hz. The spatial noise pattern consisted of 256 discrete intensity values, one per spatial unit. Every 19 ms (or 12 ms at 81 Hz), the pattern shifted, or jumped, 6. spatial units along the axis of the neuron's preferred direction, where 6., the jump size, was chosen according to a Gaussian, binary, or uniform probability distribution. The maximum spatial frequency in the pattern was limited to prevent aliasing. In the first type of experiment, 10 trials of a 30 s noise sequence, s(t), and 10 trials of its reverse, -s(t), were interleaved. A standard GWN spatial pattern and velocity modulation pattern were used for all cells, but for each cell, the stimulus was scaled for the receptive field size and aligned to the axis of preferred motion. Nine cells were tested with Gaussian noise at 81 Hz, 15 cells with binary noise at 81 Hz and 54 Hz, and 10 cells with uniform noise at 54 Hz. In another experiment, a sinusoidal spatial pattern (rather than GWN) moved according to a binary white-noise velocity signal. Thials were interleaved with all combinations of four spatial frequencies at octave intervals and four relative jump sizes: 1/4, 1/8, 1/16, and 1/32 of each spatial period. Typically 10 trials of length 3 s were run. Four cells were tested at 54 Hz and seven at 81 Hz. 3 ANALYSIS AND MODELING METHODS We used the linear reconstruction methods introduced by Bialek et al. (1991) and further analyzed by Gabbiani and Koch (1996) to compute an optimal linear estimate of the stimulus, s(t), described above, based on the neuronal response, x(t). A single neuronal response was defined as the spike train produced by s (t) minus the spike train produced by -s(t). This overcomes the neuron's limited dynamic range in response to anti-preferred direction motion (Bialek et al., 1991). 36 W. Bair; J. R. Cavanaugh and J. A. Movshon The linear filter, h(t), which when convolved with the response yields the minimum mean square error estimate, Sest, of the stimulus can be described in terms of its Fourier transform, H(w) _ Rsz( -w) Rzz(w) , (1) where Rsz(w) is the Fourier transform of the cross-correlation rsz(r) of the stimulus and the resulting spike train and Rzz(w) is the power spectrum, i.e., the Fourier transform of the auto-correlation, of the spike train (for details and references, see Bialek et aI., 1991; Gabbiani and Koch, 1996). The noise, n(t), is defined as the difference between the stimulus and the reconstruction, and the SNR is defined as n(t) = Sest(t) - s(t), SNR(w) = Rss(w) , Rnn(w) (2) (3) where Rss(w) is the Fourier power spectrum of the stimulus and Rnn(w) is the power spectrum of the noise. If the stimulus amplitude distribution is Gaussian, then SNR(w) can be integrated to give a lower bound on the rate of information transmission in bits/s (Gabbiani and Koch, 1996). The motion energy model consisted of opponent energy sub-units (Adelson and Bergen, 1985) implemented with Gabor functions (Heeger, 1987; Grzywacz and Yuille, 1990) in two spatial dimensions and time. The spatial frequency of the Gabor function was set to match the spatial frequency of the stimulus, and the temporal frequency was set to match that induced by a sequence of jumps equal to the standard deviation (SD) of the amplitude distribution (which is the jump size in the case of a binary distribution). We approximated causality by shifting the output forward in time before computing the optimal linear filter. The model operated on the same stimulus patterns and noise sequences that were used to generate stimuli for the neurons. The time-varying response of the model was broken into two halfwave rectified signals which were interpreted as the firing probabilities of two units, a neuron and an anti-neuron that preferred the opposite direction of motion. From each unit, ten 30 s long spike trains were generated with inhomogeneous Poisson statistics. These 20 model spike trains were used to reconstruct the velocity signal in the same manner as the MT neuron output. 4 RESULTS Stimulus reconstruction. Optimal linear reconstruction filters, h(t), were computed for 26 MT neurons from responses to 30 s sequences of white-noise motion. A typical h(t), shown in Fig. lA (large dots), was dominated by a single positive lobe, often preceded by a smaller negative lobe. It was thinner than the reverse correlation rsz(r) (Fig. lA, small dots) from which it was derived due to the division by the low-pass power spectrum of the spikes (see Eqn. 1). Also, rsz(r) occasionally had a slower, trailing negative lobe but did not have the preceding negative lobe of h(t). On average, h(t) peaked at -69 ms (SD 17) and was 33 ms (SD 12) wide at half-height. The peak for rsz(r) occurred at the same time, but the width was 41 ms (SD 15), ranging from 30-90 ms. The point of half-rise on the right side of the peak was -53 ms (SD 9) for h(t) and -51 ms (SD 9) for rsz(r). For all plots, Reconstructing Stimulus Velocity from Neuronal Responses in Area MT 37 vertical axes for velocity show normalized stimulus velocity, i.e., stimulus velocity was scaled to have unity SD before all computations. Fig lC (dots) shows the SNR for the reconstruction using the h(t) in panel A. For 8 cells tested with Gaussian velocity noise, the integral of the log of the SNR gives a lower bound for information transmission, which was 6.7 bits/s (SD 2.8), with a high value of 12.3 bits/so Most of the information was carried below 10 Hz, and 90% of the information was carried below 18.4 Hz (SD 2.1). In Fig. ID, the failure of the reconstruction (dots) to capture higher frequencies in the stimulus (thick line) is directly visible. Both h(t) and SNR(w) were similar but on average slightly greater in amplitude for tests using binary and uniform distributed noise. Gaussian noise has many jumps at or near zero which may induce little or no response. ~ 0.3 .... ..... u ..9 0.2 o > "'2 0.1 N ~ ~ S 0.6 A B o Z -0.1 +--..--N..,...e..,...ur..,...o..,...ll..........,..............,.........-.-.....-.-....--.--I Model -160 3 ~ .... ..... U 0 ....-4 0 > 'i:j 0 0 N ..... ....-4 ~ S 0 Z -3 100 -80 0 -80 o Time relative to spike (ms) 150 D 200 Time (ms) 3 ~2 Z CJ:) o 20 40 Frequency (Hz) 250 300 Figure 1: (A) Optimal linear filter h(t) (big dots) from Eqn. 1 and cross-correlation Tsz(r) ,(small dots) for one MT neuron. (B) h(t) (thick line) and Tsz(r) (thin line) for an opponent motion energy model. (C) SNR(w) for the neuron (dots) and the model (line). (D) Reconstruction for the neuron (dots) and model (thin line) of the stimulus velocity (thick line). Velocity was normalized to unity SD. Curves for Tsz(r) were scaled by 0.5. ·Note the different vertical scale in B. 38 W. Bair, J. R. Cavanaugh and J. A. Movshon An opponent motion energy model using Gabor functions was simulated with spatial SD 0.50 , spatial frequency 0.625 cycr, temporal SD 12.5 ms, and temporal frequency 20 Hz. The model was tested with a Gaussian velocity stimulus with SD 320 / s. Because an arbitrary scaling of the spatial parameters in the model does not affect the temporal properties of the information transmission, this was effectively the same stimulus that yielded the neuronal data shown in Fig. 1A. Spike trains were generated from the model at 20 Hz (matched to the neuron) and used to compute h(t) (Fig. 1B, thick line). The model h(t) was narrower than that for the MT neuron, but was similar to h(t) for Vi neurons that have been tested (unpublished analysis). This simple model of a putative input sub-unit to MT transmitted 29 bits/s- more than the best MT neurons studied here. The SNR ratio and the reconstruction for the model are shown in Fig. 1C,D (thin lines). The filter h(t) for the model (Fig. 1B thick line) was more symmetric than that for the neuron due to the symmetry of the Gabor function used in the model. 1 Neuron 1 Neuron 2 0.5 -120 -80 -40 0 -160 -120 -80 -40 0 1/32 1/16 -0. 5 +-----.-....:,....;.....,-.-~"T-T---r--.--.___,_~~__i -160 -120 -80 -40 0 -160 -120 -80 -40 o Time relative to spike (ms) Figure 2: The width of r 8Z (r) changes with temporal frequency, but not spatial frequency. Data from two neurons are shown, one on the left, one on the right. Top: rsx(r) is shown for binary velocity stimuli with jump sizes 1/4, 1/8, 1/16, and 1/32 (thick to thin lines) of the spatial period (10 trials, 3 s/trial). The left side of the peak shifts leftward as jump size decreases. See text for statistics. Bottom: The relative jump size, thus temporal frequency, was constant for the four cases in each panel (1/32 on the left, 1/16 on the right). The peaks do not shift left or right as spatial frequency and jump size change inversely. Thicker lines represent larger jumps and lower spatial frequencies. Reconstructing Stimulus Velocity from Neuronal Responses in Area MT 39 Changes in rsx(r). We tested 11 neurons with a set of binary white-noise motion stimuli that varied in spatial frequency and jump size. The spatial patterns were sinusoidal gratings. The peaks in rsx(r) and h(t) were wider when smaller jumps (slower velocities) were used to move the same spatial pattern. Fig. 2 shows data for two neurons plotted for constant spatial frequency (top) and constant effective temporal frequency, or contrast frequency (bottom). Jump sizes were 1/4, 1/8, 1/16, and 1/32 (thick to thin lines, top panels) ofthe period of the spatial pattern. (Note, a 1/2 period jump would cause an ambiguous motion.) Relative jump size was constant in the bottom panels, but both the spatial period and the velocity increased in octaves from thin to thick lines. One of the plots in the upper panel also appears in the lower panel for each neuron. For 26 conditions in 11 MT neurons (up to 4 spatial frequencies per neuron) the left and right half-rise points of the peak of rsx(r) shifted leftward by 19 ms (SD 12) and 4.5 ms (SD 4.0), respectively, as jump size decreased. The width, therefore, increased by 14 ms (SD 12). These changes were statistically significant (p < 0.001, t-test). In fact, the left half-rise point moved leftward in all 26 cases, and in no case did the width at half-height decrease. On the other hand, there was no significant change in the peak width or half-rise times when temporal frequency was constant, as demonstrated in the lower panels of Fig. 2. 5 DISCUSSION From other experiments using frame-based displays to present moving stimuli to MT cells, we know that roughly half of the cells can modulate to a 60 Hz signal in the preferred direction and that some provide reliable bursts of spikes on each frame but do not respond to null direction motion. Therefore, one might expect that these cells could easily be made to transmit nearly 60 bits/ s by moving the stimulus randomly in either the preferred or null direction on each video frame. However, the stimuli that we employed here did not result in such high frequency modulation, nor did our best lower bound estimate of information transmission for an MT cell, 12 bits/s, approach the 64 bits/s capacity of the motion sensitive HI neuron in the fly (Bialek et al., 1991). In recordings from seven VI neurons (not shown here), two directional complex cells responded to the velocity noise with high temporal precision and fired a burst of spikes on almost every preferred motion frame and no spikes on null motion frames. At 53 frames/s, these cells transmitted over 40 bits/so We hope that further investigation will reveal whether the lack of high frequency modulation in our MT experiments was due to statistical variation between animals, the structure of the stimulus, or possibly to anesthesia. In spite of finding less high frequency bandwidth than expected, we were able to document consistent changes, namely narrowing, of the temporal integration window of MT neurons as temporal frequency increased. Similar changes in the time constant of motion processing have been reported in the fly visual system, where it appears that neither velocity nor temporal frequency alone can account for all changes (de Ruyter et al., 1986; Borst & Egelhaaf, 1987). The narrowing of rsx(r) with higher temporal frequency does not occur in our simple motion energy model, which lacks adaptive mechanisms, but it could occur in a model which integrated signals from many motion energy units having distributed temporal frequency tuning, even without other sources of adaptation. 40 W. Bair, J. R. CavanaughandJ. A. Movshon We were not able to assess whether changes in the integration window developed quickly at the beginning of individual trials, but an analysis not described here at least indicates that there was very little change in the position and width of r 83: (T) and h(t) after the first few seconds during the 30 s trials. Acknowledgements This work was funded by the Howard Hughes Medical Institute. We thank Fabrizio Gabbiani and Christof Koch for helpful discussion, Lawrence P. O'Keefe for assistance with electrophysiology, and David Tanzer for assistance with software. References Adelson EH, Bergen JR (1985) Spatiotemporal energy models for the perception of motion. J Opt Soc Am A 2:284- 299. Bair W, Koch C (1996) Temporal precision of spike trains in extrastriate cortex of the behaving macaque monkey. Neural Comp 8:1185-1202. Bialek W, Rieke F, de Ruyter van Steveninck RR, Warland D (1991) Reading a neural code. Science 252:1854- 1857. Borst A, Egelhaaf M (1987) Temporal modulation of luminance adapts time constant of fly movement detectors. BioI Cybern 56:209-215. Emerson RC, Bergen JR, Adelson EH (1992) Directionally selective complex cells and the computation of motion energy in cat visual cortex. Vision Res 32:203218. Gabbiani F, Koch C (1996) Coding of time-varying signals in spike trains of integrate-and-fire neurons with random threshold. Neural Comp 8:44-66. Grzywacz NM, Yuille AL (1990) A model for the estimate of local image velocity by cells in the visual cortex. Proc R Soc Lond B 239:129-161. Heeger DJ (1987) Model for the extraction of image flow. J Opt Soc Am A 4:14551471. Levitt JB, Kiper DC, Movshon JA (1994) Receptive fields and functional architecture of macaque V2. J .Neurophys. 71:2517-2542. Nowlan SJ, Sejnowski TJ (1994) Filter selection model for motion segmentation and velocity integration. J Opt Soc Am A 11:3177-3200. Qian N, Andersen RA, Adelson EH (1994) Transparent motion perception as detection of unbalanced motion signals .3. Modeling. J Neurosc 14:7381-7392. de Ruyter van Steveninck R, Zaagman WH, Mastebroek HAK (1986) Adaptation of transient responses of a movement-sensitive neuron in the visual-system of the blowfly calliphora-erythrocephala. BioI Cybern 54:223-236.
1996
31
1,227
Multidimensional Triangulation and Interpolation for Reinforcement Learning Scott Davies scottd@cs.cmu.edu Department of Computer Science, Carnegie Mellon University 5000 Forbes Ave, Pittsburgh, PA 15213 Abstract Dynamic Programming, Q-Iearning and other discrete Markov Decision Process solvers can be -applied to continuous d-dimensional state-spaces by quantizing the state space into an array of boxes. This is often problematic above two dimensions: a coarse quantization can lead to poor policies, and fine quantization is too expensive. Possible solutions are variable-resolution discretization, or function approximation by neural nets. A third option, which has been little studied in the reinforcement learning literature, is interpolation on a coarse grid. In this paper we study interpolation techniques that can result in vast improvements in the online behavior of the resulting control systems: multilinear interpolation, and an interpolation algorithm based on an interesting regular triangulation of d-dimensional space. We adapt these interpolators under three reinforcement learning paradigms: (i) offline value iteration with a known model, (ii) Q-Iearning, and (iii) online value iteration with a previously unknown model learned from data. We describe empirical results, and the resulting implications for practical learning of continuous non-linear dynamic control. 1 GRID-BASED INTERPOLATION TECHNIQUES Reinforcement learning algorithms generate functions that map states to "cost-t<r go" values. When dealing with continuous state spaces these functions must be approximated. The following approximators are frequently used: • Fine grids may be used in one or two dimensions. Above two dimensions, fine grids are too expensive. Value functions can be discontinuous, which (as we will see) can lead to su boptimalities even with very fine discretization in two dimensions . • Neural nets have been used in conjunction with TD [Sutton, 1988] and Q-Iearning [Watkins, 1989] in very high dimensional spaces [Tesauro, 1991, Crites and Barto, 1996]. While promising, it is not always clear that they produce the accurate value functions that might be needed for fine nearoptimal control of dynamic systems, and the most commonly used methods of applying value iteration or policy iteration with a neural-net value function are often unstable. [Boyan and Moore, 1995]. 1006 S. Davies Interpolation over points on a coarse grid is another potentially useful approximator for value functions that has been little studied for reinforcement learning. This paper attempts to rectify this omission. Interpolation schemes may be particularly attractive because they are local averagers, and convergence has been proven in such cases for offline value iteration [Gordon, 1995]. All of the interpolation methods discussed here split the state space into a regular grid of d-dimensional boxes; data points are associated with the centers or the corners of the resulting boxes. The value at a given point in the continuous state space is computed as a weighted average of neighboring data points. 1.1 MULTILINEAR INTERPOLATION When using multilinear interpolation, data points are situated at the corners of the grid's boxes. The interpolated value within a box is an appropriately weighted average of the 2d datapoints on that box's corners. The weighting scheme assures global continuity of the interpolated surface, and also guarantees that the interpolated value at any grid corner matches the given value of that corner. In one-dimensional space, multilinear interpolation simply involves piecewise linear interpolations between the data points. In a higher-dimensional space, a recursive (though not terribly efficient) implementation can be described as follows: • Pick an arbitrary axis. Project the query point along this axis to each of the two opposite faces of the box containing the query point. • Use two (d - 1)-dimensional multilinear interpolations over the 2d - 1 datapoints on each of these two faces to calculate the values at both of these projected points. • Linearly interpolate between the two values generated in the previous step. Multilinear interpolation processes 2d data points for every query, which becomes prohibitively expensive as d increases. 1.2 SIMPLEX-BASED INTERPOLATION It is possible to interpolate over d + 1 of the data points for any given query in only O( d log d) time and still achieve a continuous surface that fits the datapoints exactly. Each box is broken into d! hyperdimensional triangles, or simplexes, according to the Coxeter-Freudenthal-Kuhn triangulation [Moore, 1992]. Assume that the box is the unit hypercube, with one corner at (Xl, X2, . .. , Xd) = (0,0, ... ,0), and the diagonally opposite corner at (1,1, ... ,1). Then, each simplex in the Kuhn triangulation corresponds to one possible permutation p of (1,2, ... , d), and occupies the set of points satisfying the equation o ~ Xp(l) ~ Xp(2) ~ ... ~ Xp(d) ~ 1. Triangulating each box into d! simplexes in this manner generates a conformal mesh: any two elements with a (d - 1 )-dimensional surface in common have entire faces in common, which ensures continuity across element boundaries when interpolating. We use the Kuhn triangulation for interpolation as follows: • Translate and scale to a coordinate system in which the box containing the query point is the unit hypercube. Let the new coordinate of the query point be (x~, ... ,x~). • Use a sorting algorithm to rank x~ through x~. This tells us the simplex of the Kuhn triangulation in which the query point lies. Triangulation and Interpolation for Reinforcement Learning ]007 • Express (x~, . .. , x~) as a convex combination of the coordinates of the relevant simplex's (d + 1) corners . • Use the coefficients determined in the previous step as the weights for a weighted sum of the data values stored at the corresponding corners. At no point do we explicitly represent the d! different simplexes. All of the above steps can be performed in Oed) time except the second, which can be done in O( d log d) time using conventional sorting routines. 2 PROBLEM DOMAINS CAR ON HILL: In the Hillcar domain, the goal is to park a car near the top of a one-dimensional hill. The hill is steep enough that the driver needs to back up in order to gather enough speed to get to the goal. The state space is two-dimensional (position, velocity). See [Moore and Atkeson, 1995] for further details, but note that our formulation is harder than the usual formulation in that the goal region is restricted to a narrow range of velocities around 0, and trials start at random states. The task is specified by a reward of -1 for any action taken outside the goal region, and 0 inside the goal. No discounting is used, and two actions are available: maximum thrust backwards, and maximum thrust forwards. ACROBOT: The Acrobot is a two-link planar robot acting in the vertical plane under gravity with a weak actuator at its elbow joint joint. The shoulder is unactuated. The goal is to raise the hand to at least one link's height above the unactuated pivot [Sutton, 1996]. The state space is four-dimensional: two angular positions and two angular velocities. Trials always start from a stationary position hanging straight down. This task is formulated in the same way as the car-on-thehill. The only actions allowed are the two extreme elbow torques. 3 APPLYING INTERPOLATION: THREE CASES 3.1 CASE I: OFFLINE VALUE ITERATION WITH A KNOWN MODEL First, we precalculate the effect of taking each possible action from each state corresponding to a datapoint in the grid. Then, as suggested in [Gordon, 1995], we use these calculations to derive a completely discrete MDP. Taking any action from any state in this MDP results in c possible successor states, where c is the number of datapoints used per interpolation. Without interpolation, c is 1; with multilinear interpolation, 2d; with simplex-based interpolation, d + 1. We calculate the optimal policy for this derived MDP offline using value iteration [Ross, 1983]; because the value iteration can be performed on a completely discrete MDP, the calculations are much less computationally expensive than they would have been with many other kinds of function approximators. The value iteration gives us values for the datapoints of our grid, which we may then use to interpolate the values at other states during online control. 3.1.1 Hillcar Results: value iteration with known model We tested the two interpolation methods on a variety of quantization levels by first performing value iteration offline, and then starting the car from 1000 random states and averaging the number of steps taken to the goal from those states. We also recorded the number of backups required before convergence, as well as the execution time required for the entire value iteration on a 85 MHz Sparc 5. See Figure 1 for the results. All steps-to-goal values are means with an expected error of 2 steps. 1008 S. Davies Grid size Interpolation Method l1:l 21:l 51:l 301:l None Steps to Goal : 237 131 133 120 Backups: 2.42K 15.4K 156K 14.3M Time (sec) : 0.4 1.0 4.1 192 MultlLm Steps to Goal : 134 116 lOS 107 Backups: 4.S4K 18.1K 205K 17.8M Time _Lsec): 0.6 1.3 7.1 405 Simplex Steps to Goal: 134 118 109 107 Backups: 6.17K 18.1K 195K 17.9M Time (sec) : 0.5 1.2 5.7 328 Figure 1: Hillcar: value iteration with known model Grid size Interpolation Method 84 94 104 11" 124 13" 14'1 15" None Steps to Goal: 44089 26952 > 100000 Backups: 280K 622K 1.42M Time (sec) : 15 30 53 MultlLm Steps to Goal : 3340 2006 1136 3209 1300 1820 1518 1802 Backups: 233K 1.01M 730K 2.01M 2.03M 3.74M 4.45M 6.78M Time (sec): 17 43 42 83 99 164 197 284 Simplex Steps to Goal: 4700 8007 2953 3209 4663 2733 1742 9613 Backups: 196K 1.16M 590K 2.28M 1.62M 4.03M 3.65M 6.73M Time(sec): 9 24 22 47 47 86 93 142 Figure 2: Acrobot: value iteration with known model The interpolated functions require more backUps for convergence, but this is amply compensated by dramatic improvement in the policy. Surprisingly, both interpolation methods provide improvements even at extremely high grid resolutions - the noninterpolated grid with 301 datapoints along each axis fared no better than the interpolated grids with only 21 datapoints along each axis(!). 3.1.2 Acrobot Results: value iteration with known model We used the same value iteration algorithm in the acrobot domain. In this case our test trials always began from the same start state, but we ran tests for a larger set of grid sizes (Figure 2). Grids with different resolutions place grid cell boundaries at different locations, and these boundary locations appear to be important in this problem the performance varies unpredictably as the grid resolution changes. However, in all cases, interpolation was necessary to arrive at a satisfactory solution; without interpolation, the value iteration often failed to converge at all. With relatively coarse grids it may be that any trajectory to the goal passes through some grid box more than once, which would immediately spell disaster for any algorithm associating a constant value over that entire grid box. Controllers using multilinear interpolation consistently fared better than those employing the simplex-based interpolation; the smoother value function provided by multilinear interpolation seems to help. However, value iteration with the simplexbased interpolation was about twice as fast as that with multilinear interpolation. In higher dimensions this speed ratio will increase. Triangulation and Interpolation for Reinforcement Learning 1009 3.2 CASE II: Q-LEARNING Under a second reinforcement learning paradigm, we do not use any model. Rather, we learn a Q-function that directly maps state-action pairs to long-term rewards [Watkins, 1989]. Does interpolation help here too? In this implementation we encourage exploration by optimistically initializing the Q-function to zero everywhere. After travelling a sufficient distance from our last decision point, we perform a single backup by changing the grid point values according to a perceptron-like update rule, and then we greedily select the action for which the interpolated Q-function is highest at the current state. 3.2.1 Hillcar Results: Q-Learning We used Q-Learning with a grid size of 112. Figure 3 shows learning curves for three learners using the three different interpolation techniques. Both interpolation methods provided a significant improvement in both initial and final online performance. The learner without interpolation achieved a final average performance of about 175 steps to the goal; with multilinear interpolation, 119; with simplex-based interpolation, 122. Note that these are all significant improvements over the corresponding results for offline value iteration with a known model. Inaccuracies in the interpolated functions often cause controllers to enter cycles; because the Q-Iearning backups are being performed online, however, the Q-Iearning controller can escape from these control cycles by depressing the Q-values in the vicinities of such cycles. 3.2.2 Acrobot Results: Q-Learning We used the same algorithms on the acrobot domain with a grid size of 154 ; results are shown in Figure 3. ...,.,.,. . -,eoooo -,eoooo ~ J -1.,..07 15 1 ~ - -'5&+07 l -_._-·200000 -2 ... 01 L---_'-----'_---'_--'_---'-_~_~ _ __' o !iO 100 150 200 2iO 3DO 360 400 450 500 0 !iO 100 150 200 250 300 !fiO 400 ~ofl,.. ~otT!U Figure 3: Left: Cumulative performance of Q-Iearning hillcar on an 112 grid. (Multilinear interpolation comes out on top; no interpolation on the bottom.) Right: Q-Iearning acrobot on a. 154 grid. (The two interpolations come out on top with nearly identical performance.) For each learner, the y-axis shows the sum of rewards for all trials to date. The better the average performance, the shallower the gradient. Gradients are always negative because each state transition before reaching the goal results in a reward of -1. Both Q-Iearners using interpolation improved rapidly, and eventually reached the goal in a relatively small number of steps per trial. The learner using multilinear interpolation eventually achieved an average of 1,529 steps to the goal per trial; the learner using simplex-based interpolation achieved 1,727 steps per trial. On the other hand, the learner not using any interpolation fared much worse, taking 1010 s. Davies an average of more than 27,000 steps per trial. (A controller that chooses actions randomly typically takes about the same number of steps to reach the goal.) Simplex-based interpolation provided on-line performance very close to that provided by multilinear interpolation, but at roughly half the computational cost. 3.3 CASE III: VALUE ITERATION WITH MODEL LEARNING Here, we use a model of the system, but we do not assume that we have one to start with. Instead, we learn a model of the system as we interact with it; we assume this model is adequate and calculate a value function via the same algorithms we would use if we knew the true model. This approach may be particularly beneficial for tasks in which data is expensive and computation is cheap. Here, models are learned using very simple grid-based function approximators without interpolation for both the reward and transition functions of the model. The same grid resolution is used for the value function grid and the model approximator. We strongly encourage exploration by initializing the model so that every state is initially assumed to be an absorbing state with zero reward. While making transitions through the state space, we update the model and use prioritized sweeping [Moore and Atkeson, 1993] to concentrate backups on relevant parts of the state space. We also occasionally stop to recalculate the effects of all actions under the updated model and then run value iteration to convergence. As this is fairly time-consuming, it is done rather rarely; we rely on the updates performed by prioritized sweeping to guide the system in the meantime . . ,00000 .",.". ~_'------'_~_-.L_--'-_--'-_--'-_~ o 50 100 150 200 2!10 !OO 350 -400 4SO 500 0 50 100 150 200 250 ~oo .:ISO 400 ~o1 Tn'" ~o1 T "'1I Figure 4: Left: Cumulative performance, model-learning on hillcar with a 112 grid. Right: Acrobot with a 154 grid. In both cases, multilinear interpolation comes out on top, while no interpolation winds up on the bottom. 3.3.1 Hillcar Results: value iteration with learned model We used the algorithm described above with an ll-by-ll grid. An average of about two prioritized sweeping backups were performed per transition; the complete recalculations were performed every 1000 steps throughout the first two trials and every 5000 steps thereafter. Figure 4 shows the results for the first 500 trials. Over the first 500 trials, the learner using simplex-based interpolation didn't fare much better than the learner using no interpolation. However, its performance on trials 1500-2500 (not shown) was close to that of the learner using multilinear interpolation, taking an average of 151 steps to the goal per trial while the learner using multilinear interpolation took 147. The learner using no interpolation did significantly worse than the others in these later trials, taking 175 steps per trial. Triangulation and Interpolation for Reinforcement Learning 1011 The model-learners' performance improved more quickly than the Q-Iearners' over the first few trials; on the other hand, their final performance was significantly worse that the Q-Iearners'. 3.3.2 Acrobot Results: value iteration with learned model We used the same algorithm with a 154 grid on the acrobot domain, this time performing the complete recalculations every 10000 steps through the first two trials and every 50000 thereafter. Figure 4 shows the results. In this case, the learner using no interpolation took so much time per trial that the experiment was aborted early; after 100 trials, it was still taking an average of more than 45,000 steps to reach the goal. The learners using interpolation, however, fared much better. The learner using multilinear interpolation converged to a solution taking 938 steps per trial; the learner using simplex-based interpolation averaged about 2450 steps. Again, as the graphs show, these three learners initially improve significantly faster than did the Q-Learners using similar grid sizes. 4 CONCLUSIONS We have shown how two interpolation schemes- one based on a weighted average of the 2d points in a square cell, the other on a d- dimensional triangulation-may be used in three reinforcement learning paradigms: Optimal policy computation with a known model, Q-Iearning, and online value iteration while learning a model. In each case our empirical studies demonstrate interpolation resoundingly decreasing the quantization level necessary for a satisfactory solution. Future extensions of this research will explore the use of variable resolution grids and triangulations, multiple low-dimensional interpolations in place of one high-dimension interpolation in a manner reminiscent ofCMAC [Albus, 1981], memory-based approximators, and more intelligent exploration. This research was funded in part by a National Science Foundation Graduate Fellowship to Scott Davies, and a Research Initiation Award to Andrew Moore. References [Albus, 1981] J. S. Albus. Brains, BehaVIour and Robottcs. BYTE Books, McGraw-Hili , 1981. [Boyan and Moore, 1995] J. A. Boyan and A. W. Moore. Generalization in Reinforcement Learning: Safely Approximating the Value Function. In Neural Information Processing Systems 7, 1995. [Crites and Barto, 1996] R. H. Crites and A. G. Barto. Improving Elevator Performance using Reinforcement Learning. In D. Touretzky, M. Mozer, and M. Hasselmo, editors, Neural Information Processing Systems 8, 1996. [Gordon, 1995] G. Gordon. Stable Function Approximation in Dynamic Programming. In Proceedmgs of the 12th International Conference on Machme Learning. Morgan Kaufmann, June 1995. [Moore and Atkeson, 1993] A. W. Moore and C . G. Atkeson . Prioritized Sweeping: Reinforcement Learning with Less Data and Less Real Time. Machme Learning, 13, 1993. [Moore and Atkeson, 1995] A. W. Moore and C. G. Atkeson. The Parti-game Algorithm for Variable Resolution Reinforcement Learning in Multidimensional State-spaces. Machine Learning, 21, 1995. [Moore, 1992] D. W . Moore. Simplical Mesh Generation with Applications. PhD. Thesis. Report no. 92-1322, Cornell University, 1992. [Ross, 1983] S. Ross. Introduction to Stochastic Dynamic Programming. Academic Press, New York, 1983. [Sutton, 1988] R. S. Sutton. Learning to Predict by the Methods of Temporal Differences. Machine Learning, 3:9-44, 1988. [Sutton, 1996] R. S. Sutton. Generalization in Reinforcement Learning: Successful Examples Using Sparse Coarse Coding. In D. Touretzky, M. Mozer, and M. Hasselmo, editors, Neural Information Processing Systems 8, 1996. [Tesauro, 1991] G. J. Tesauro. Practical Issues in Temporal Difference Learning. RC 17223 (76307), IBM T. J . Watson Research Center, NY, 1991. [Watkins, 1989] C. J . C . H . Watkins. Learning from Delayed Rewards. PhD. Thesis, King's College, University of Cambridge, May 1989.
1996
32
1,228
Viewpoint invariant face recognition using independent component analysis and attractor networks Marian Stewart Bartlett University of California San Diego The Salk Institute La Jolla, CA 92037 marni@salk.edu Terrence J. Sejnowski University of California San Diego Howard Hughes Medical Institute The Salk Institute, La Jolla, CA 92037 terry@salk.edu Abstract We have explored two approaches to recogmzmg faces across changes in pose. First, we developed a representation of face images based on independent component analysis (ICA) and compared it to a principal component analysis (PCA) representation for face recognition. The ICA basis vectors for this data set were more spatially local than the PCA basis vectors and the ICA representation had greater invariance to changes in pose. Second, we present a model for the development of viewpoint invariant responses to faces from visual experience in a biological system. The temporal continuity of natural visual experience was incorporated into an attractor network model by Hebbian learning following a lowpass temporal filter on unit activities. When combined with the temporal filter, a basic Hebbian update rule became a generalization of Griniasty et al. (1993), which associates temporally proximal input patterns into basins of attraction. The system acquired representations of faces that were largely independent of pose. 1 Independent component representations of faces Important advances in face recognition have employed forms of principal component analysis, which considers only second-order moments of the input (Cottrell & Metcalfe, 1991; Turk & Pentland 1991). Independent component analysis (ICA) is a generalization of principal component analysis (PCA), which decorrelates the higher-order moments of the input (Comon, 1994). In a task such as face recognition, much of the important information is contained in the high-order statistics of the images. A representational basis in which the high-order statistics are decorrelated may be more powerful for face recognition than one in which only the second order statistics are decorrelated, as in PCA representations. We compared an ICAbased representation to a PCA-based representation for recognizing faces across changes in pose. 818 M. S. Bartlett and T. J. Sejnowski -30" -IS" 0" IS" 30" Figure 1: Examples from image set (Beymer, 1994). The image set contained 200 images of faces, consisting of 40 subjects at each of five poses (Figure 1). The images were converted to vectors and comprised the rows of a 200 x 3600 data matrix, X. We consider the face images in X to be a linear mixture of an unknown set of statistically independent source images S, where A is an unknown mixing matrix (Figure 2). The sources are recovered by a matrix of learned filters, W, which produce statistically independent outputs, U. s X U ~ X ~ X ~ ~ III Sources Unknown Face Learned Separated Mixing Images Weights Outputs Process Figure 2: Image synthesis model. The weight matrix, W, was found through an unsupervised learning algorithm that maximizes the mutual information between the input and the output of a nonlinear transformation (Bell & Sejnowski, 1995). This algorithm has proven successful for separating randomly mixed auditory signals (the cocktail party problem), and has recently been applied to EEG signals (Makeig et al., 1996) and natural scenes (see Bell & Sejnowski, this volume). The independent component images contained in the rows of U are shown in Figure 3. In contrast to the principal components, all 200 independent components were spatially local. We took as our face representation the rows of the matrix A = W- 1 which provide the linear combination of source images in U that comprise each face image in X. 1.1 Face Recognition Performance: leA vs. Eigenfaces We compared the performance of the leA representation to that of the peA representation for recognizing faces across changes in pose. The peA representation of a face consisted of its component coefficients, which was equivalent to the "Eigenface" Viewpoint Invariant Face Recognition 819 Figure 3: Top: Four independent components of the image set. Bottom: First four principal components. representation (Turk & Pentland, 1991). A test image was recognized by assigning it the label of the nearest of the other 199 images in Euclidean distance. Classification error rates for the ICA and PCA representations and for the original graylevel images are presented in Table 1. For the PCA representation, the best performance was obtained with the 120 principal components corresponding to the highest eigenvalues. Dropping the first three principal components, or selecting ranges of intermediate components did not improve performance. The independent component sources were ordered by the magnitude of the weight vector, row of W, used to extract the source from the image.1 Best performance was obtained with the 130 independent components with the largest weight vectors. Performance with the ICA representation was significantly superior to Eigenfaces by a paired t-test (0 < 0.05). Graylevel Images PCA ICA Mutual Information .89 .10 .007 Percent Correct Recognition .83 .84 .87 Table 1: Mean mutual information between all pairs of 10 basis images, and between the original graylevel images. Face recognition performance is across all 200 images. For the task of recognizing faces across pose, a statistically independent basis set provided a more powerful representation for face images than a principal component representation in which only the second order statistics are decorrelated. IThe magnitude of the weight vector for optimally projecting the source onto the sloping part of the nonlinearity provides a measure of the variance of the original source (Tony Bell, personal communication). 820 M. S. Bartlett and T. 1. Sejnowski 2 Unsupervised Learning of Viewpoint Invariant Representations of Faces in an Attractor Network Cells in the primate inferior temporal lobe have been reported that respond selectively to faces despite substantial changes in viewpoint (Hasselmo, Rolls, Baylis, & Nalwa, 1989). Some cells responded independently of viewing angle, whereas other cells gave intermediate responses between a viewer-centered and an object centered representation. This section addresses how a system can acquire such invariance to viewpoint from visual experience. During natural visual experience, different views of an object or face tend to appear in close temporal proximity as an animal manipulates the object or navigates around it, or as a face changes pose. Capturing such temporal relationships in the input is a way to automatically associate different views of an object without requiring three dimensional descriptions. Attractor Network CoDlpetitive Hebbian Learning Figure 4: Model architecture. Hebbian learning can capture these temporal relationships in a feedforward system when the output unit activities are passed through a lowpass temporal filter (Foldiak, 1991; Wallis & Rolls, 1996). Such lowpass temporal filters have been related to the time course of the modifiable state of a neuron based on the open time of the NMDA channel for calcium influx (Rhodes, 1992). We show that 1) this lowpass temporal filter increases viewpoint invariance of face representations in a feedforward system trained with competitive Hebbian learning, and 2) when the input patterns to an attractor network are passed through a lowpass temporal filter, then a basic Hebbian weight update rule associates sequentially proximal input patterns into the same basin of attraction. This simulation used a subset of 100 images from Section 1, consisting of twenty faces at five poses each. Images were presented to the model in sequential order as the subject changed pose from left to right (Figure 4). The first layer is an energy model related to the output of V1 complex cells (Heeger, 1991). The images were filtered by a set of sine and cosine Gabor filters at 4 spatial scales and 4 orientations at 255 spatial locations. Sine and cosine outputs were squared and summed. The set of V1 model outputs projected to a second layer of 70 units, grouped into two Viewpoint Invariant Face Recognition 821 inhibitory pools. The third stage of the model was an attractor network produced by lateral interconnections among all of the complex pattern units. The feedforward and lateral connections were trained successively. 2.1 Competitive Hebbian learning of temporal relationships The Competitive Learning Algorithm (Rumelhart & Zipser, 1985) was extended to include a temporal lowpass filter on output unit activities (Bartlett & Sejnowski, 1996). This manipulation gives the winner in the previous time steps a competitive advantage for winning, and therefore learning, in the current time step. winner = maxj [y/t)] y/t) = AY} + (1 _ A)y/t-1) (1) The output activity of unit j at time t, y/t), is determined by the trace, or running average, of its activation, where y} is the weighted sum of the feedforward inputs, a is the learning rate, Xitl is the value of input unit i for pattern u, and 8t1 is the total amount of input activation for pattern u. The weight to each unit was constrained to sum to one. This algorithm was used to train the feedforward connections. There was one face pattern per time step and A was set to 1 between individuals. 2.2 Lateral connections in the output layer form an attractor network Hebbian learning of lateral interconnections, in combination with a lowpass temporal filter on the unit activities in (1), produces a learning rule that associates temporally proximal inputs into basins of attraction. We begin with a simple Hebbian learning rule P Wij = ~ L(Y~ - yO)(y} _ yO) t=1 (2) where N is the number of units, P is the number of patterns, and yO is the mean activity over all of the units. Replacing y~ with the activity trace y/t) defined in (1), substituting yO = Ayo + (1 - A)YO and multiplying out the terms leads to the following learning rule: +k2 [(y/t-1) _ yO)(y/t-1) _ yO)] (3) h k >'(1;>') d k (1_;)2 were 1= >. an 2= >. The first term in this equation is basic Hebbian learning, the second term associates pattern t with pattern t - 1, and the third term is Hebbian association of the trace activity for pattern t - 1. This learning rule is a generalization of an attractor network learning rule that has been shown to associate random input patterns 822 M. S. Bartlett and T. 1. Sejnowski into basins of attraction based on serial position in the input sequence (Griniasty, Tsodyks & Amit, 1993). The following update rule was used for the activation V of unit i at time t from the lateral inputs (Griniasty, Tsodyks, & Amit, 1993): Vi(t + lSt) = ¢ [I: Wij Vj(t) - 0] Where 0 is a neural threshold and ¢(x) = 1 for x > 0, and 0 otherwise. In these simulations, 0 = 0.007, N = 70, P = 100, yO = 0.03, and A = 0.5 gave kl = k2 = 1. 2.3 Results Temporal association in the feedforward connections broadened the pose tuning of the output units (Figure 5 Left) . When the lateral connections in the output layer were added, the attractor network acquired responses that were largely invariant to pose (Figure 5 Right). 80.8 :c ,.!g ~ 0.6 ... <3 ~0.4 Q) ~0 .2 o,....-o_-u, [J Same Face, with Trace o Same Face, no Trace -- Different Faces -600 -450 _30° _15° 0° 15° 30° 45° 60° ~Pose • Hebb plus trace DTesiset .. Griniasly 01. al. • Hebbonly -600 _45° _30° _15° 0° 15° 30° 45° 60° ~Pose Figure 5: Left: Correlation of the outputs of the feedforward system as a function of change in pose. Correlations across different views of the same face (- ) are compared to correlations across different faces (--) with the temporal trace parameter A = 0.5 and A = O. Right: Correlations in sustained activity patterns in the attractor network as a function of change in pose. Results obtained with Equation 3 (Hebb plus trace) are compared to Hebb only, and to the learning rule in Griniasty et al. (1993). Test set results for Equation 3 (open squares) were obtained by alternately training on four poses and testing on the fifth, and then averaging across all test cases. Attractor Network leA F N FIN % Correct % Correct 5 70 .07 1.00 .96 10 70 _14 .90 .86 20 70 .29 .61 .89 20 160 .13 .73 .89 Table 2: Face classification performance of the attractor network for four ratios of the number of desired memories, F, to the number of units, N. Results are compared to ICA for the same subset of faces. Viewpoint Invariant Face Recognition 823 Classification accuracy of the attractor network was calculated by nearest neighbor on the activity states (Table 2). Performance of the attractor network depends both on the performance of the feedforward system, which comprises its input, and on the ratio of the number of patterns to be encoded in memory, F, to the number of units, N, where each individual in the face set comprises one memory pattern. The attractor network performed well when this ratio was sufficiently high. The ICA representation also performed well, especially for N=20. The goal of this simulation was to begin with structured inputs similar to the responses of VI complex cells, and to explore the performance of unsupervised learning mechanisms that can transform these inputs into pose invariant responses. We showed that a lowpass temporal filter on unit activities, which has been related to the time course of the modifiable state of a neuron (Rhodes, 1992), cooperates with Hebbian learning to (1) increase the viewpoint invariance of responses to faces in a feedforward system, and (2) create basins of attraction in an attractor network which associate temporally proximal inputs. These simulations demonstrated that viewpoint invariant representations of complex objects such as faces can be developed from visual experience by accessing the temporal structure of the input. Acknowledgments This project was supported by Lawrence Livermore National Laboratory ISCR Agreement B291528, and by the McDonnell-Pew Center for Cognitive Neuroscience at San Diego. References Bartlett, M. Stewart, & Sejnowski, T., 1996. Unsupervised learning of invariant representations of faces through temporal association. Computational Neuroscience: Int. Rev. Neurobio. Suppl. 1 J.M Bower, Ed., Academic Press, San Diego, CA:317-322. Beymer, D. 1994. Face recognition under varying pose. In Proceedings of the 1994 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Los Alamitos, CA: IEEE Comput. Soc. Press: 756-61. Bell, A. & Sejnowski, T., (1997). The independent components of natural scenes are edge filters. Advances in Neural Information Processing Systems 9. Bell, A., & Sejnowski, T., 1995. An information Maximization approach to blind separation and blind deconvolution. Neural Compo 7: 1129-1159. Comon, P. 1994. Independent component analysis - a new concept? Signal Processing 36:287-314. Cottrell & Metcalfe, 1991. Face, gender and emotion recognition using Holons. In Advances in Neural Information Processing Systems 3, D. Touretzky, (Ed.), Morgan Kaufman, San Mateo, CA: 564 - 571. Foldiak, P. 1991. Learning invariance from transformation sequences. Neural Compo 3:194200. Griniasty, M., Tsodyks, M., & Amit, D. 1993. Conversion of temporal correlations between stimuli to spatial correlations between attractors. Neural Compo 5:1-17. Hasselmo M. Rolls E. Baylis G. & Nalwa V. 1989. Object-centered encoding by faceselective neurons in the cortex in the superior temporal sulcus of the monkey. Experimental Brain Research 75(2):417-29. Heeger, D. (1991). Nonlinear model of neural responses in cat visual cortex. Computational Models of Visual Processing, M. Landy & J. Movshon, Eds. MIT Press, Cambridge, MA. Makeig, S, Bell, AJ, Jung, T-P, and Sejnowski, TJ 1996. Independent component analysis of Electroencephalographic data, In: Advances in Neural Information Processing Systems 8, 145-151. Rhodes, P. 1992. The long open time of the NMDA channel facilitates the self-organization of invariant object responses in cortex. Soc. Neurosci. Abst. 18:740. Rumelhart, D. & Zipser, D. 1985. Feature discovery by competitive learning. Cognitive Science 9: 75-112. Turk, M., & Pentland, A. 1991. Eigenfaces for Recognition. J. Cog. Neurosci. 3(1):71-86. Wallis, G. & Rolls, E. 1996. A model of invariant object recognition in the visual system. Technical Report, Oxford University Department of Experimental Psychology.
1996
33
1,229
On the Effect of Analog Noise in Discrete-Time Analog Computations Wolfgang Maass Institute for Theoretical Computer Science Technische Universitat Graz* Abstract Pekka Orponen Department of Mathematics University of Jyvaskylat We introduce a model for noise-robust analog computations with discrete time that is flexible enough to cover the most important concrete cases, such as computations in noisy analog neural nets and networks of noisy spiking neurons. We show that the presence of arbitrarily small amounts of analog noise reduces the power of analog computational models to that of finite automata, and we also prove a new type of upper bound for the VC-dimension of computational models with analog noise. 1 Introduction Analog noise is a serious issue in practical analog computation. However there exists no formal model for reliable computations by noisy analog systems which allows us to address this issue in an adequate manner. The investigation of noise-tolerant digital computations in the presence of stochastic failures of gates or wires had been initiated by [von Neumann, 1956]. We refer to [Cowan, 1966] and [Pippenger, 1989] for a small sample of the nllmerous results that have been achieved in this direction. In all these articles one considers computations which produce a correct output not with perfect reliability, but with probability ~ t + p (for some parameter p E (0, t D· The same framework (with stochastic failures of gates or wires) hac; been applied to analog neural nets in [Siegelmann, 1994]. The abovementioned approaches are insufficient for the investigation of noise in analog computations, because in analog computations one has to be concerned not only with occasional total failures of gates or wires, but also with "imprecision", i.e. with omnipresent smaller (and occa<;ionally larger) perturbations of analog outputs • Klosterwiesgasse 32/2, A-BOlO Graz, Austria. E-mail: maass@igi.tu-graz.ac.at. t P. O. Box 35, FIN-4035l JyvaskyHi, Finland. E-mail: orponen@math.jyu.fi. Part of this work was done while this author was at the University of Helsinki, and during visits to the Technische Universitat Graz and the University of Chile in Santiago. On the Effect of Analog Noise in Discrete-Time Analog Computations 219 of internal computational units. These perturbations may for example be given by Gaussian distributions. Therefore we introduce and investigate in this article a notion of noise-robust computation by noisy analog systems where we assume that the values of intermediate analog values are moved according to some quite arbitrary probability distribution. We consider - as in the traditional framework for noisy digital computations - arbitrary computations whose output is correct with some given probability 2: ~ + P (for p E (O,~]) . We will restrict our attention to analog computation with digital output. Since we impose no restriction (such as continuity) on the type of operations that can be performed by computational units in an analog computational system, an output unit of such system can convert an analog value into a binary output via "thresholding". Our model and our Theorem 3.1 are somewhat related to the analysis of probabilistic finite automata in [Rabin, 1963]. However there the finiteness of the state space simplifies the setup considerably. [Casey, 1996] addresses the special case of analog computations on recurrent neural nets (for those types of analog noise that can move an internal state at most over a distance c) whose digital output is perfectly reliable (Le. p = 1/2 in the preceding notation).l The restriction to perfect reliability in [Casey, 1996] has immediate consequences for the types of analog noise processes that can be considered, and for the types of mathematical arguments that are needed for their investigation. In a computational model with perfect reliability of the output it cannot happen that an intermediate state §. occurs at some step t both in a computation for an input !!2 that leads to output "0" , and at step t in a computation for the same input "!!2" that leads to output "I" . Hence an analysis of perfectly reliable computations can focus on partitions of intermediate states §. according to the computations and the computation steps where they may occur. Apparently many important concrete cases of noisy analog computations require a different type of analysis. Consider for example the special case of a sigmoidal neural net (with thresholding at the output), where for each input the output of an internal noisy sigmoidal gate is distributed according to some Gaussian distribution (perhaps restricted to the range of all possible output values which this sigmoidal gate can actually produce). In this case an intermediate state §. of the computational system is a vector of values which have been produced by these Gaussian distributions. Obviously each such intermediate state ~ can occur at any fixed step t in any computation (in particular in computations with different network output for the same network input). Hence perfect reliability of the network output is unattainable in this case. For an investigation of the actual computational power of a sigmoidal neural net with Gaussian noise one haC) to drop the requirement of perfect reliability of the output, and one has to analyze how probable it is that a particular network output is given, and how probable it is that a certain intermediate state is assumed. Hence one has to analyze for each network input and each step t the different IThere are relatively few examples for nontrivial computations on common digital or analog computational models that can achieve perfect reliability of the output in spite of noisy internal components. Most constructions of noise-robust computational models rely on the replication of noisy computational units (see [von Neumann, 1956], [Cowan, 1966]). The idea of this method is that the average of the outputs of k identical noisy computational units (with stochastically independent noise processes) is with high probability close to the expected value of their output, if k is sufficiently large. However for any value of k there exists in general a small but nonzero probability that this average deviates strongly from its expected value. In addition, if one assumes that the computational unit that produces the output of the computations is also noisy, one cannot expect that the reliability of the output of the computation is larger than the reliability of this last computational unit. Consequently there exist many methods for reducing the error-probability of the output to a small value, but these methods cannot achieve error probability 0 at the output. 220 W. Maass and P. Orponen probability distributions over intermediate states §.. that are induced by computations of the noisy analog computational system. In fact, one may view the set of these probability distributions over intermediate states §.. as a generalized set of "states" of a noisy analog computational system. In general the mathematical structure of this generalized set of "states" is substantially more complex than that of the original set of intermediate states §.. • We have introduced in [Maass, Orponen, 1996] some basic methods for analyzing this generalized set of "states", and the proofs of the main results in this article rely on this analysis. The preceding remarks may illustrate that if one drops the assumption of perfect reliability of the output, it becomes more difficult to prove upper bounds for the power of noisy analog computations. We prove such upper bounds even for the case of stochastic dependencies among noises for different internal units, and for the case of nonlinear dependencies of the noise on the current internal state. Our results also cover noisy computations in hybrid analog/digital computational models, such as for example a neural net combined with a binary register, or a network of noisy spiking neurons where a neuron may temporarily assume the discrete state "not-firing". Obviously it becomes quite difficult to analyze the computational effect of such complex (but practically occuring) types of noise without a rigorous mathematical framework. We introduce in section 2 a mathematical framework that is general enough to subsume all these cases. The traditional case of noisy digital computations is captured as a special case of our definition. One goal of our investigation of the effect of analog noise is to find out which features of analog noise have the most detrimental effect on the computational power of an analog computational system. This turns out to be a nontrivial question.2 As a first step towards characterizing those aspects and parameters of analog noise that have a strong impact on the computational power of a noisy analog system, the proof of Theorem 3.1 (see [Maass, Orponen, 1996]) provides an explicit bound on the number of states of any finite automaton that can be implemented by an analog computational system with a given type of analog noise. It is quite surprising to see on which specific parameters of the analog noise the bound depends. Similarly the proofs of Theorem 3.4 and Theorem 3.5 provide explicit (although very large) upper bounds for the VC-dimension of noisy analog neural nets with batch input, which depend on specific parameters of the analog noise. 2 Preliminaries: Definitions and Examples An analog discrete-time computational system (briefly: computational system) M is defined in a general way as a 5-tuple (0, pO, F, 1:, s), where 0, the set of states, is a bounded subset of R d , po E 0 is a distinguished initial state, F ~ 0 is the set of accepting states, 1: is the input domain, and s : 0 x E ~ 0 is the transition function. To avoid unnecessary pathologies, we impose the conditions that 0 and F are Borel subsets of R d, and for each a E 1:, s(p, a) is a measurable function of p. We also assume that E contains a distinguished null value U, which may be used to pad the actual input to arbitrary length. The nonnull input domain is denoted by 1:0 = 1: - {U}. 2For example, one might think that analog noise which is likely to move an internal state over a large distance is more harmful than another type of analog noise which keeps an internal state within its neighborhood. However this intuition is deceptive. Consider the extreme case of analog noise in a Sigmoidal neural net which moves a gate output x E [-1,1] to a value in the e-neighborhood of -x. This type of noise moves some values x over large distances, but it appears to be less harmful for noise-robust computing than noise which moves x to an arbitrary value in the lOe-neighborhood of x . On the Effect of Analog Noise in Discrete-Time Analog Computations 221 The intended noise-free dynamics of such a system M is as follows. The system starts its computation in state pO, and on each single computation step on input element a E Eo moves from its current state p to its next state s(p, a). After the actual input sequence has been exhausted, M may still continue to make pure computation steps. Each pure computation step leads it from a state p to the state s(p, U). The system accepts its input if it enters a state in the class F at some point after the input has finished. For instance, the recurrent analog neural net model of [Siegelmann, Sontag, 1991] (also known as the "Brain State in a Box" model) is obtained from this general framework as follows. For a network N with d neurons and activation values between -1 and 1, the state space is 0 = [-1, 1] d. The input domain may be chosen as either E = R or E = {-l,O, I} (for "online" input) or E = R n (for "batch" input). Feedforward analog neural nets may also be modeled in the same manner, except that in this case one may wish to select as the state set 0 := ([-1, 1] U {dormant})d, where dormant is a distinguished value not in [-1, 1]. This special value is used to indicate the state of a unit whose inputs have not all yet been available at the beginning of a given computation step (e.g. for units on the l-th layer of a net at computation steps t < 1). The completely different model of a network of m stochastic spiking neurons (see e.g. [Maass, 1997]) is also a special case of our general framework. 3 Let us then consider the effect of noise in a computational system M. Let Z (p, B) be a function that for each state p E 0 and Borel set B ~ 0 indicates the probability of noise moving state p to some state in B. The function Z is called the noise process affecting M, and it should satisfy the mild conditions of being a stochastic kernel, i.e., for each p E 0, Z(p,.) should be a probability distribution, and for each Borel set B, Z(-, B) should be a measurable function. We assume that there is some measure IL over 0 so that Z(p, ·) is absolutely continuous with respect to It for each p EO, i.e. IL(B) = 0 implies Z(p, B) = 0 for every measurable B ~ 0 . By the Radon-Nikodym theorem, Z then possesses a density kernel with respect to IL, i.e. there exists a function z(·,·) such that for any state p E 0 and Borel set B ~ 0, Z(p, B) = JqEB z(p, q) dJL. We assume that this function z(',·) has values in [0,00) and is measurable. (Actually, in view of our other conditions this can be assumed without loss of generality.) The dynamics of a computational system M affected by a noise process Z is now defined as follows. 4 If the system starts in a state p, the distribution of states q obtained after a single computation step on input a E E is given by the density kernel 1f'a(P, q) = z(s(p, a), q). Note that as a composition of two measurable func3In this case one wants to set nsp := (U~=l [0, T)i U {not-firing})m, where T > 0 is a sufficiently large constant so that it suffices to consider only the firing history of the network during a preceding time interval of length T in order to determine whether a neuron fires (e.g. T = 30 ms for a biological neural system). If one partitions the time axis into discrete time windows [0, T) , [T, 2T) ,. .. , then in the noise-free case the firing events during each time window are completely determined by those in the preceding one. A component Pi E [0, T)i of a state in this set nsp indicates that the corresponding neuron i has fired exactly j times during the considered time interval, and it also specifies the j firing times of this neuron during this interval. Due to refractory effects one can choose 1< 00 for biological neural systems, e.g. 1= 15 for T = 30 ms. With some straightforward formal operations one can also write this state set nsp as a bounded subset of Rd for d:= l·m. 4We would like to thank Peter Auer for helpful conversations on this topic. 222 W. Maass and P. Orponen tions, 1ra is again a measurable function. The long-term dynamics of the system is given by a Markov process, where the distribution 1rza (p, q) of states after Ixal computation steps with input xa E E* starting in state p is defined recursively by 1rza (p,q) = IrEo1rz(p,r) '1ra(r,q) dp,. Let us denote by 1rz(q) the distribution 1rz(pO,q), i.e. the distribution of states of M after it has processed string x, starting from the initial state pO. Let p > 0 be the required reliability level. In the most basic version the system M accepts (rejects) some input x E I':o if IF 1rz (q) dll 2: ! + p (respectively ~ ~ - p). In less trivial cases the system may also perform pure computation steps after it has read all of the input. Thus, we define more generally that the system M recognizes a set L ~ I':o with reliability p if for any x E Eo: x E L ¢:> 11rzu (q) dp, 2: ~ + P for some u E {u}* x ¢ L ¢:> 11rzu (q) dp, ~ ~ - p for all u E {u}*. This covers also the case of batch input, where Ixl = 1 and Eo is typically quite large (e.g. Eo = Rn). 3 Results The proofs of Theorems 3.1, 3.4, 3.5 require a mild continuity assumption for the density functions z(r,·) , which is satisfied in all concrete cases that we have examined. We do not require any global continuity property over 0 for the density functions z(r,·) because there are important special cases (see [Maass, o rponen , 1996]), where the state space 0 is a disjoint union of subspaces 0 1 , .•• ,Ok with different measures on each subspace. We only assume that for some arbitrary partition of n into Borel sets 0 1 , ... ,Ok the density functions z(r,·) are uniformly continuous over each OJ , with moduli of continuity that can be bounded independently of r . In other words, we require that z(·, .) satisfies the following condition: We call a function 1r(" .) from 0 2 into R piecewise uniformly continuous if for every c> 0 there is a 8 > 0 such that for every rEO, and for all p, q E OJ, j = 1, ... , k: II p - q II ~ 8 implies 11r(r,p) - 1r(r, q)1 ~ c. (1) If z(',') satisfies this condition, we say that the re~mlting noise process Z is piecewise uniformly continuous. Theorem 3.1 Let L ~ I':o be a set of sequences over an arbitrary input domain Eo. Assume that some computational system M, affected by a piecewise uniformly continuous noise process Z, recognizes L with reliability p, for some arbitrary p > O. Then L is regular. The proof of Theorem 3.1 relies on an analysis of the space of probability density functions over the state set 0 . An upper bound on the number of states of a deterministic finite automaton that simulates M can be given in terms of the number k of components OJ of the state set 0 , the dimension and diameter of 0 , a bound on the values of the noise density function z , and the value of 8 for c = p/4p,(0) in condition (1). For details we refer to [Maass, Orponen, 1996].5 • '" A corresponding result is claimed in Corollary 3.1 of [Casey, 1996] for the special case On the Effect of Analog Noise in Discrete-TIme Analog Computations 223 Remark 3.2 In stark contrast to the results of [Siegelmann, Sontag, 1991} and [Maass, 1996} for the noise-free case, the preceding Theorem implies that both recurrent analog neural nets and recurrent networks of spiking neurons with online input from ~o can only recognize regular languages in the presence of any reasonable type of analog noise, even if their computation time is unlimited and if they employ arbitrary real-valued parameters. Let us say that a noise process Z defined on a set 0 ~ Rd is bounded by 11 if it can move a state P only to other states q that have a distance $ 11 from p in the LI -norm over Rd , Le. if its density kernel z has the property that for any p = (PI, ... ,Pd) and q = (ql, ... , qd) E 0, z(p, q) > 0 implies that Iqi - Pil $ 11 for i = 1, ... , d. Obviously 11-bounded noise processes are a very special class. However they provide an example which shows that the general upper bound of Theorem 3.1 is a sense optimal: Theorem 3.3 For every regular language L ~ {-1, 1}* there is a constant 11 > 0 such that L can be recognized with perfect reliability (i. e. p = ~) by a recurrent analog neural net in spite of any noise process Z bounded by 11. • We now consider the effect of analog noise on discrete time analog computations with batch-input. The proofs of Theorems 3.4 and 3.5 are quite complex (see [Maass, Orponen, 1996]). Theorem 3.4 There exists a finite upper bound for the VC-dimension of layered feedforward sigmoidal neural nets and feedforward networks of spiking neurons with piecewise uniformly continuous analog noise (for arbitrary real-valued inputs, Boolean output computed with some arbitrary reliability p > OJ and arbitrary realvalued ''programmable parameters") which does not depend on the size or structure of the network beyond its first hidden layer. • Theorem 3.5 There exists a finite upper bound for the VC-dimension of recurrent sigmoidal neural nets and networks of spiking neurons with piecewise uniformly continuous analog noise (for arbitrary real valued inputs, Boolean output computed with some arbitrary reliability p > 0, and arbitrary real valued ''programmable parameters") which does not depend on the computation time of the network, even if the computation time is allowed to vary for different inputs. • 4 Conclusions We have introduced a new framework for the analysis of analog noise in discretetime analog computations that is better suited for "real-world" applications and of recurrent neural nets with bounded noise and p = 1/2 , i.e. for certain computations with perfect reliability. This case may not require the consideration of probability density functions. However it turns out that the proof for this special case in [Casey, 1996J is wrong. The proof of Corollary 3.1 in [Casey, 1996J relies on the argument that a compact set "can contain only a finite number of disjoint sets with Jlonempty interior" . This argument is wrong, as the counterexample of the intervals [1/{2i + 1), 1/2iJ for i = 1,2, ... shows. These infinitely many disjoint intervals are all contained in the compact set [0, 1 J . In addition, there is an independent problem with the structure of the proof of Corollary 3.1 in [Casey, 1996J. It is derived as a consequence ofthe proof of Theorem 3.1 in [Casey, 1996]. However that proof relies on the assumption that the recurrent neural net accepts a regular language. Hence the proof via probability density functions in [Maass, Orponen, 1996] provides the first valid proof for the claim of Corollary 3.1 in [Casey, 1996]. 224 W Maass and P. Orponen more flexible than previous models. In contrast to preceding models it also covers important concrete cases such as analog neural nets with a Gaussian distribution of noise on analog gate outputs, noisy computations with less than perfect reliability, and computations in networks of noisy spiking neurons. Furthermore we have introduced adequate mathematical tools for analyzing the effect of analog noise in this new framework. These tools differ quite strongly from those that have previously been used for the investigation of noisy computations. We show that they provide new bounds for the computational power and VCdimension of analog neural nets and networks of spiking neurons in the presence of analog noise. Finally we would like to point out that our model for noisy analog computations can also be applied to completely different types of models for discrete time analog computation than neural nets, such as arithmetical circuits, the random access machine (RAM) with analog inputs, the parallel random access machine (PRAM) with analog inputs, various computational discrete-time dynamical systems and (with some minor adjustments) also the BSS model [Blum, Shub, Smale, 1989]. Our framework provides for each of these models an adequate definition of noise-robust computation in the presence of analog noise, and our results provide upper bounds for their computational power and VC-dimension in terms of characteristica of their analog noise. References [Blum, Shub, Smale, 1989] L. Blum, M. Shub, S. Smale, On a theory of computation over the real numbers: NP-completeness, recursive functions and universal machines. Bulletin of the Amer. Math. Soc. 21 (1989), 1-46. [Casey, 1996] M. Casey, The dynamics of discrete-time computation, with application to recurrent neural networks and finite state machine extraction. Neural Computation 8 (1996), 1135-1178. [Cowan, 1966] J. D. Cowan, Synthesis of reliable automata from unreliable components. Automata Theory (E. R. Caianiello, ed.), 131-145. Academic Press, New York, 1966. [Maass, 1996] W. Maass, Lower bounds for the computational power of networks of spiking neurons. Neural Computation 8 (1996), 1-40. [Maass, 1997] W. Maass, Fast sigmoidal networks via spiking neurons, to appear in Neural Computation 9, 1997. FTP-host: archive.cis.ohio-state.edu, FTP-filename: /pub/neuroprose /maass.sigmoidal-spiking.ps.Z. [Maass, Orponen, 1996] W. Maass, P. Orponen, On the effect of analog noise in discrete-time analog computations Uournal version), submitted for publication; see http://www.math.jyu.fi/ ... orponen/papers/noisyac.ps. [Pippenger, 1989] N. Pippenger, Invariance of complexity measures for networks with unreliable gates. J. Assoc. Comput. Mach. 36 (1989), 531-539. [Rabin, 1963] M. Rabin, Probabilistic automata. Information and Control 6 (1963), 230245. [Siegelmann, 19941 H. T. Siegelmann, On the computational power of probabilistic and faulty networks. Proc. 21st International Colloquium on Automata, Languages, and Programming, 23-34. Lecture Notes in Computer Science 820, Springer-Verlag, Berlin, 1994. [Siegelmann, Sontag, 1991] H. T. Siegelmann, E. D. Sontag, Turing computability with neural nets. Appl. Math. Letters 4(6) (1991), 77-80. [von Neumann, 1956] J. von Neumann, Probabilistic logics and the synthesis of reliable organisms from unreliable components. Automata Studies (C. E. Shannon, J. E. McCarthy, eds.), 329-378. Annals of Mathematics Studies 34, Princeton University Press, Princeton, NJ, 1956.
1996
34
1,230
Combined Weak Classifiers Chuanyi Ji and Sheng Ma Department of Electrical, Computer and System Engineering Rensselaer Polytechnic Institute, Troy, NY 12180 chuanyi@ecse.rpi.edu, shengm@ecse.rpi.edu Abstract To obtain classification systems with both good generalization performance and efficiency in space and time, we propose a learning method based on combinations of weak classifiers, where weak classifiers are linear classifiers (perceptrons) which can do a little better than making random guesses. A randomized algorithm is proposed to find the weak classifiers. They· are then combined through a majority vote. As demonstrated through systematic experiments, the method developed is able to obtain combinations of weak classifiers with good generalization performance and a fast training time on a variety of test problems and real applications. 1 Introduction The problem we will investigate in this work is how to develop a classifier with both good generalization performance and efficiency in space and time in a supervised learning environment. The generalization performance is measured by the probability of classification error of a classifier. A classifier is said to be efficient if its size and the (average) time needed to develop such a classifier scale nicely (polynomiaUy) with the dimension of the feature vectors, and other parameters in the training algorithm. The method we propose to tackle this problem is based on combinations of weak classifiers [8][6] , where the weak classifiers are the classifiers which can do a little better than random guessing. It has been shown by Schapire and Freund [8][4] that the computational power of weak classifiers is equivalent to that of a welltrained classifier, and an algorithm has been given to boost the performance of weak classifiers. What has not been investigated is the type of weak classifiers that can be used and how to find them. In practice, the ideas have been applied with success in hand-written character recognition to boost the performance of an already well-trained classifier. But the original idea on combining a large number of weak classifiers has not been used in solving real problems. An independent work Combinations of Weak Classifiers 495 by Kleinberg[6] suggests that in addition to a good generalization performance, combinations of weak classifiers also provide advantages in computation time, since weak classifiers are computationally easier to obtain than well-trained classifiers. However, since the proposed method is based on an assumption which is difficult to realize, discrepancies have been found between the theory and the experimental results[7]. The recent work by Breiman[1][2] also suggests that combinations of classifiers can be computationally efficient, especially when used to learn large data sets. The focus of this work is to investigate the following problems: (1) how to find weak classifiers, (2) what are the performance and efficiency of combinations of weak classifiers, and (3) what are the advantages of using combined weak classifiers compared with other pattern classification methods? We will develop a randomized algorithm to obtain weak classifiers. We will then provide simulation results on both synthetic real problems to show capabilities and efficiency of combined weak classifiers. The extended version of this work with some of the theoretical analysis can be found in [5]. 2 Weak Classifiers In the present work, we choose linear classifiers (perceptrons) as weak classifiers. Let t - ~ be the required generalization error of a classifier, where v 2: 2, is called the weakness factor which is used to characterize the strength of a classifier. The larger the v, the weaker the weak classifier. A set of weak classifiers are combined through a simple majority vote. 3 Algorithm Our algorithm for combinations of weak classifiers consists of two steps: (1) generating individual weak classifiers through a simple randomized algorithm; and (2) combining a collection of weak classifiers through a simple majority vote. Three parameters need to be chosen a priori for the algorithm: a weakness factor v, a number (J (~ ~ (J < 1) which will be used as a threshold to partition the training set, and the number of weak classifiers 2£ + 1 to be generated, where £ is a positive integer. 3.1 Partitioning the Training Set The method we use to partition a training set is motivated by what given in [4]. Suppose a combined classifier consists of K (K ~ 1) weak classifiers already. In order to generate a (new) weak classifier, the entire training set of N training samples is partitioned into two subsets: a set of Ml samples which contain all the misclassified samples and a small fraction of samples correctly-classified by the existing combined classifier; and the remaining N - Ml training samples. The set of Ml samples are called "cares", since they will be used to select a new weak classifier, while the rest of the samples are the "don't-cares". The threshold (J is used to determine which samples should be assigned as cares. For instance, for the n-th training sample (1 ~ n ~ N), the performance index a( n) is recorded, where a( n) is the fraction of the weak classifiers in the existing combined classifier which classify the n-th sample correctly. If a(n) < (J, this sample is assigned to the cares. Otherwise, it is a don't-care. This is done for all N samples. 496 C. Ji and S. Ma Through partitioning a training set in this way, a newly-generated weak classifier is forced to learn the samples which have not been learned by the existing weak classifiers. In the meantime, a properly-chosen () can ensure that enough samples are used to obtain each weak classifier. 3.2 Random Sampling To achieve a fast training time, we obtain a weak classifier by randomly sampling the classifier-space of all possible linear classifiers. Assume that a feature vector x E Rd is distributed over a compact region D. The direction of a hyperplane characterized by a linear classifier with a weight vector, is first generated by randomly selecting the elements of the weight vector based on a uniform distribution over (-1, l)d . Then the threshold of the hyperplane is determined by randomly picking an xED, and letting the hyperplane pass through x . This will generate random hyperplanes which pass through the region D, and whose directions are randomly distributed in all directions. Such a randomly selected classifier will then be tested on all the cares. If it misclassifies a fraction of cares no more than k - ~ - € (€ > 0 and small), the classifier is kept and will be used in the combination. Otherwise, it is discarded. This process is repeated until a weak classifier is finally obtained. A newly-generated weak classifier is then combined with the existing ones through a simple majority vote. The entire training set will then be tested on the combined classifier to result in a new set of cares, and don't-cares. The whole process will be repeated until the total number 2L + 1 of weak classifiers are generated. The algorithm can be easily extended to multiple classes. Details can be found in [5] . 4 Experimental Results Extensive simulations have been carried out on both synthetic and real problems using our algorithm. One synthetic problem is chosen to test the efficiency of our method. Real applications from standard data bases are selected to compare the generalization performance of combinations of weak classifiers (CW) with that of other methods such as K-Nearest-Neighbor classifiers (K-NN)l, artificial neural networks (ANN), combinations of neural networks (CNN), and stochastic discriminations (SD). 4.1 A Synthetic Problem: Two Overlapping Gaussians To test the scaling properties of combinations of weak classifiers, a non-linearly separable problem is chosen from a standard database called ELENA 2. The problem is a two-class classification problem, where the distributions of samples in both classes are multi-variate Gaussians with the same mean but different variances for each independent variable. There is a considerable amount of overlap between the samples in two classes, and the problem is non-linearly separable. The average generalization error and the standard deviations are given in Figure 1 for our algorithm based on 20 runs, and for other classifiers. The Bayes error is also given to show the theoretical limit. The results show that the performance of kNN degrades very quickly. The performance of ANN is better than that of kNN but still deviates more and more from the Bayes error as d gets large. The combination of weak classifiers 1 The best result of different k is reported. 2 I pu b I neural-nets I ELEN AI databases IBenchmarks. ps. Z on ft p.dice. ucl. ac. be Combinations of Weak Classifiers - "- k-NN -_ ANN 10 -0- cw - ..... IV_ " I., Figure 1: Performance versus the dimension of the feature vectors Algorithms Card1 Diabetes1 Gene1 (%) Error /er (%) Error / er (%1 Error/ er Combined Weak Classifiers 11.3/ 0.85 22.70/ 0.70 11.80/0.52 k Nearest Neighbor 15.67 25.8 22.87 Neural Networks 13.64/ 0.85 23.52/ 0.72 13.47/0.44 Combined Neural Networks 13.02/0.33 22.79/0.57 12.08/0.23 497 Table 1: Performance on Card1, Diabetes1 and Gene1. er: standard deviation continues to follow the trend of the Bayes error. 4.2 Proben1 Data Sets Three data sets, Card1, Diabetes1 and Gene1 were selected to test our algorithm from Proben1 databases which contain data sets from real applications3 . Card1 data set is for a problem on determining whether a credit-card application from a customer can be approved based on information given in 51-dimensional feature vectors. 345 out of 690 examples are used for training and the rest for testing. Diabetes1 data set is for determining whether diabetes is present based on 8-dimensional input patterns. 384 examples are used for training and the same number of samples for testing. Gene1 data set is for deciding whether a DNA sequence is from a donor, an acceptor or neither from 120 dimensional binary feature vectors. 1588 samples out of total of 3175 were used for training, and the rest for testing. The average generalization error as well as the standard deviations are reported in Table 1. The results from combinations of weak classifiers are based on 25 runs. The results of neural networks and combinations of well-trained neural networks are from the database. As demonstrated by the results, combinations of weak classifiers have been able to achieve the generalization performance comparable to or better than that of combinations of well-trained neural networks. 4.3 Hand-written Digit Recognition Hand-written digit recognition is chosen to test our algorithm, since one of the previously developed method on combinations of weak classifiers (stochastic discrimination[6]) was applied to this problem. For the purpose of comparison, the 3 Available by anonymous ftp from ftp.ira.uka.de, as / pub/papers/techreports/1994/1994-21. ps.z. 498 C. Ji and S. Ma Algorithms (%) Error/O' Combined Weak Classifiers 4.23/ 0.1 k Nearest Neighbor 4.84 Neural Networks 5.33 Stochastic Discriminations 3.92 Table 2: Performance on handwritten digit recognition. Parameters Gaussians Card1 Diabetes1 Gene1 Digits 1/2 + l/v 0.51 0.51 0.51 0.55 0.54 e 0.51 0.51 0.54 0.54 0.53 2L+1 2000 1000 1000 4000 20000 A verage Tries 2 3 7 4 2 Table 3: Parameters used in our experiments. same set of data as used in [6](from the NIST data base) is utilized to train and to test our algorithm. The data set contains 10000 digits written by different people. Each digit is represented by 16 by 16 black and white pixels. The first 4997 digits are used to form a training set, and the rest are for testing. Performance of our algorithm, k-NN, neural networks, and stochastic discriminations are given in Table 2. The results for our methods are based on 5 runs, while the results for the other methods are from [6]. The results show that the performance of our algorithm is slightly worse (by 0.3%) than that of stochastic discriminations, which uses a different method for multi-class classification [6] . 4.4 Effects of The Weakness Factor Experiments are done to test the effects of v on the problem of two 8-dimensional overlapping Gaussians. The performance and the average training time (CPUtime on Sun Spac-10) of combined weak classifiers based on 10 runs are given for different v's in Figures 2 and 3, respectively. The results indicate as v increases an individual weak classifier is obtained more quickly, but more weak classifiers are needed to achieve good performance. When a proper v is chosen, a nice scaling property can be observed in training time. A record of the parameters used in all the experiments on real applications are provided in Table 3. The average tries, which are the average number of times needed to sample the classifier space to obtain an acceptable weak classifier, are also given in the table to characterize the training time for these problems. 4.5 Training Time To compare learning time with off-line BackPropagation (BP), feedforward two layer neural network with 10 sigmoidal hidden units are trained by gradient-descent to learn the problem on the two 8-dimensional overlapping Gaussians. 2500 training samples are used. The performance versus CPU time4 are plotted for both our algorithm and BP in Figure 4. For our algorithm, 2000 weak classifiers are combined. For BP, 1000 epoches are used. The figure shows that our algorithm is much faster than the BP algorithm. Moreover, when several well-trained neural networks are combined to achieve a better performance, the cost on training time will be 4Both algorithms are run on a Sun Sparc-lO sun workstation Combinations of Weak Classifiers 499 '°0 200 400 too 100 1000 1200 1400 ,'00 1100 2000 NuntMI'oII .... ~ Figure 2: Performance versus the number of weak classifiers for different 1/. nu: 1/. I' -0- 1 AwaO 005 -x- lJnu-o.Ol0 ~ • -·- · 1~O'5 ~ -+-: lA'1u-002O u , : lhN-0025 .' , . _ "::;~~=1;:~-~:~~~~-~~:::: : ::::: m _ _ _ _ _ _ _ _ ~oIIw.-a...r ... Figure 3: Training time versus the number of weak classifiers for different 1/. even higher. Therefore, compared to combinations of well-trained neural networks, combining weak classifiers is computationally much cheaper. 5 Discussions From the experimental results, we observe that the performance of the combined weak classifiers is comparable or even better than combinations of well-trained classifiers, and out-performs individual neural network classifiers and k-Nearest Neighbor classifiers. In the meantime whereas the k-nearest neighbor classifiers suffer from the curse of dimensionality, a nice scaling property in terms of the dimension of feature vectors has been observed for combined weak classifiers. Another .. , , .. , , 35 ; IS l --_ TranngcurteollBP T_CV't'eollBP T,......cuwolCW - r_cwy.otcw I v_ -1- _ I- _ _ __ t ____ _ 10 -----.. I .............. --t-___ , Figure 4: Performance versus CPU time 500 C. Ii and S. Ma important observation obtained from the experiments is that the weakness factor directly impacts the size of a combined classifier and the training time. Therefore, the choice of the weakness factor is important to obtain efficient combined weak classifiers. It has been shown in our theoretical analysis on learning an underlying perceptron [5] that v should be at least large as O( dlnd) to obtain a polynomial training time, and the price paid to accomplish this is a space-complexity which is polynomial in d as well. This cost can be observed from our experimental results for the need of a large number of weak classifiers. Acknowledgement Specials thanks are due to Tin Kan Ho for providing NIST data, related references and helpful discussions. Support from the National Science Foundation (ECS9312594 and (CAREER) IRI-9502518) is gratefully acknowledged. References [1] L. Breiman, "Bias, Variance and Arcing Classifiers," Technical Report, TR460, Department of Statistics, University of California, Berkeley, April, 1996. [2] 1. Breiman, "Pasting, Bites Together for Prediction in Large Data sets and On-Line," ftp.stat.berkeley.edu/users/breiman, 1996. [3] H. Drucker, R. Schapire and P. Simard, "Improving Performance in Neural Networks Using a Boosting Algorithm," Neural Information Processing Symposium, 42-49, 1993. [4] Y. Freund and R. Schapire, "A Decision-Theoretic Generalization of On-Line Learning and An Application to Boosting," http://www.research.att.com/orgs/ssr/people/yoav or schapire. [5] C. Ji and S. Ma, "Combinations of Weak Classifiers," IEEE Trans. Neural Networks, Special Issue on Neural Networks and Pattern Recognition, vol. 8, 32-42, Jan., 1997. [6] E.M. Kleinberg, "Stochastic Discrimination," Annals of Mathematics and Artificial Intelligence, voU , 207-239, 1990. [7] E.M. Kleinberg and T. Ho, "Pattern Recognition by Stochastic Modeling," Proceedings of the Third International Workshop on Frontiers in Handwriting Recognition, 175-183, Buffalo, May 1993. [8] R.E. Schapire, "The Strength of Weak Learnability," Machine Learning, vol. 5, 197-227, 1990.
1996
35
1,231
Reinforcement Learning for Dynamic C·hannel Allocation in Cellular Telephone Systems Satinder Singh Department of Computer Science University of Colorado Boulder, CO 80309-0430 bavej a@cs.colorado.edu Dimitri Bertsekas Lab. for Info. and Decision Sciences MIT Cambridge, MA 02139 bertsekas@lids.mit.edu Abstract In cellular telephone systems, an important problem is to dynamically allocate the communication resource (channels) so as to maximize service in a stochastic caller environment. This problem is naturally formulated as a dynamic programming problem and we use a reinforcement learning (RL) method to find dynamic channel allocation policies that are better than previous heuristic solutions. The policies obtained perform well for a broad variety of call traffic patterns. We present results on a large cellular system with approximately 4949 states. In cellular communication systems, an important problem is to allocate the communication resource (bandwidth) so as to maximize the service provided to a set of mobile callers whose demand for service changes stochastically. A given geographical area is divided into mutually disjoint cells, and each cell serves the calls that are within its boundaries (see Figure 1a). The total system bandwidth is divided into channels, with each channel centered around a frequency. Each channel can be used simultaneously at different cells, provided these cells are sufficiently separated spatially, so that there is no interference between them. The minimum separation distance between simultaneous reuse of the same channel is called the channel reuse constraint. When a call requests service in a given cell either a free channel (one that does not violate the channel reuse constraint) may be assigned to the call, or else the call is blocked from the system; this will happen if no free channel can be found. Also, when a mobile caller crosses from one cell to another, the call is "handed off" to the cell of entry; that is, a new free channel is provided to the call at the new cell. If no such channel is available, the call must be dropped/disconnected from the system. RLfor Dynamic Channel Allocation 975 One objective of a channel allocation policy is to allocate the available channels to calls so that the number of blocked calls is minimized. An additional objective is to minimize the number of calls that are dropped when they are handed off to a busy cell. These two objectives must be weighted appropriately to reflect their relative importance, since dropping existing calls is generally more undesirable than blocking new calls. To illustrate the qualitative nature of the channel assignment decisions, suppose that there are only two channels and three cells arranged in a line. Assume a channel reuse constraint of 2, i.e., a channel may be used simultaneously in cells 1 and 3, but may not be used in channel 2 if it is already used in cell 1 or in cell 3. Suppose that the system is serving one call in cell 1 and another call in cell 3. Then serving both calls on the same channel results in a better channel usage pattern than serving them on different channels, since in the former case the other channel is free to be used in cell 2. The purpose of the channel assignment and channel rearrangement strategy is, roughly speaking, to create such favorable usage patterns that minimize the likelihood of calls being blocked. We formulate the channel assignment problem as a dynamic programming problem, which, however, is too complex to be solved exactly. We introduce approximations based on the methodology of reinforcement learning (RL) (e.g., Barto, Bradtke and Singh, 1995, or the recent textbook by Bertsekas and Tsitsiklis, 1996). Our method learns channel allocation policies that outperform not only the most commonly used policy in cellular systems, but also the best heuristic policy we could find in the literature. 1 CHANNEL ASSIGNMENT POLICIES Many cellular systems are based on a fixed assignment (FA) channel allocation; that is, the set of channels is partitioned, and the partitions are permanently assigned to cells so that all cells can use all the channels assigned to them simultaneously without interference (see Figure 1a). When a call arrives in a cell, if any preassigned channel is unused; it is assigned, else the call is blocked. No rearrangement is done when a call terminates. Such a policy is static and cannot take advantage of temporary stochastic variations in demand for service. More efficient are dynamic channel allocation policies, which assign channels to different cells, so that every channel is available to every cell on a need basis, unless the channel is used in a nearby cell and the reuse constraint is violated. The best existing dynamic channel allocation policy we found in the literature is Borrowing with Directional Channel Locking (BDCL) of Zhang & Yum (1989). It numbers the channels from 1 to N, partitions and assigns them to cells as in FA. The channels assigned to a cell are its nominal channels. If a nominal channel is available when a call arrives in a cell, the smallest numbered such channel is assigned to the call. If no nominal channel is available, then the largest numbered free channel is borrowed from the neighbour with the most free channels. When a channel is borrowed, careful accounting of the directional effect of which cells can no longer use that channel because of interference is done. The call is blocked if there are no free channels at all. When a call terminates in a cell and the channel so freed is a nominal channel, say numbered i, of that cell, then if there is a call in that cell on a borrowed channel, the call on the smallest numbered borrowed channel is reassigned to i and the borrowed channel is returned to the appropriate cell. If there is no call on a borrowed channel, then if there is a call on a nominal channel numbered larger than i, the call on the highest numbered nominal channel is reassigned to i. If the call just terminated was itself on a borrowed channel, the 976 S. Singh and D. Bertsekas call on the smallest numbered borrowed channel is reassigned to it and that channel is returned to the cell from which it was borrowed. Notice that when a borrowed channel is returned to its original cell, a nominal channel becomes free in that cell and triggers a reassignment. Thus, in the worst case a call termination in one cell can sequentially cause reassignments in arbitrarily far away cells making BDCL somewhat impractical. BOCL is quite sophisticated and combines the notions of channel-ordering, nominal channels, and channel borrowing. Zhang and Yum (1989) show that BOCL is superior to its competitors, including FA. Generally, BOCL has continued to be highly regarded in the literature as a powerful heuristic (Enrico et.al., 1996). In this paper, we compare the performance of dynamic channel allocation policies learned by RL with both FA and BOCL. 1.1 DYNAMIC PROGRAMMING FORMULATION We can formulate the dynamic channel allocation problem using dynamic programming (e.g., Bertsekas, 1995). State transitions occur when channels become free due to call departures, or when a call arrives at a given cell and wishes to be assigned a channel, or when there is a handoff, which can be viewed as a simultaneous call departure from one cell and a call arrival at another cell. The state at each time consists of two components: (1) The list of occupied and unoccupied channels at each cell. We call this the configuration of the cellular system. It is exponential in the number of cells. (2) The event that causes the state transition (arrival, departure, or handoff). This component of the state is uncontrollable. The decision/control applied at the time of a call departure is the rearrangement of the channels in use with the aim of creating a more favorable channel packing pattern among the cells (one that will leave more channels free for future assignments). Unlike BDCL, our RL solution will restrict this rearrangement to the cell with the current call departure. The control exercised at the time of a call arrival is the assignment of a free channel, or the blocking of the call if no free channel is currently available. In general, it may also be useful to do admission control, i.e., to allow the possibility of not accepting a new call even when there exists a free channel to minimize the dropping of ongoing calls during handoff in the future. We address admission control in a separate paper and here restrict ourselves to always accepting a call if a free channel is available. The objective is to learn a policy that assigns decisions (assignment or rearrangement depending on event) to each state so as to maximize J = E {l CO e- f3te(t)dt} , where E{-} is the expectation operator, e(t) is the number of ongoing calls at time t, and j3 is a discount factor that makes immediate profit more valuable than future profit. Maximizing J is equivalent to minimizing the expected (discounted) number of blocked calls over an infinite horizon. 2 REINFORCEMENT LEARNING SOLUTION RL methods solve optimal control (or dynamic programming) problems by learning good approximations to the optimal value function, J*, given by the solution to RLfor Dynamic Channel Allocation 977 the Bellman optimality equation which takes the following form for the dynamic channel allocation problem: J(x) Ee{ max [E~dc(x,a,~t)+i(~t)J(Y)}]}, aEA(r,e) (1) where x is a configuration, e is the random event (a call arrival or departure), A( x, e) is the set of actions available in the current state (x, e), ~t is the random time until the next event, c(x, a, ~t) is the effective immediate payoff with the discounting, and i(~t) is the effective discount for the next configuration y. RL learns approximations to J* using Sutton's (1988) temporal difference (TD(O)) algorithm. A fixed feature extractor is used to form an approximate compact representation of the exponential configuration of the cellular array. This approximate representation forms the input to a function approximator (see Figure 1) that learns/stores estimates of J*. No partitioning of channels is done; all channels are available in each cell. On each event, the estimates of J* are used both to make decisions and to update the estimates themselves as follows: Call Arrival: When a call arrives, evaluate the next configuration for each free channel and assign the channel that leads to the configuration with the largest estimated value. If there is no free channel at all, no decision has to be made. Call Termination: When a call terminates, one by one each ongoing call in that cell is considered for reassignment to the just freed channel; the resulting configurations are evaluated and compared to the value of not doing any reassignment at all. The action that leads to the highest value configuration is then executed. On call arrival, as long as there is a free channel, the number of ongoing calls and the time to next event do not depend on the free channel assigned. Similarly, the number of ongoing calls and the time to next event do not depend on the rearrangement done on call departure. Therefore, both the sample immediate payoff which depends on the number of ongoing calls and the time to next event, and the effective discount factor which depends only on the time to next event are independent of the choice of action. Thus one can choose the current best action by simply considering the estimated values of the next configurations. The next configuration for each action is deterministic and trivial to compute. When the next random event occurs, the sample payoff and the discount factor become available and are used to update the value function as follows: on a transition from configuration x to y on action a in time ~t, (1- a)Jo/d(X) + a (c(x, a, ~t) + i(~t)Jo/d(y» (2) where x is used to indicate the approximate feature-based representation of x. The parameters ofthe function approximator are then updated to best represent Jnew(x) using gradient descent in mean-squared error (Jnew(x) - JO/d(x»2 . 3 SIMULATION RESULTS Call arrivals are modeled as Poisson processes with a separate mean for each cell, and call durations are modeled with an exponential distribution. The first set of results are on the 7 by 7 cellular array of Figure ??a with 70 channels (roughly 7049 configurations) and a channel reuse constraint of 3 (this problem is borrowed from Zhang and Yum's (1989) paper on an empirical comparison of BDCL and its competitors). Figures 2a, b & c are for uniform call arrival rates of 150, 200, and 350 calls/hr respectively in each cell. The mean call duration for all the experiments 978 S. Singh and D. Bertsekas reported here is 3 minutes. Figure 2d is for non-uniform call arrival rates. Each curve plots the cumulative empirical blocking probability as a function of simulated time. Each data point is therefore the percentage of system-wide calls that were blocked up until that point in time. All simulations start with no ongoing calls. a) b) Con figuration Feature Extractor Features TO(O) Training , , ,-----FunCtion Value , (Availability Ap5ioXimator and I , Packing) , Figure 1: a) Cellular Array. The market area is divided up into cells, shown here as hexagons. The available bandwidth is divided into channels. Each cell has a base station responsible for calls within its area. Calls arrive randomly, have random durations and callers may move around in the market area creating handoffs. The channel reuse constraint requires that there be a minimum distance between simultaneous reuse of the same channel. In a fixed assignment channel allocation policy (assuming a channel reuse constraint of 3), the channels are partitioned into 7 lots labeled 1 to 7 and assigned to the cells in the compact pattern shown here. Note that the minimum distance between cells with the same number is at least three. b) A block diagram of the RL system. The exponential configuration is mapped into a feature-based approximate representation which forms the input to a function approximation system that learns values for configurations. The parameters of the function approximator are trained using gradient descent on the squared TD(O) error in value function estimates (c.L Equation 2). The RL system uses a linear neural network and two sets of features as input: one availability feature for each cell and one packing feature for each cell-channel pair. The availability feature for a cell is the number offree channels in that cell, while the packing feature for a cell-channel pair is the number of times that channel is used in a 4 cell radius. Other packing features were tried but are not reported because they were insignificant. The RL curves in Figure 2 show the empirical blocking probability whilst learning. Note that learning is quite rapid. As the mean call arrival rate is increased the relative difference between the 3 algorithms decreases. In fact, FA can be shown to be optimal in the limit of infinite call arrival rates (see McEliece and Sivarajan, 1994). With so many customers in every cell there are no short-term fluctuations to exploit. However, as demonstrated in Figure 2, for practical traffic rates RL consistently gives a big win over FA and a smaller win over BnCL. One difference between RL and BnCL is that while the BnCL policy is independent of call traffic, RL adapts its policy to the particulars of the call traffic it is trained on and should therefore be less sensitive to different patterns of non-uniformity of call traffic across cells. Figure 3b presents multiple sets of bar-graphs of asymptotic blocking probabilities for the three algorithms on a 20 by 1 cellular array with 24 channels and a channel reuse constraint of 3. For each set, the average per-cell call arrival rate is the same (120 calls/hr; mean duration of 3 minutes), but the pattern of call arrival rates across the 20 cells is varied. The patterns are shown on the left of the bar-graphs and are explained in the caption of Figure 3b. From Figure 3b it is apparent that RL is much less sensitive to varied patterns of non-uniformity than both BnCL and FA. We have showed that RL with a linear function approximator is able to find better dynamic channel allocation policies than the BnCL and FA policies. Using nonlinear neural networks as function approximators for RL did in some cases improve RLfor Dynamic Channel Allocation a) c) ~ .. t. J '" j ~ .oo CD ~ ~ w I" r .... h. ...... 150 calslhr FA BDCL -- RL 2500.0 5000.0 Time 350 calslhr .... ~==:;c::-==;::~.=.==:J ~ L / ~.... / l FA B~L go." :Ii! ! "20 ! .t. E w RL ......... -----26...., .. ..,.. .• -----... ~O'. TIIYIe b) d) ~ ! 0.20 .0 £ '" :i ~ ell 0.10 i E w -. . ..... ~ i £ .20 go :Ii! ~ CD 200 calslhr 2500.0 TIIYIe Non-Unilorm FA BDCL RL ...... FA BDCL ~ .. t. :~ .,...----------1 ~ RL ~ .......... -----26--.. ~ .. -----.... ...J .• TIIYIe 979 Figure 2: a), b)t c) & d) These figures compare performance of RL, FA, and BDCL on the 7 by 7 cellular array of Figure lao The means of the call arrival (Poisson) processes are shown in the graph titles. Each curve presents the cumulative empirical blocking probability as a function of time elapsed in minutes. All simulations start with no ongoing calls and therefore the blocking probabilities are low in the early minutes of the performance curves. The RL curves presented here are for a linear function approximator and show performance while learning. Note that learning is quite rapid. performance over linear networks by a small amount but at the cost of a big increase in training time. We chose to present results for linear networks because they have the advantage that even though training is centralized, the policy so learned is decentralized because the features are local and therefore just the weights from the local features in the trained linear network can be used to choose actions in each cell. For large cellular arrays, training itself could be decentralized because the choice of action in a particular cell has a minor effect on far away cells. We will explore the effect of decentralized training in future work. 4 CONCLUSION The dynamic channel allocation problem is naturally formulated as an optimal control or dynamic programming problem, albeit one with very large state spaces. Traditional dynamic programming techniques are computationally infeasible for such large-scale problems. Therefore, knowledge-intensive heuristic solutions that ignore the optimal control framework have been developed. Recent approximations to dynamic programming introduced in the reinforcement learning (RL) community make it possible to go back to the channel assignment probiem and solve it as an optimal control problem, in the process finding better solutions than previously available. We presented such a solution using Sutton's (1988) TD(O) with a feature-based linear network and demonstrated its superiority on a problem with approximately 7049 states. Other recent examples of similar successes are the game 980 ~LJ.."j ".LLd.d..J -.I a) ~~~~~;.;~ ~tI!C~"I""""CIhtuT~."" "'1Ir1d~ 1 __ 10 11't-,. ~...:=.:.J - '-'~ _ ",,, ' .......... , -=-....... ~"'0011_~ fWnfore ...... tl .. nlng A_A ..... n_' ....... I' ............ e .......... Patterne b) (mmm ....... , .•.. m) 01 .......... 11) (II ................. "" .... S. Singh and D. Bertsekas FA _ BDCL _ AL Figure 3: a) Screen dump of a Java Demonstration available publicly at http:/ / www.cs.colorado.edurbaveja/Demo.html b) Sensitivity of channel assignment methods to non-uniform traffic patterns. This figure plots asymptotic empirical blocking probability for RL, BDCL, and FA for a linear array of cells with different patterns (shown at left) of mean call arrival rates chosen so that the average per cell call arrival rate is the same across patterns. The symbol I is for low, m for medium, and h for high. The numeric values of I, h, and m are chosen separately for each pattern to ensure that the average per cell rate of arrival is 120 calls/hr. The results show that RL is able to adapt its allocation strategy and thereby is better able to exploit the non-uniform call arrival rates. of backgammon (Tesauro, 1992), elevator-scheduling (Crites & Barto, 1995), and job-shop scheduling (Zhang & Dietterich, 1995). The neuro-dynamic programming textbook (Bertsekas and Tsitsiklis, 1996) presents a variety ofrelated case studies. References Barto, A.G., Bradtke, S.J. & Singh, S. (1995) Learning to act using real-time dynamic programming. Artificial Intelligence, 72:81-138. Bertsekas, D.P. (1995) Dynamic Programming and Optimal Control: Vols 1 and 2. AthenaScientific, Belmont, MA. Bertsekas, D.P. & Tsitsiklis, J. (1996) Neuro-Dynamic Programming Athena-Scientific, Belmont, MA. Crites, R.H. & Barto, A.G. (1996) Improving elevator performance using reinforcement learning. In Advances is Neural Information Processing Systems 8. Del Re, W., Fantacci, R. & Ronga, L. (1996) A dynamic channel allocation technique based on Hopfield Neural Networks. IEEE Transactions on Vehicular Technology, 45:1. McEliece, R.J. & Sivarajan, K.N. (1994), Performance limits for channelized cellular telephone systems. IEEE Trans. Inform. Theory, pp. 21-34, Jan. Sutton, R.S. (1988) Learning to predict by the methods of temporal differences. Machine Learning, 3:9-44. Tesauro, G.J. (1992) Practical issues in temporal difference learning. Machine Learning, 8{3/4):257-277. Zhang, M. & Yum, T.P. (1989) Comparisons of Channel-Assignment Strategies in Cellular Mobile Telephone Systems. IEEE Transactions on Vehicular Technology Vol. 38, No.4. Zhang, W. & Dietterich, T .G. (1996) High-performance job-shop scheduling with a timedelay TD{lambda) network. In Advances is Neural Information Processing Systems 8.
1996
36
1,232
Neural Learning in Structured Parameter Spaces Natural Riemannian Gradient Shun-ichi Amari RIKEN Frontier Research Program, RIKEN, Hirosawa 2-1, Wako-shi 351-01, Japan amari@zoo.riken.go.jp Abstract The parameter space of neural networks has a Riemannian metric structure. The natural Riemannian gradient should be used instead of the conventional gradient, since the former denotes the true steepest descent direction of a loss function in the Riemannian space. The behavior of the stochastic gradient learning algorithm is much more effective if the natural gradient is used. The present paper studies the information-geometrical structure of perceptrons and other networks, and prove that the on-line learning method based on the natural gradient is asymptotically as efficient as the optimal batch algorithm. Adaptive modification of the learning constant is proposed and analyzed in terms of the Riemannian measure and is shown to be efficient. The natural gradient is finally applied to blind separation of mixtured independent signal sources. 1 Introd uction Neural learning takes place in the parameter space of modifiable synaptic weights of a neural network. The role of each parameter is different in the neural network so that the parameter space is structured in this sense. The Riemannian structure which represents a local distance measure is introduced in the parameter space by information geometry (Amari, 1985). On-line learning is mostly based on the stochastic gradient descent method, where the current weight vector is modified in the gradient direction of a loss function. However, the ordinary gradient does not represent the steepest direction of a loss function in the Riemannian space. A geometrical modification is necessary, and it is called the natural Riemannian gradient. The present paper studies the remarkable effects of using the natural Riemannian gradient in neural learning. 128 S. Amari We first studies the asymptotic behavior of on-line learning (Opper, NIPS'95 Workshop). Batch learning uses all the examples at any time to obtain the optimal weight vector, whereas on-line learning uses an example once when it is observed. Hence, in general, the target weight vector is estimated more accurately in the case of batch learning. However, we prove that, when the Riemannian gradient is used, on-line learning is asymptotically as efficient as optimal batch learning. On-line learning is useful when the target vector fluctuates slowly (Amari, 1967). In this case, we need to modify a learning constant TJt depending on how far the current weight vector is located from the target function. We show an algorithm adaptive changes in the learning constant based on the Riemannian criterion and prove that it gives asymptotically optimal behavior. This is a generalization of the idea of Sompolinsky et al. [1995]. We then answer the question what is the Riemannian structure to be introduced in the parameter space of synaptic weights. We answer this problem from the point of view of information geometry (Amari [1985, 1995], Amari et al [1992]). The explicit form of the Riemannian metric and its inverse matrix are given in the case of simple perceptrons. We finally show how the Riemannian gradient is applied to blind separation of mixtured independent signal sources. Here, the mixing matrix is unknown so that the parameter space is the space of matrices. The Riemannian structure is introduced in it. The natural Riemannian gradient is computationary much simpler and more effective than the conventional gradient. 2 Stochastic Gradient Descent and On-Line Learning Let us consider a neural network which is specified by a vector parameter w (Wl' ... wn) E Rn. The parameter w is composed of modifiable connection weights and thresholds. Let us denote by I(~, w) a loss when input signal ~ is processed by a network having parameter w . In the case of multilayer perceptrons, a desired output or teacher signal y is associated with ~, and a typical loss is given 1 2 I(~ , y,w) = 211 y I(~ , w) II , (1) where z = I(~, w) is the output from the network. When input ~, or input-output training pair (~ , y) , is generated from a fixed probability distribution, the expected loss L( w) of the network specified by w is L(w) = E[/(~,y;w)], (2) where E denotes the expectation. A neural network is trained by using training examples (~l ' Yl)'(~2 ' Y2) ''' ' to obtain the optimal network parameter w· that minimizes L(w). If L(w) is known, the gradient method'is described by t = 1,2, ,, , where TJt is a learning constant depending on t and "ilL = oL/ow. Usually L(w) is unknown. The stochastic gradient learning method (3) was proposed by an old paper (Amari [1967]). This method has become popular since Rumelhart et al. [1986] rediscovered it. It is expected that, when TJt converges to 0 in a certain manner, the above Wt converges to w". The dynamical behavior of Neural Learning and Natural Gradient Descent 129 (3) was studied by Amari [1967], Heskes and Kappen [1992] and many others when "It is a constant. It was also shown in Amari [1967] that (4) works well for any positive-definite matrix, in particular for the metric G. Geometrically speaking Ol/aw is a covariant vector while Llwt = Wt+l - Wt is a contravariant vector. Therefore, it is natural to use a (contravariant) metric tensor C- 1 to convert the covariant gradient into the contravariant form 1 01 (" . . a ) VI = G- = L.; g'J -(w) , ow . aWj J (5) where C- 1 = (gij) is the inverse matrix of C = (gij). The present paper studies how the matrix tensor matrix C should be defined in neural learning and how effective is the new gradient learning rule (6) 3 Gradient in Riemannian spaces Let S = {w} be the parameter space and let I(w) be a function defined on S. When S is a Euclidean space and w is an orthonormal coordinate system, the squared length of a small incremental vector dw connecting wand w + dw is given by n Idwl 2 = L(dwd2 . (7) i=l However, when the coordinate system is non-orthonormal or the space S is Riemannian, the squared length is given by a quadratic form Idwl 2 = Lgjj(w)dwidwj = w'Gw. i ,j (8) Here, the matrix G = (gij) depends in general on wand is called the metric tensor. It reduces to the identity matrix I in the Euclidean orthonormal case. It will be shown soon that the parameter space S of neural networks has Riemannian structure (see Amari et al. [1992], Amari [1995], etc.). The steepest descent direction of a function I( w) at w is defined by a vector dw that minimize I(w+dw) under the constraint Idwl 2 = [2 (see eq.8) for a sufficiently small constant [. Lemma 1. The steepest descent direction of I (w) in a Riemannian space is given by -V/(w) = -C-1(w)VI(w). We call V/(w) = C-1(w)V/(w) the natural gradient of I( w) in the Riemannian space. It shows the steepest descent direction of I, and is nothing but the contravariant form of VI in the tensor notation. When the space is Eu~lidean and the coordinate system is orthonormal, C is the unit matrix I so that VI = V/' 130 S. Amari 4 Natural gradient gives efficient on-line learning Let us begin with the simplest case of noisy multilayer analog perceptrons. Given input ~, the network emits output z = f(~, 10) + n, where f is a differentiable deterministic function of the multilayer perceptron with parameter 10 and n is a noise subject to the normal distribution N(O,1) . The probability density of an input-output pair (~, z) is given by p(~,z;1O) = q(~)p(zl~;1O), where q( ~) is the probability distribution of input ~, and p(zl~; 10) = vk exp {-~ II z - f(~, 10) 112} . The squared error loss function (1) can be written as l(~, z, 10) = -logp(~, z; 10) + logq(~) -log.;z;:. Hence, minimizing the loss is equivalent to maximizing the likelihood function p(~, Zj 10). Let DT = {(~l' zI),· · ·, (~T, ZT)} be T independent input-output examples generated by the network having the parameter 10" . Then, maximum likelihood estimator WT minimizes the log loss 1(~,z;1O) = -logp(~,z;1O) over the training data DT , that is, it minimizes the training error (9) The maximum likelihood estimator is efficient (or Fisher-efficient), implying that it is the best consistent estimator satisfying the Cramer-Rao bound asymptotically, lim TE[(wT - 1O")(WT - 10")'] = C- 1 , T ..... oo (10) where G-l is the inverse of the Fisher information matrix G = (9ii) defined by [8 logp(~, z; 10) 8 logp(~, z; 10)] gij = E 8 8 Wi Wj (11) in the component form. Information geometry (Amari, 1985) proves that the Fisher information G is the only invariant metric to be introduced in the space S = {1o} of the parameters of probability distributions. Examples (~l' zd, (~2' Z2) . . . are given one at a time in the case of on-line learning. Let Wt be the estimated value at time t. At the next time t + 1, the estimator Wt is modified to give a new estimator Wt+l based on the observation (~t+ll zt+d. The old observations (~l' zd, .. . , (~t, zt) cannot be reused to obtain Wt+l, so that the learning rule is written as Wt+l = m( ~t+l, Zt+l , wt}. The process {wd is hence Markovian. Whatever a learning rule m we choose, the behavior of the estimator Wt is never better than that of the optimal batch estimator Wt because of this restriction. The conventional on-line learning rule is given by the following gradient form Wt+l = Wt -1JtV'I(~t+l,zt+l;Wt). When 1Jt satisfies a certain condition, say 1Jt = cit, the stochastic approximation guarantees that Wt is a consistent estimator converging to 10" . However, it is not efficient in general. There arises a question if there exists an on-line learning rule that gives an efficient estimator. If it exists, the asymptotic behavior of on-line learning is equivalent to Neural Learning and Natural Gradient Descent 131 that of the batch estimation method. The present paper answers the question by giving an efficient on-line learning rule _ _ 1"/( _ ) Wt+l = Wt - - v ~t+l, Zt+l; Wt . t (12) Theorem 1. The natural gradient on-line learning rule gives an Fisher-efficient estimator, that is, (13) 5 Adaptive modification of learning constant We have proved that TJt = lit with the coefficient matrix C- 1 is the asymptotically best choice for on-line learning. However, when the target parameter w· is not fixed but fluctuating or changes suddenly, this choice is not good, because the learning system cannot follow the change if TJt is too small. It was proposed in Amari [1967] to choose TJt adaptively such that TJt becomes larger when the current target w· is far from Wt and becomes small when it is close to Wt adaptively. However, no definite scheme was analyzed there. Sompolinsky et at. [1995] proposed an excellent scheme of an adaptive choice of TJt for a deterministic dichotomy neural networks. We extend their idea to be applicable to stochastic cases, where the Riemannian structure plays a role. We assume that I( ~, z; w) is differentiable with respect to w. (The nondifferentiable case is usually more difficult to analyze. Sompolinsky et al [1995] treated this case.) We moreover treat the realizable teacher so that L(w·) = o. We propose the following learning scheme: Wt+l = Wt TJt't7/(~t+1,Zt+I;Wt) TJt+l = TJt + aTJt [B/( ~t+l, Zt+l; wd - TJt], (14) (15) where a and (3 are constants. We try to analyze the dynamical behavior of learning by using the continuous version of the algorithm, ! Wt = -TJt 't7/(~t, Zt; Wt), (16) !TJt = aTJt[B/(~t,zt;wd - TJt]. (17) In order to show the dynamical behavior of (Wt, TJt), we use the averaged version of the above equation with respect to the current input-output pair (~t, Zt). We introduce the squared error variable et = ~(Wt - w·)'C·(Wt - w·). (18) By using the average and continuous time version tVt = -TJtC- 1(Wt) (a~ I(~t, Zt; Wt») , * = aTJd{3(I(~t, Zt;Wt») - TJd, where· denotes dldt and ( ) the average over the current (~, z), we have et = - 2TJtet , * = a{3TJt et - aTJi· (19) (20) 132 S.Amari The behavior of the above equation is interesting: The origin (0,0) is its attractor. However, the basin of attraction has a fractal boundary. Anyway, starting from an adequate initial value, it has the solution of the form et ~ ~ (~ - ±) ~, (21) 1 1Jt ~ 2t' (22) This proves the lit convergence rate of the generalization error, that is optimal in order for any estimator tVt converging to w· . 6 Riemannian structures of simple perceptrons We first study the parameter space S of simple perceptrons to obtain an explicit form of the metric G and its inverse G-1. This suggests how to calculate the metric in the parameter space of multilayer perceptrons. Let us consider a simple perceptron with input ~ and output z. Let w be its connection weight vector. For the analog stochastic perceptron, its input-output behavior is described by z = f( Wi ~) + n, where n denotes a random noise subject to the normal distribution N(O, (72) and f is the hyperbolic tangent, 1- e- u f(u)=l+e- u In order to calculate the metric G explicitly, let ew be the unit column vector in the direction of w in the Euclidean space R n , w ew=~, where II w II is the Euclidean norm. We then have the following theorem. Theorem 2. The Fisher metric G and its in verse G-l are given by G(w) = Cl(W)! + {C2(W) C1(w)}ewe~, -1 1 (1 1) I G (w) = -( -)! + -( -) - -( -) ewew' Cl W C2 W C1 W where W = Iwl (Euclidean norm) and C1(W) and C2(W) are given by C1(W) 4~(72 J {f2(wc) _1}2 exp { _~c2} dc, C2(W) = 4~(72 J {f2(wc) - 1}2c2 exp {_~c2} de. Theorem 3. The Jeffrey prior is given by 1 vIG(w)1 = Vn VC2(W){C1(W)}n-1. 7 The natural gradient for blind separation of mixtured signals (23) (24) (25) (26) (27) Let s = (51, ... , 5 n ) be n source signals which are n independent random variables. We assume that their n mixtures ~ = (Xl,' ", X n ), ~ = As (28) Neural Learning and Natural Gradient Descent 133 are observed. Here, A is a matrix. When 8 is time serieses, we observe ~(1), . .. , ~(t) . The problem of blind separation is to estimate W = A-I adaptively from z(t), t = 1,2,3,· ·· without knowing 8(t) nor A. We can then recover original 8 by y=W~ (29) when W = A-I. Let W E Gl(n), that is a nonsingular n x n-matrix, and ¢>(W) be a scalar function. This is given by a measure of independence such as ¢>(W) = f{ L[P(y);p(y»)' which is represented by the expectation of a loss function. We define the natural gradient of ¢>(W). Now we return to our manifold Gl(n) of matrices. It has the Lie group structure : Any A E Gl(n) maps Gl(n) to Gl(n) by W -+ W A. We impose that the Riemannian structure should be invariant by this operation A. We can then prove that the natural gradient in this case is fl¢> = \7¢>W'W. (30) The natural gradient works surprisingly well for adaptive blind signal separation Amari et al. (1995), Cardoso and Laheld [1996]. References [1] S. Amari. Theory of adaptive pattern classifiers, IEEE Trans., EC-16, No.3, 299-307, 1967. [2] S. Amari. Differential-Geometrical Methods in Statistics, Lecture Notes in Statistics, vol.28, Springer, 1985. [3] S. Amari. Information geometry of the EM and em algorithms for neural networks, Neural Networks, 8, No.9, 1379-1408, 1995. [4] S. Amari, A. Cichocki and H.H. Yang. A new learning algorithm for blind signal separation, in NIPS'95, vol.8, 1996, MIT Press, Cambridge, Mass. [5] S. Amari, K. Kurata, H. Nagaoka. Information geometry of Boltzmann machines, IEEE Trans. on Neural Networks, 3, 260-271, 1992. [6] J . F. Cardoso and Beate Laheld. Equivariant adaptive source separation, to appear IEEE Trans. on Signal Processing, 1996. [7] T. M. Heskes and B. Kappen. Learning processes in neural networks, Physical Review A , 440, 2718-2726, 1991. [8] D. Rumelhart, G.E. Hinton and R. J . Williams. Learning internal representation, in Parallel Distributed Processing: Explorations in the Microstructure of Cognition, 1, Foundations, MIT Press, Cambridge, MA, 1986. [9] H. Sompolinsky,N. Barkai and H. S. Seung. On-line learning of dichotomies: algorithms and learning curves, Neural Networks: The statistical Mechanics Perspective, Proceedings of the CTP-PBSRI Joint Workshop on Theoretical Physics, J .-H. Oh et al eds, 105- 130, 1995.
1996
37
1,233
Extraction of temporal features in the electrosensory system of weakly electric fish Fabrizio GabbianiDivision of Biology 139-74 Caltech Pasadena, CA 91125 RalfWessel Department of Biology Univ. of Cal. San Diego La J oBa, CA 92093-0357 Walter Metzner Department of Biology Univ. of Cal. Riverside Riverside, CA 92521-0427 Christof Koch Division of Biology 139-74 Caltech Pasadena, CA 91125 Abstract The encoding of random time-varying stimuli in single spike trains of electrosensory neurons in the weakly electric fish Eigenmannia was investigated using methods of statistical signal processing. At the first stage of the electrosensory system, spike trains were found to encode faithfully the detailed time-course of random stimuli, while at the second stage neurons responded specifically to features in the temporal waveform of the stimulus. Therefore stimulus information is processed at the second stage of the electrosensory system by extracting temporal features from the faithfully preserved image of the environment sampled at the first stage. 1 INTRODUCTION The weakly electric fish, Eigenmannia, generates a quasi sinusoidal, dipole-like electric field at individually fixed frequencies (250 - 600 Hz) by discharging an electric organ located in its tail (see Bullock and Heilgenberg, 1986 for reviews). The fish sense local changes in the electric field by means of two types of tuberous electroreceptors located on the body surface. T-type electroreceptors fire phase-locked to the zero-crossing of the electric field once per cycle of the electric organ discharge -email: gabbiani@klab.caltech.edu, wmetzner@mail.ucr.edu, rwessel@jeeves.ucsd.edu, koch@klab.caltech.edu. Extraction of Temporal Features in Weakly Electric Fish 63 (EOD) and are thus able to encode phase changes. P-type electroreceptors fire in a more loosely phase-locked manner with a probability smaller than 1 per EOD. Their probability of firing increases with the mean amplitude of the field thereby allowing them to encode amplitude changes (Zakon, 1986). This information is used by the fish in order to locate objects (electrolocation, Bastian 1986) as well as for communication with conspecifics (Hopkins, 1988). One behavior which has been most thoroughly studied (Heiligenberg, 1991), the jamming avoidance response, occurs when two fish with similar EOD frequency (less than 15 Hz difference) approach close enough to sense each other's field. In order to minimize beat patterns resulting from their summed electric fields, the fish with the higher (resp. lower) EOD raises further (resp. lowers) its own EOD frequency. The resulting frequency difference increase reduces the dis torsions in the interfering EODs. The fish is known to correlate phase differences computed across body regions with local amplitude increases or decreases in order to determine whether it should raise or lower its own EOD. At the level of the first central nervous nucleus of the electrosensory pathway, the electrosensory lateral line lobe of the hindbrain (ELL), phase and amplitude information are processed nearly independently of each other (Maler et al., 1981). Amplitude information is encoded in the spike trains of ELL pyramidal cells that receive input from P-receptor afferents and transmit their information to higher order levels of the electrosensory system. Two functional classes of pyramidal cells are distinguished: E-type pyramidal cells respond by raising their firing frequency to increases in the amplitude of an externally applied electric field while I-type pyramidal cells increase their firing frequency when the amplitude decreases (Bastian, 1981). The aim of this study was to characterize the temporal information processing performed by ELL pyramidal cells on random electric field amplitude modulations and to relate it to the information carried by P-receptor afferents. To this end we recorded the responses of P-receptor afferents and pyramidal cells to random electric field amplitude modulations and analyzed them using two different methods: a signal estimation method characterizing to what extent the neuronal response encodes the detailed . time-course of the stimulus and a signal-detection method developed to identify features encoded in spike trains. These two methods as well as the electrophysiology are explained in the next section followed by the result section and a short discussion. 2 METHODS 2.1 ELECTROPHYSIOLOGY Adult specimens of Eigenmannia were immobilized by intramuscular injection of a curare-like drug (Flaxedil). This also strongly attenuated the fish's EODs. The fish were held in place by a foam-lined clamp in an experimental tank and an EOD substitute electric field was established by two electrodes placed in the mouth and near the tail of the fish. The carrier frequency of the electric field, fcarrier, was chosen equal to the EOD frequency prior to curarization and the voltage generating the stimulus was modulated according to Vet) = Ao(1 + set)) cos (27rfcarrier) , where Ao is the mean amplitude and set) is a random, zero-mean modulation having a flat (white) spectrum up to a variable cut-off frequency fc and a variable standard deviation u. The modulation set) was generated by playing a blank tape on a tape 64 A } F. Gabbiani, W. Metzner, R. Wessel and C. Koch B -200m. Figure 1: A. Schematic drawing of the experimental set-up. Tape recorder (T), variable cutoff frequency Bessel filter (BF) and function generator (FG). B. Sample amplitude modulation set) and intracellular recording from a pyramidal cell (I-type, Ie = 12 Hz). Spikes occurring in bursts are marked with an asterisk (see sect. 2.3.2). The intracellular voltage trace reveals a high frequency noise caused by the EOD substitute signal and a high electrode resistance. recorder, passing the signal through a variable cut-off frequency low-pass filter before mUltiplying it by the frequency carrier signal in a function generator (fig. 1A). Extracellular recordings from P-receptor afferents were made by exposing the anterior lateral line nerve. Intracellular recordings from ELL pyramidal cells were obtained by positioning electrodes in the central region of the pyramidal cell layer. Intracellular recoroing electrodes were filled with neurotracer (neurobiotin) to be used for subsequent intracellular labeling if the recordings were stable long enough. This allowed to verify the cell type and its location within the ELL. In case no intracellular labeling could be made the recording site was verified by setting electrolytic lesions at the conclusion of the experiment. In the subsequent data analysis, data from E- and I-type pyramidal cells and from two different maps (centromedial and lateral, Carr et al., 1982) were pooled. For further experimental details, see Wessel et al. (1996), Metzner and Heiligenberg (1991), Metzner (1993). 2.2 SIGNAL ESTIMATION The ability of single spike trains to carry detailed time-varying stimulus information was assessed by estimating the stimulus from the spike train. The spike trains, x(t) = E 6(t - ti), where ti are the spike occurrence times, were convolved with a filter h (Wiener-Kolmogorov filtering; Poor, 1994; Bialek et al. 1991), Se3t(t) = J dtl h(tl)X(t - t 1) chosen in order to minimize the mean square error between the true stimulus and the estimated stimulus, f2 = (s(t) - Se3t(t»2}. The optimal filter h(t) is determined in Fourier space as the ratio of the Fourier transform of the cross-correlation between stimulus and spike train, R X3 ( r) = (x(t)s(t + r)} and the Fourier transform of the autocorrelation (power spectrum) of the spike train, Rxx(r) = (x(t)x(t + r)}. The accuracy at which single spike trains transmit information about the stimulus was characterized in the time domain by the coding fraction, defined as 'Y = 1 - flu, were u is the standard deviation of the stimulus. The coding fraction is normalized between 1 when the stimulus is perfectly estimated by the spike train «( = 0) and 0, when the estimation performance of the spike train is at chance level (f = u). Thus, the coding fraction can be Extraction a/Temporal Features in Weakly Electric Fish 65 compared across experiments. For further details and references on this stimulus estimation method in the context of neuronal sensory processing, see Gabbiani and Koch (1996) and Gabbiani (1996). 2.3 FEATURE EXTRACTION 2.3.1 General procedure The ability of single spikes to encode the presence of a temporal feature in the stimulus waveform was assessed by adapting a Fisher linear discriminant classification scheme to our data (Anderson, 1984; sect. 6.5). Each random stimulus wave-form and spike response of pyramidal cells (resp. P-receptor afferents) were binned. The bin size ~ was varied between ~min = 0.5 ms, corresponding to the sampling ratio and ~max, corresponding to the longest interval leading to a maximum of one spike per bin. The sampling interval yielding the best performance (see below) was finally retained. Typical bin sizes were ~ = 7 ms for pyramidal cells (typical mean firing rate: 30 Hz) and ~ = 1 ms for P-receptor afferents (typical firing rate: 200 Hz). The mean stimulus preceding a spike containing bin (m1) or no-spike containing bin (mO) as well as the covariances eEl, 'Eo) of these distributions were computed (Anderson, 1984; sect. 3.2). Mean vectors (resp. covariance matrices) had at most 100 (resp. 100 x 100) components. The optimal linear feature vector f predicting the occurrence or non-occurrence of a spike was found by maximizing the signal-to-noise ratio (see fig. 2A and Poor, 1994; sect. IIIB) SNR(f) = [f · ~ml - mo)]2 . f· 2('EO + 'El)f (1) The vector f is solution of (m1 - mO) = ('EO + 'E1)f. This equation was solved by diagonalizing 'EO + 'E 1 and retaining the first n largest eigenvalues accounting for 99% of the variance (Jolliffe, 1986; sect. 6.1 and 8.1). The optimal feature vector f thus obtained corresponded to up- or downstrokes in the stimulus amplitude modulation for E- and I-type pyramidal cells respectively, as expected from their mean response properties to changes in the electric field amplitude (see sect. 1). Similarly, optimal feature vectors for P-receptor afferents corresponded to upstrokes in the electric field amplitude (see sect. 1). Once f was determined, we projected the stimuli preceding a spike or no spike onto the optimal feature vector (fig. 2A) and computed the probability of correct classification between the two distributions so obtained by the resubstitution method (Raudys and Jain, 1991). The probability of correct classification (Pee) is obtained by optimizing the value of the threshold used to separate the two distributions in order to maximize (2) where the probabilities of false alarm (PFA) and correct detection (PeD ) depend on the threshold. 2.3.2 Distinction between isolated spikes and burst-like spike patterns A large fraction (56% ± 21%, n = 30) of spikes generated by pyramidal cells in response to random electric field amplitude modulations occurred in bursts (mean burst length: 18 ± 9 ms, mean number of spikes per burst: 2.9 ± 1.3, n = 30, fig. 1B). In order to verify whether spikes occurring in bursts corresponded to a more reliable encoding of the feature vector, we separated the distribution of stimuli occurring before a spike in two distributions, conditioned on whether the 66 F. Gabbiani, W. Metzner, R. Wessel and C. Koch A B ·2000 ·1000 0 1000 2000 projection onto the feature vector Figure 2: A. 2-dimensional example of two random distributions (circles and squares) as well as the optimal discrimination direction determined by maximizing the signal-to-noise ratio of eq. 1. The I-dimensional projection of the two distributions onto the optimal direction is also shown (compare with B). B. Example of the distribution of stimuli projected onto the optimal feature vector (same cell as in fig. 1B). Stimuli preceding a bin containing no spike (null), an isolated spike (isolated) and a spike belonging to a burst (burst). Horizontal scale is arbitrary (see eq. 1). spike belonged to a burst or not. The stimuli were then projected onto the feature vector (fig. 2B), as described in 2.3.1, and the probability of correct classification between the distribution of stimuli occurring before no spike and isolated spike bins was compared to the probability of correct classification between the distribution of stimuli occurring before no spike and burst spike bins (see sect. 3). 3 RESULTS The results are summarized in fig. 3. Data were analyzed from 30 pyramidal cells (E- and I-type) and 20 P-receptor afferents for a range of stimulus parameters (Ie = 2 - 40 Hz, (J' = 0.1 - 0.4, Ao was varied in order to obtain ±20 dB changes around the physiological value of the mean electric field amplitude which is of the order of 1 m V / cm). Fig. 3A reports the best probability of correct classification (eq. 2) obtained for each pyramidal cell (white squares) / P-receptor afferent (black dots) as a function of the coding fraction observed in the same experiment (note that for pyramidal cells only burst spikes are shown, see sect. 2.3.2 and fig. 3B). The horizontal axis shows that while the coding fraction of P-receptors afferents can be very high (up to 75% of the detailed stimulus time-course is encoded in a single spike train), pyramidal cells only poorly transmit information on the detailed time-course of the stimulus (less than 20% in most cases). In contrast, the vertical axis shows that pyramidal cells outperform P-receptor afferent in the classification task: it is possible to classify with up to 85% accuracy whether a given stimulus will cause a short burst of spikes or not by comparing it to a single feature vector f. This indicates that the presence of an up- or downstroke (the feature vector) is reliably encoded by pyramidal cells. Fig. 3B shows for each experiment on the ordinate the discrimination performance (eq. 2) for stimuli preceding isolated spikes against stimuli preceding no spike. The abscissa plots the discrimination performance (eq. 2) for stimuli preceding spikes occurring in bursts (white squares, fig. 3A) or stimuli preceding all spikes (black squares) against stimuli preceding Extraction o/Temporal Features in Weakly Electric Fish 67 A B c c 1.0 0 D pyramidal cells .2 D burst spikes ;:; 1.0 Preceptors 'tii . all spikes tJ . u !E !E 0.9 CD = 0.9 = D D '" .... l1li l1li U U 0.8 D D DD ill· 0.8 II DtiJ D ~ \ . . . • U D D ~ D .. D . '~ 0.7 .D lill" 0 0.7 D.p, . 0 D J' , ... u o '\, . u o ... 0 . . '0 0.6 ~ ~~ . " 0 o D .. ' . 0.6 Do . .. ~ ~ o . .. :a 0.5 :a 0.5 .! l1li 0.5 0.6 0.7 0.8 0.9 1.0 0.0 0.2 0.4 0.6 0.8 .a 2 2 probability of correct classification 0. coding fraction 0. Figure 3: A. Coding fraction and probability of correct classification for pyramidal cells (white squares, burst spikes only) and P-receptor afferents (black circles). B. Probability of correct classification against stimuli preceding no spikes for stimuli preceding burst spikes or all spikes vs. stimuli preceding isolated spikes. Dashed line: identical performance. no spike. The distribution of stimuli occurring before burst spikes (all spikes) is more easily distinguished from the distribution of stimuli occurring before no spike than the distribution of stimuli preceding isolated spikes. This clearly indicates that spikes occurring in bursts carry more reliable information than isolated spikes. 4 DISCUSSION We have analyzed the response of P-receptor afferents and pyramidal cells to random electric field amplitude modulations using methods of statistical signal processing. The previously studied mean responses of P-receptor afferents and pyramidal cells to step amplitude changes or sinusoidal modulations of an externally applied electric field left several alternatives open for the encoding and processing of stimulus information in single spike trains. We find that, while P-receptor afferents encode reliably the detailed time-course of the stimulus, pyramidal cells do not. In contrast, pyramidal cells perform better than P-receptor afferents in discriminating the occurrence of up- and downstrokes in the amplitude modulation. The presence of these features is signaled most reliably to higher order stations in the electrosensory system by short bursts of spikes emitted by pyramidal cells in response to the stimulus. This code can be expected to be robust against possible subsequent noise sources, such as synaptic unreliability. The temporal pattern recognition task solved at the level of the ELL is particularly appropriate for computations which have to rely on the temporal resolution of up- and downstrokes, such as those underlying the jamming avoidance response. 68 F Gabbiani, W. Metzner, R. Wessel and C. Koch Acknowledgments We thank Jenifer Juranek for computer assistance. Support: UCR and NSF grants, Center of Neuromorphic Systems Engineering as a part of the NSF ERC Program, and California Trade and Commerce Agency, Office of Strategic Technology. References Anderson, T.W. (1984) An introduction to Multivariate Statistical Analysis. Wiley, New York. Bastian, J. (1981) Electrolocation 2. The effects of moving objects and other electrical stimuli on the activities of two categories of posterior lateral line lobe cells in apteronotus albifrons. J. Compo Physiol. A, 144: 481-494. Bialek, W., de Ruyter van Steveninck, R.R. & Warland, D. (1991) Reading a neural code. Science, 252: 1854-1857. Bullock, T .H. & Heiligengerg, W. (1986) Electroreception. Wiley, New York. Carr, C.C., Maler, L. & Sas, E. (1982). Peripheral Organization and Central Projections of the Electrosensory Nerves in Gymnotiform Fish. J. Compo Neurol., 211:139-153. Gabbiani, F. & Koch, C. (1996) Coding of Time-Varying Signals in Spike Trains of Integrate-and-Fire Neurons with Random Threshold. Neur. Comput., 8: 44-66. Gabbiani, F. (1996) Coding of time-varying signals in spike trains of linear and half-wave rectifying neurons. Network: Compo Neur. Syst., 7:61-85. Heiligenberg, W . (1991) Neural Nets in electric fish. MIT Press, Cambridge, MA. Hopkins, C.D. (1976) Neuroethology of electric communication. Ann. Rev. Neurosci., 11:497-535. Jolliffe, I.T. (1986) Principal Component Analysis. Springer-Verlag, New York. Maler, L., Sas, E.K.B. & Rogers, J . (1981) The cytology of the posterior lateral line lobe of high-frequency weakly electric fish (gymnotidae): Dendritic differentiation and synaptic specificity. J. Compo Neurol., 255: 526-537. Metzner, W. (1993) The jamming avoidance response in Eigenmannia is controlled by two separate motor pathways. J. Neurosci., 13:1862-1878. Metzner, W. & Heiligenberg, W. (1991). The coding of signals in the electric communication of the gymnotiform fish Eigenmannia: From electroreceptors to neurons in the torus semicircularis of the midbrain. J. Compo Physiol. A, 169: 135-150. Poor, H.V. (1994) An introduction to Signal Detection and Estimation. Springer Verlag, New York. Raudys, S.J. & Jain, A.K. (1991) Small sample size effects in statistical pattern recognition: Recommendations for practitioners. IEEE Trans. Patt. Anal. Mach. Intell., 13: 252-264. Wessel, R., Koch, C. & Gabbiani F. (1996) Coding of Time-Varying Electric Field Amplitude Modulations in a Wave-Type Electric Fish J. Neurophysiol. 75:22802293. Zakon, H. (1986) The electroreceptive periphery. In: Bullock, T .H. & Heiligenberg, W. (eds) , Electroreception, pp. 103-156. Wiley, New York.
1996
38
1,234
A Model of Recurrent Interactions in Primary Visual Cortex ElDanuel Todorov, Athanassios Siapas and David SOlDers Dept. of Brain and Cognitive Sciences E25-526, MIT, Cambridge, MA 02139 Email: {emo, thanos,somers }@ai.mit.edu Abstract A general feature of the cerebral cortex is its massive interconnectivity - it has been estimated anatomically [19] that cortical neurons receive upwards of 5,000 synapses, the majority of which originate from other nearby cortical neurons. Numerous experiments in primary visual cortex (VI) have revealed strongly nonlinear interactions between stimulus elements which activate classical and non-classical receptive field regions. Recurrent cortical connections likely contribute substantially to these effects. However, most theories of visual processing have either assumed a feedforward processing scheme [7], or have used recurrent interactions to account for isolated effects only [1, 16, 18]. Since nonlinear systems cannot in general be taken apart and analyzed in pieces, it is not clear what one learns by building a recurrent model that only accounts for one, or very few phenomena. Here we develop a relatively simple model of recurrent interactions in VI, that reflects major anatomical and physiological features of intracortical connectivity, and simultaneously accounts for a wide range of phenomena observed physiologically. All phenomena we address are strongly nonlinear, and cannot be explained by linear feedforward models. 1 The Model We analyze the mean firing rates observed in oriented VI cells in response to stimuli consisting of an inner circular grating and all outer annular grating. Mean responses of individual cells are modeled by single-valued "cellular" response functions, whose arguments are the mean firing rates of the cell's inputs and their maximal synaptic conductances. A Model of Recurrent Interactions in Primary Visual Cortex 119 1.1 Neuronal model Each neuron is modeled as a single voltage compartment in which the membrane potential V is given by: C dV(t) m dt gex(t)(Eex - V{t)) + ginh(t)(Einh - V(t)) + gleak(Eleak - V{t)) + gahp(t)(Eahp - V(t)) where em is the membrane capacitance, Ex is the reversal potential for current x, and gx is the conductance for that current. If the voltage exceeds a threshold V9, a spike is generated, and afterhyperpolarizing currents are activated. The conductances for excitatory and inhibitory currents are modeled as sums of a-functions, and the ahp conductance is modeled as a decaying exponential. The model consists of two distinct cell types, excitatory and inhibitory, with realistic cellular parameters [13], similar to the ones used in [17]. To compute the response functions for the two cell types, we simulated one cell of each type, receiving excitatory and inhibitory Poisson inputs. The synaptic strengths were held constant, while the rates of the excitatory and inhibitory inputs were varied independently. Although the driving forces for excitation and inhibition vary, we found that single cell responses can be accurately modeled if incoming excitation and inhibition are combined linearly, and the net input is passed through a response function that is approximately threshold-linear, with some smoothing around threshold. This is consistent with the results of intracellular experiments that show linear synaptic interactions in visual cortex[5]. Note that the cellular functions are not sigmoids, and thus saturating responses could be achieved only through network interactions. 1.2 Cortical connectivity The visual cortex shares with many other cortical areas a similar pattern of intraareal connections [12]. Excitatory cells make dense local projections, as well as long-range horizontal projections that usually contact cells with similar response properties. Inhibitory cells make only local projections, which are spread further in space than the local excitatory connections [10]. We assume that cells with similar response properties have a higher probability of connection, and that probability falls down with distance in "feature" space. For simplicity, we consider only two feature dimensions: orientation and RF center in visual space. Since we are dealing with stimuli with radial symmetry, one spatial dimension is sufficient. The extension to more dimensions, i.e. another spatial dimension, direction selectivity, ocularity, etc., is straightforward. We assume that the feature space is filled uniformly with excitatory and inhibitory cells. Rather than modeling individual cells, we model a grid of locations, and for each location we compute the mean firing rate of cells present there. The connectivity is defined by two projection kernels Kex, Kin (one for each presynaptic cell type) and weights Wee, Wei, Wie, Wii, corresponding to the number and strength of synapses made onto excitatory/inhibitory cells. The excitatory projection has sharper tuning in orientation space, and bigger spread in visual space (Figure 1). 1.3 Visual stimuli and thalamocortical input The visual stimulus is defined by five parameters: diameter d of the inner grating (the outer is assumed infinite), log contrast Cl, C2, and orientation (Jl, (J2 of each grating. The two gratings are always centered in the spatial center of the model. 120 E. Todorov, A. Siapas and D. Somers VlslU.a.1 Spa.oe Figure 1: Excitatory (solid) and Inhibitory (dashed) connectivity kernels. visual apace Figure 2: LGN input for a stimulus with high contrast, orthogonal orientations of center and surround gratings. Each cortical cell receives LGN input which is the product of log contrast, orientation tuning, and convolution of the stimulus with a spatial receptive field. The LGN input computed in this way is multiplied by LGNex , LGNin and sent to the cortical cells. Figure 2 shows an example of what the LGN input looks like. 2 Results For given input LGN input, we computed the steady-state activity in cortex iteratively (about 30 iteration were required). Since we studied the model for gradually changing stimulus parameters, it was possible to use the solution for one set of parameters as an initial guess for the next solution, which resulted in significant speedup. The results presented here are for the excitatory population, since i) it provides the output of the cortex; ii) contains four times more cells than the inhibitory popUlation, and therefore is more likely to be recorded from. All results were obtained for the same set of parameters. 2.1 Classical RF effects First we simulate responses to a central grating (ldeg diameter) for increasing logcontrast levels. It has been repeatedly observed that although LGN firing increases linearly with stimulus log-contrast, the contrast response functions observed in VI saturate, and may even supersaturate or decline[ll, 2]. The most complete and recent model of that phenomena [6, 3] assunws ~hunting inhibition, which contradicts recent intracellular observations f4, 5]. Furthermore, that model cannot explain supersaturation. Our model achieves saturation (Figure 3A) for high contrast levels, A Model of Recurrent Interactions in Primary Visual Cortex 12I A) B) 80 60 70 50 60 , , I 40 r::i"SO r::i" = , = i 40 I i 30 , CD I ~ a:: 30 I 20 20 I I I 10 10 I , 0 Log Contrast Figure 3: Classical RF effects. A) Contrast response function for excitatory (solid) and inhibitory (dashed) cells. B) Orientation tuning for different contrast levels (solid); scaled LGN input (dashed). and can easily achieve supersaturation for different parameter setting. Instead of using shunting (divisive) inhibition, we suggest that inhibitory neurons have higher contrast thresholds than excitatory neurons, i.e. the direct LGN input to inhibitory cells is weaker. Note that we only need a subpopulation of inhibitory neurons with that property; the rest can have the same response threshold as the excitatory population. Another well-known property of VI neurons is that their orientation tuning is roughly contrast-invariant[14]. The LGN input tuning is invariant, therefore this property is easily obtained in models where VI responses are almost linear, rather than saturating [1, 16]. Achieving both contrast-invariant tuning and contrast saturation for the entire population (while singe cell feedforward response functions are non-saturating) is non-trivial. The problem is that invariant tuning requires the responses at all orientations saturate simultaneously. This is the case in our model (Figure 3B) - we found that the tuning (half width at half amplitude) varied within a 5deg range as contrast increased. 2.2 Extraclassical RF effects N ext we consider stimuli which include both a center and a surround grating. In the first set of simulations we held the diameter constant (at Ideg) and varied stimulus contrast and orientation. It has been obE.erved [9, 15] that a high contrast isoorientation surround stimulus facilitates responses to a low contrast, but suppresses responses to a high contrast center stimulus. Tpis behavior is captured very well by our model (Figure 4A). The strong response to the surround stimulus alone is partially due to direct thalamic input (Le. the thalamocortical projective field is larger than the classical receptive field of a Vl cell). The response to an orthogonal surround is between the center and iso-orientation surround responses, as observed in [9]. Many neurons in VI respond optimally to a center grating with a certain diameter, but their response decreases as the diameter increases (end-stopping). End-stopping in the model is shown in Figure 4B - responses to increasing grating diameter reach a peak and then decrease. In experiments it has been observed that the 122 A) 60r-------------~----------~ 50 40 10 /.i / I " I I -' I I I I I I O~~~--~L-o-g-C~o-n~t-ra-a-t~----~ E. Todorov, A. Siapas and D. Somers 8) 60~------------------------~ Figure 4: Extraclassical RF effects. A) Contrast response functions for center (solid), center + iso-orientation surround (dashed), center + orthogonal surround (dash-dot). B) Length tuning for 4 center log contrast levels (1, .75, .5, .4), response to surround of high contrast (dashed). border between the excitatory and inhibitory regions shifts outward (rightward) as stimulus contrast levels decline [8] . Our model also achieves this effect (for other parameter settings it can shift more downward than rightward). Note also that a center grating with maximal contrast reaches its peak response for a diameter value which is 3 times smaller than the diameter for which responses to a surround grating disappear. This is interesting because both the peak response to a central grating, and the first response to a surround grating can be used to define the extent of the classical RF - in this case clearly leading to very different definitions. This effect (shown in Figure 4B) has also been recently observed in VI [15]. 3 Population Modeling and Variability So far we have analyzed the population firing rates in the model, and compared them to physiological observations. Unfortunately, in many cases the limited sample size, or the variability in a given physiological experiment does not allow an accurate estimate of what the population response might be. In such cases researchers only describe individual cells, which are not necessarily representative. How can findings reported in this way be captured in a population model? The parameters of the model could be modified in order capture the behavior of individual cells on the population level; or, further subdivisions into neuronal subpopulations may be introduced explicitly. The approach we prefer is to increase the variance of the number of connections made across individual cells, while maintaining the same mean values. We consider variations in the amount of excitatory and inhibitory synapses that a particular cell receives from other cortical neurons. Note that the presynaptic origin of these inputs is still the "average" cortical cell , and is not chosen from a subpopulation with special response properties. The two examples we show have recently been reported in [15]. If we increase the cortical input (excitation by 100%, inhibition by 30%), we obtain a completely patch-suppressed cell, i.e. it does not respond to a center + iso-orientation surround stimulus - Figure 5A. Figure 5B shows that this cell is an "orientation contrast detector", i.e. it responds well to Odeg center + 90deg surround, and to 90deg A Model of Recurrent Interactions in Primary Visual Cortex 123 center + Odeg surround. Interestingly, the cells with that property reported in [15] were all patch-suppressed. Note also that our cell has a supersaturating center response - we found that it always accompanies patch suppression in the model. A) 70r-------------~----------~ 60 50 20 10 " " " " " " " " O~-£----~--~----~------~ Log Contrast 8) 35r-------------------------~ 30 25 10 5 I I , I I I . I , , I O~--~~--~--~~~------~ Orientation Difference Figure 5: Orientation discontinuity detection for a strongly patch- suppressed cell. A) Contrast response functions for center (solid) and center + iso-orientation surround (dashed). B) Cell's response for Odeg cmter, 0 - 90deg surround (solid) and 90deg center, 90 - Odeg surround. A) 8) 60 60 70 , 70 ...60 .60 ...o:::i"50 .o:::i"SO , = .e.. i 40 , i 40 , CD .~30 r::r::30 ..,-----.20 .20 10 .10 .0 Log Contarst 0 1.1 Sur Ort S ... O Cen Figure 6: Nonlinear summation for a non patch-suppressed cell. A) Contrast response functions for center (solid) and center + iso-orientation surround (dashed). B) Cell's response for iso-orientation surround, orthogonal center, surround + center, center only. The second example is a cell receIvmg 15% of the average cortical input. Not surprisingly, its contrast response function does not saturate - Figure 6A. However, this cell exhibits an interesting nonlinear property - it respond well to a combination of an iso-orientation surround + orthogonal center, but does not respond to either stimulus alone (Figure 6B). It. is not clear from [15] whether the cells with this property had saturating contrast response functions. 124 E. Todorov, A. Siapas and D. Somers 4 Conclusion We presented a model of recurrent computation in primary visual cortex that relies on a limited set of physiologically plausible mechanisms. In particular, we used the different cellular properties of excitatory and inhibitory neurons and the nonisotropic shape of lateral connections (Figure 1). Due to space limitations, we only presented simulation results, rather than analyzing the effects of specific parameters. A preliminary version of such analysis is given in [17] for local effects, and will be developed elsewhere for lateral interaction effects. Our goal here Was to propose a framework for studying recurrent computation in VI, that is relatively simple, yet rich enough to simultaneously account for a wide range of physiological observations. Such a framework is needed if we are to analyse systematically the fundamental role of recurrent interactions in neocortex. References [1] Ben-Yishai, R., Lev Bar-Or, R. & Sompolinsky, H. Proc. Nail. Acad. Sci. U.S.A. 92, 3844-3848 (1995). [2] Bonds, A.B. Visual Neurosci. 6, 239-255 (1991). [3] Carandini, M. & Heeger, D.J. Science. 264, 1333-1336 (1994). [4] Douglas R.J., Martin K.C., Whitteridge D. An intracellular analysis of the visual responses on neurones in cat visual cortex. J Physiology 44, 659-696, 1991. [5] Ferster,D & Jagadeesh, B. J. Neurosci. 12, 1262-1274(1992). [6] Heeger, D.J. Visual Neurosci. 70, 181-197 (1992). [7] Hubel, D.H. & Wiesel, T.N. J. Neurophyszol. 148, 574-591 (1959). [8] Jagadeesh, B. & Ferster, D. Soc Neursci Abstr. 16 130.11. (1990). [9] Knierim, J.J. & Van Essen, D.C. J. Neurophysiol. 67, 961-980 (1992). [10] Kisvarday, Z.F., Martin, K.A.C., Freund, T.F., Magloczky, Z.F., Whitteridge, D., and Somogyi, D. Exp. Brain Res. 64, 541-552. [11] Li, C.Y. & Creutzfeldt, O.D. Pflugers Arch. 401, 304-314 (1984). [12] Lund J .S., Yoshioka T., Levitt, J .B. Cereb. Cortex 3, 148-162. [13] McCormick, D.A., Connors, B.W., Lighthall, J.W. & Prince, D.A. J. Neurophysiol. 54, 782-806 (1985). [14] Sclar, G. & Freeman, R.D. Exp. Brain Res. 46,457-461. [15] Sillito, A.M., Grieve, K.L., Jones, H.E., Cudeiro, J., & Davis, J. Nature, Nov 1995. [16] Somers, D.C., Nelson, S.B. & Sur, M. J. Neurosci. 15, 5448-5465 (1995). [17] Siapas A, Todorov E, Somers D. Computing the mean firing rates of ensembles of realistic neurons. Soc Neuroscience Abstract, 1995. [18] Stemmler, M., Usher, M. & Niebur, E Science 269, 1877-1880, (1995). [19] White, E.L. Cortical Circuits 46-82 (Birkhauser, Boston, 1989). PART III THEORY
1996
39
1,235
The CONDENSATION algorithm conditional density propagation and applications to visual tracking A. Blake and M. IsardDepartment of Engineering Science, University of Oxford, Oxford OXI 3PJ, UK. Abstract The power of sampling methods in Bayesian reconstruction of noisy signals is well known. The extension of sampling to temporal problems is discussed. Efficacy of sampling over time is demonstrated with visual tracking. 1 INTRODUCTION The problem of tracking curves in dense visual clutter is a challenging one. Trackers based on Kalman filters are of limited power; because they are based on Gaussian densities which are unimodal they cannot represent simultaneous alternative hypotheses. Extensions to the Kalman filter to handle multiple data associations (Bar-Shalom and Fortmann, 1988) work satisfactorily in the simple case of point targets but do not extend naturally to continuous curves. Tracking is the propagation of shape and motion estimates over time, driven by a temporal stream of observations. The noisy observations that arise in realistic problems demand a robust approach involving propagation of probability distributions over time. Modest levels of noise may be treated satisfactorily using Gaussian densities, and this is achieved effectively by Kalman filtering (Gelb, 1974). More pervasive noise distributions, as commonly arise in visual background clutter, demand a more powerful, non-Gaussian approach. One very effective approach is to use random sampling. The CONDENSATION algorithm, described here, combines random sampling with learned dynamical models to propagate an entire probability distribution for object position and shape, over time. The result is accurate tracking of agile motion in clutter, decidedly more • Web: http://www.robots.ox.ac.uk/ ... ab/ 362 A. Blake and M. lsard robust than what has previously been attainable by Kalman filtering. Despite the use of random sampling, the algorithm is efficient, running in near real-time when applied to visual tracking. 2 SAMPLING METHODS A standard problem in statistical pattern recognition is to find an object parameterised as x with prior p(x), using data z from a single image. The posterior density p(xlz) represents all the knowledge about x that is deducible from the data. It can be evaluated in principle by applying Bayes' rule (Papoulis, 1990) to obtain p(xlz) = kp(zlx)p(x) (1) where k is a normalisation constant that is independent of x. However p(zlx) may become sufficiently complex that p(xlz) cannot be evaluated simply in closed form. Such complexity arises typically in visual clutter, when the superfluity of observable features tends to suggest multiple, competing hypotheses for x. A one-dimensional illustration of the problem is illustrated in figure 1 in which multiple features give Measured features z, Z2 Z:J Z4 Z!j Ze + + + + + + • x I I I I I I I I I I I I p(z I x) I I I I I I I I I I I I a x Figure 1: One-dimensional observation model. A probabilistic observation model allowing for clutter and the possibility of missing the target altogether is specified here as a conditional density p( z I x ) . rise to a multimodal observation density function p(zlx). When direct evaluation of p(xlz) is infeasible, iterative sampling techniques can be used (Geman and Geman, 1984; Ripley and Sutherland, 1990; Grenander et al., 1991; Storvik, 1994). The factored sampling algorithm (Grenander et al., 1991). generates a random variate x from a distribution p(x) that approximates the posterior p(xlz). First a sample-set {s(1), ... , s(N)} is generated from the prior density p(x) and then a sample x = Xi, i E {I, ... , N} is chosen with probability p(zlx = s(i») 7ri = N . . Lj=l p(zlx = s(3») Sampling methods have proved remarkably effective for recovering static objects from cluttered images. For such problems x is multi-dimensional, a set of parameters for curve position and shape. In that case the sample-set {s(1), ... , s(N)} represents The CONDENSATION Algorithm 363 a distribution of x-values which can be seen as a distribution of curves in the image plane, as in figure 2. Figure 2: Sample-set representation of shape distributions for a curve with parameters x, modelling the outline (a) of the head of a dancing girl. Each sample s(n) is shown as a curve (of varying position and shape) with a thickness proportional to the weight 1r(n). The weighted mean of the sample set (b) serves as an estimator of mean shape 3 THE CONDENSATION ALGORITHM The CONDENSATION algorithm is based on factored sampling but extended to apply iteratively to successive images in a sequence. Similar sampling strategies have appeared elsewhere (Gordon et al., 1993; Kitigawa, 1996), presented as developments of Monte-Carlo methods. The methods outlined here are described in detail elsewhere. Fuller descriptions and derivation of the CONDENSATION algorithm are in (Isard and Blake, 1996; Blake and Isard, 1997) and details of the learning of dynamical models, which is crucial to the effective operation of the algorithm are in (Blake et al., 1995). Given that the estimation process at each time-step is a self-contained iteration of factored sampling, the output of an iteration will be a weighted, time-stamped sample-set, denoted s~n), n = 1, ... I N with weights 1r~n) I representing approximately the conditional state-density p(xtIZe) at time t, where Zt = (Zl, ... I Zt). How is this sample-set obtaine-d? Clearly the process must begin with a prior density and the effective prior for time-step t should be p(xtIZt-t}. This prior is of course multi-modal in general and no functional representation of it is available. It is derived from the sample set representation (S~~)ll 1r~~)1)' n = 1, ... , N of p(Xt-lIZt-l), the output from the previous time-step, to which prediction must then be applied. The iterative process applied to the sample-sets is depicted in figure 3. At the top of the diagram, the output from time-step t - 1 is the weighted sample-set {(st)l' 1rt?l) , n = I, . .. ,N}. The aim is to maintain, at successive time-steps, sample sets of fixed size N, so that the algorithm can be guaranteed to run within a given computational resource. The first operation therefore is to sample (with 364 A. Blake and M. !sard p(x11 Z,-1 ) p(x11 Z,) Figure 3: One time-step in the CONDENSATION algorithm. Blob centres represent sample values and sizes depict sample weights. replacement) N times from the set {S~~)l}' choosing a given element with probability ll't)l' Some elements, especially those with high weights, may be chosen several times, leading to identical copies of elements in the new set. Others with relatively low weights may not be chosen at all. Each element chosen from the new set is now subjected to a predictive step. {The dynamical model we generally use for prediction is a linear stochastic differential equation (s.d.e.) learned from training sets of sample object motion (Blake et al., 1995).) The predictive step includes a random component, so identical elements may now split as each undergoes its own independent random motion step. At this stage, the sample set {s~n)} for the new time-step has been generated but, as yet, without its weights; it is approximately a fair random sample from the effective prior density p(XtIZt-l) for time-step t. Finally, the observation step from factored sampling is applied, generating weights from the observation density p(Zt IXt) to obtain the sample-set representation {(s~n), ll'}n»} of state-density for time t. The algorithm is specified in detail in figure 4. The process for a single time-step consists of N iterations to generate the N elements of the new sample set. Each iteration has three steps, detailed in the figure, and we comment below on each. 1. Select nth new sample s~(n) to be some S~~l from the old sample set, sampled with replacement with probability 1l'~~1' This is achieved efficiently by using cumulative weights C~~l (constructed in step 3). 2. Predict by sampling randomly from the conditional density for the dynamical model to generate a sample for the new sample-set. 3. Measure in order to generate weights ll'~n) for the new sample. Each weight The CONDENSATION Algorithm 365 is evaluated from the observation density function which, being multimodal in general, "infuses" multi modality into the state density. Iterate From the "old" sample-set {s~~t 7ltL c~~t n ;;:: 1, ... , N} at time-step t - 1, construct a "new" sample-set {s~n),7r~n),c~n)},n;;:: 1, .. . ,N for time t. Construct the nth of N new samples as follows: 1. Select a sample s~(n) as follows: (a) generate a random number r E [0,1], uniformly distributed. (b) find, by binary subdivision, the smallest j for which C~~l ~ r (c) set s~(n) = S~~l 2. Predict by sampling from p(XtIXt-l = S/~~)l) to choose each s~n). 3. Measure and weight the new position in terms of the measured features Zt: then normalise so that Ln 7r~n) ;;:: 1 and store together with cumulative Probability as (s(n) 7r(n) c(n») where t , t , t (0) Ct (n) Ct 0, (n-l) + (n) Ct 7rt (n = 1 .. . N). Figure 4: The CONDENSATION algorithm. At any time-step, it is possible to "report" on the current state, for example by evaluating some moment of the state density as N £(f(xd] ;;:: l:= 7r~n) f (s~n») . (2) n=l 4 RESULTS A good deal of experimentation has been performed in applying the CONDENSATION algorithm to the tracking of visual motion, including moving hands and dancing figures. Perhaps one of the most stringent tests was the tracking of a leaf on a bush, in which the foreground leaf is effectively camouflaged against the background. A 12 second (600 field) sequence shows a bush blowing in the wind, the task being to track one particular leaf. A template was drawn by hand around a still of one chosen leaf and allowed to undergo affine deformations during tracking. Given that a clutter-free training sequence is not available, the motion model was learned by means of a bootstrap procedure (Blake et al., 1995). A tracker with default dynamics proved capable of tracking the first 150 fields of a training sequence before losing
1996
4
1,236
On a Modification to the Mean Field EM Algorithm in Factorial Learning A. P. Dunmur D. M. Titterington Department of Statistics Maths Building University of Glasgow Glasgow G12 8QQ, UK alan~stats.gla.ac.uk mike~stats.gla.ac.uk Abstract A modification is described to the use of mean field approximations in the E step of EM algorithms for analysing data from latent structure models, as described by Ghahramani (1995), among others. The modification involves second-order Taylor approximations to expectations computed in the E step. The potential benefits of the method are illustrated using very simple latent profile models. 1 Introduction Ghahramani (1995) advocated the use of mean field methods as a means to avoid the heavy computation involved in the E step of the EM algorithm used for estimating parameters within a certain latent structure model, and Ghahramani & Jordan (1995) used the same ideas in a more complex situation. Dunmur & Titterington (1996a) identified Ghahramani's model as a so-called latent profile model, they observed that Zhang (1992,1993) had used mean field methods for a similar purpose, and they showed, in a simulation study based on very simple examples, that the mean field version of the EM algorithm often performed very respectably. By this it is meant that, when data were generated from the model under analysis, the estimators of the underlying parameters were efficient, judging by empirical results, especially in comparison with estimators obtained by employing the 'correct' EM algorithm: the examples therefore had to be simple enough that the correct EM algorithm is numerically feasible, although any success reported for the mean field 432 A. P. Dunmur and D. M. Titterington version is, one hopes, an indication that the method will also be adequate in more complex situations in which the correct EM algorithm is not implement able because of computational complexity. In spite of the above positive remarks, there were circumstances in which there was a perceptible, if not dramatic, lack of efficiency in the simple (naive) mean field estimators, and the objective of this contribution is to propose and investigate ways of refining the method so as to improve performance without detracting from the appealing, and frequently essential, simplicity of the approach. The procedure used here is based on a second order correction to the naive mean field well known in statistical physics and sometimes called the cavity or TAP method (Mezard, Parisi & Virasoro , 1987). It has been applied recently in cluster analysis (Hofmann & Buhmann, 1996). In Section 2 we introduce the structure of our model, Section 3 explains the refined mean field approach, Section 4 provides numerical results, and Section 5 contains a statement of our conclusions. 2 The Model The model under study is a latent profile model (Henry, 1983), which is a latent structure model involving continuous observables {x r : r = 1 ... p} and discrete latent variables {Yi : i = 1 ... d}. The Yi are represented by indicator vectors such that for each i there is a single j such that Yij = 1 and Yik = 0, for all k f; j. The latent variables are connected to the observables by a set of weight matrices Wi in such a way that the distribution of the observations given the latent variables is a multivariate Gaussian with mean Ei WiYi and covariance matrix r. To ease the notation, the covariance matrix is taken to be the identity matrix, although extension is quite easy to the case where r is a diagonal matrix whose elements have to be estimated (Dunmur & Titterington, 1996a). Also to simplify the notation, the marginal distributions of the latent variables are taken to be uniform, so that the totality of unknown parameters is made up of the set of weight matrices, to be denoted by W = (WI, W2 , •• ·, Wd). 3 Methodology In order to learn about the model we have available a dataset V = {xl-' : J.L = 1 ... N} of N independent, p dimensional realizations of the model, and we adopt the Maximum Likelihood approach to the estimation of the weight matrices. As is typical of latent structure models, there is no explicit Maximum Likelihood estimate of the parameters of the model, but there is a version of the EM algorithm (Dempster, Laird & Rubin, 1977) that can be used to obtain the estimates numerically. The EM algorithm consists of a sequence of double steps, E steps and M steps. At stage m the E step, based on a current estimate wm-I of the parameters, calculates where, the expectation ( . ) is over the latent variables y, and is conditional on V and wm-l, and Cc denotes the crucial part of the complete-data log-likelihood, given A Modification to Mean Field EM 433 by The M step then maximizes Q with respect to W and gives the new parameter estimate W m . For the simple model considered here, the M step gives where W = (WI, W2 , ..• , Wd ) and yT = (Yf, Yf, ... ,yJ) and, for brevity, explicit mention of the conditioned quantities in the expectations (.) has been omitted. The above formula differs somewhat from that given by Ghahramani (1995). Hence we need to evaluate the sets of expectations (Yi) and (YiyJ) for each example in the dataset. (The superscript J.t is omitted, for clarity.) As pointed out in Ghahramani (1995), it is possible to evaluate these expectations directly by summing over all possible latent states. This has the disadvantage of becoming exponentially more expensive as the size of the latent space increases. The mean field approximation is well known in physics and can be used to reduce the computational complexity. At its simplest level, the mean field approximation replaces the joint expectations of the latent variables by the products of the individual expectations; this can be interpreted as bounding the likelihood from below (Saul, Jaakkola, Jordan, 1996). Here we consider a second order approximation, as outlined below. Since the latent variables are categorical, it is simple to sum over the state space of a single latent variable. Hence, following Parisi (1988), the expectations of the latent variables are given by (1) where /j(fi) is the lh component of the softmax function; exp(fij)/ ~k exp(fik), and the expectation (.) is taken over the remaining latent variables. The vector fi = {fij} contains the log probabilities (up to a constant) associated with each category of the latent variable for each example in the data set. For the simple model under study fij is given by f' 0 = {W!(x - '" WkYk)} - !(W!Wo) 0 0. I) t L.J k#i j 2 t 1 11 (2) The expectation in (1) can be expanded in a Taylor series about the average, (fi), giving (3) where D.fij = fij - (fij ). The naive mean field approximation simply ignores all corrections. We can postulate that the second order fluctuations are taken care of 434 A. P. Dunmur and D. M. Titterington by a so called cavity field, (see, for instance, Mezard, Parisi & Virasoro , 1987, p.16), that is, (4) where the vector of fields hi = {hik} has been introduced to take care of the correction terms. This equation may also be expanded in a Taylor series to give Then, equating coefficients with (3) and after a little algebra, we get (5) where djk is the Kronecker delta and, for the model under consideration, (tl€ijtl€ik) = (wr L, Wm (( YmY~) - (Ym) (y~») W!Wi) (6) mn#~ jk The naive mean field assumption may be used in (6), giving (7) Within the E step, for each realization in the data set, the mean fields (4), along with the cavity fields (5), can be evaluated by an iterative procedure which gives the individual expectations of the latent variables. The naive mean field approximation (7) is then used to evaluate the joint expectations (YiyJ). In the next section we report, for a simple model, the effect on parameter estimation of the use of cavity fields. 4 Results Simulations were carried out using latent variable models (i) with 5 observables and 4 binary hidden variables and (ii) with 5 observables and 3 3-state hidden variables. The weight matrices were generated from zero mean Gaussian variables with standard deviation w. In order to make the M step trivial it was assumed that the matrices were known up to the scale parameter w, and this is the parameter estimated by the algorithm. A data set was generated using the known parameter and this was then estimated using straight EM, naive mean field (MF) and mean field with cavity fields (MF cay ). Although datasets of sizes 100 and 500 were generated, only the results for N = 500 are presented here since both scenarios showed the same qualitative behaviour. Also, the estimation algorithms were started from different initial positions; this too had no effect on the final estimates of the parameters. A representative selection of results follows; Table 1 shows the results for both the 5 x 4 x 2 model and the 5 x 3 x 3 model. The results show that, when the true value, Wtr, of the parameter was small, there is little difference among the three methods. This is due to the fact that at these A Modification to Mean Field EM 435 Table 1: Table of results N=500, results averaged over 50 simulations for 5 observables with 4 binary latent variables and for 5 observables with 3 3-state latent variables. The figures in brackets give the standard deviation of the estimates, West, in units according to the final decimal place. RMS is the root mean squared error of the estimate compared to the true value. 5x4x2 5x3x3 Method Wtr Winit West RMS West RMS EM 0.1 0.05 0.09(1) 0.014 0.10(2) 0.024 MF 0.1 0.05 0.09(1) 0.014 0.10(2) 0.023 MFcav 0.1 0.05 0.09(1) 0.014 0.10(2) 0.023 EM 0.5 0.1 0.49(2) 0.016 0.50(2) 0.019 MF 0.5 0.1 0.47(2) 0.029 0.46(2) 0.038 MFcav 0.5 0.1 0.48(2) 0.026 0.47(2) 0.032 EM 1.0 0.1 0.99(2) 0.016 1.00(2) 0.018 MF 1.0 0.1 0.96(2) 0.040 0.98(2) 0.032 MFcav 1.0 0.1 0.99(2) 0.018 1.00(2) 0.018 EM 2.0 0.2 1.99(1) 0.014 1.99(1) 0.015 MF 2.0 0.2 1.98(2) 0.021 1.98(2) 0.023 MFcav 2.0 0.2 1.97(2) 0.027 1.97(2) 0.031 EM 5.0 0.1 4.99(1) 0.013 5.00(1) 0.013 MF 5.0 0.1 4.99(1) 0.016 4.96(2) 0.047 MFcav 5.0 0.1 4.97(2) 0.032 4.88(3) 0.114 small values there is little separation among the mixtures that are used to generate the data and hence the methods are all equally good (or bad) at estimating the parameter. As the true parameter increases and becomes close to one, the cavity field method performs significantly better then naive mean field; in fact it performs as well as EM for Wtr=1. For values of Wtr greater than one, the cavity field method performs less well than naive mean field. This suggests that the Taylor expansions (3) and (5) no longer provide reliable approximations. Since (8) it is easy to see that if the elements in Wi are much less than one then the corrections to the Taylor expansion are small and hence the cavity fields are small, so the approximations hold. If W is much larger than one, then the mean field estimates become closer to zero and one since the energies f.ij (equation 2) become more extreme. Hence if the mean fields correctly estimate the latent variables the corrections are indeed small, but if the mean fields incorrectly estimate the latent variable the error term is substantial, leading to a reduction in performance. Another simulation, similar to that presented in Ghahramani (1995), was also studied to compare the modification to both the 'correct' EM and the naive mean field. The model has two latent variables that correspond to either horizontal or vertical 436 A. P. Dunmur and D. M. Titterington lines in one of four positions. These are combined and zero mean Gaussian noise added to produce a 4 x 4 example image. A data set is created from many of these examples with the latent variables chosen at random. From the data set the weight matrices connecting the observables to the latent variables are estimated and compared to the true weights which consist of zeros and ones. Typical results for a sample size of 160 and Gaussian noise of variance 0.2 are presented in Figure 1. The number of iterations needed to converge were similar for all three methods. EM 0.158 MF 0.154 MFcav 0.153 liD Figure 1: Estimated weights for a sample size of N = 160 and noise variance 0.2 added. The three rows correspond to EM, naive MF and MF cay respectively. The number on the left hand end of the row is the mean squared error of the estimated weights compared with the true weights. The first four images are the estimates of the first latent vector and the remaining four images are the estimates of second latent vector. As can be seen from Figure 1 there is very little difference between the estimates of the weights and the mean squared errors are all very close. The mean field method converged in approximately four iterations which means that for this simple model the MF E step is taking approximately 32 steps as compared to 16 steps for the straight EM. This is due to the simplicity of the latent structure for this model. For a more complicated model the MF algorithm should take fewer iterations. Again the results are encouraging for the MF method, but they do not show any obvious benefit from using the cavity field correction terms. 5 Conclusion The cavity field method can be applied successfully to improve the performance of naive mean field estimates. However, care must be taken when the corrections become large and actually degrade the performance. Predicting the failure modes of the algorithm may become harder for larger (more realistic) models. The message seems to be that where the mean field does well the cavity fields will improve the situation, but where the mean field performs less well the cavity fields can degrade performance. This suggests that the cavity fields could be used as a check on the mean field method. Where the cavity fields are small we can be reasonably confident that the mean field is producing sensible answers. However where the cavity fields become large it is likely that the mean field is no longer producing accurate estimates. Further work would consider larger simulations using more realistic models. It A Modification to Mean Field EM 437 might no longer be feasible to compare these simulations with the 'correct' EM algorithm as the size of the model increases, though other techniques such as Gibbs sampling could be used instead. It would also be interesting to look at the next level of approximation where, instead of approximating the joint expectations by the product of the individual expectations in equation (6), the joint expectations are evaluated by summing over the joint state space (c.f. equation (1)) and possibly evaluating the corresponding cavity fields (Dunmur & Titterington, 1996b). This would perhaps improve the quality of the approximation without introducing the exponential complexity associated with the full Estep. Acknowledgements This research was supported by a grant from the UK Engineering and Physical Sciences Research Council. References DEMPSTER, A. P., LAIRD, N. M. & RUBIN, D. B. (1977). Maximum likelihood estimation from incomplete data via the EM algorithm (with discussion). J. R. Statist. Soc. B 39, 1-38. DUNMUR, A. P. & TITTERINGTON, D. M. (1996a). Parameter estimation in latent structure models. Tech. Report 96-2, Dept. Statist., Univ. Glasgow. DUNMUR, A. P. & TITTERINGTON, D. M. (1996b). Higher order mean field approximations. In preparation. GHAHRAMANI, Z. (1995). Factorial learning and the EM algorithm. In Advances in Neural Information Processing Systems 7, Eds. G. Tesauro, D. S. Touretzky & T. K. Leen. Cambridge MA: MIT Press. GHAHRAMANI, Z. & JORDAN, M. I. (1995). Factorial hidden Markov models. Computational Cognitive Science Technical Report 9502, MIT. HENRY, N. W. (1983). Latent structure analysis. In Encyclopedia of Statistical Sciences, Volume 4, Eds. S. Kotz, N. L. Johnson & C. B. Read, pp.497-504. New York: Wiley. HOFMANN, T. & BUHMANN, J. M. (1996) Pairwise Data Clustering by Deterministic Annealing. Tech. Rep. IAI-TR-95-7, Institut fur Informatik III, Universitiit Bonn. MEZARD, M., PARISI, G. & VIRASORO, M. A. (1987) Spin Glass Theory and Beyond. Lecture Notes in Physics, 9. Singapore: World Scientific. PARISI, G. (1988). Statistical Field Theory. Redwood City CA: Addison-Wesley. SAUL, L. K., JAAKKOLA, T. & JORDAN, M. I. (1996) Mean Field Theory for Sigmoid Belief Networks. J. Artificial Intelligence Research 4,61-76. ZHANG, J. (1992). The Mean Field Theory in EM procedures for Markov random fields. 1. E. E. E. Trans. Signal Processing 40, 2570-83. ZHANG, J. (1993). The Mean Field Theory in EM procedures for blind Markov random field image restoration. 1. E. E. E. Trans. Image Processing 2, 27-40.
1996
40
1,237
VLSI Implementation of Cortical Visual Motion Detection Using an Analog Neural Computer Ralph Etienne-Cummings Electrical Engineering, Southern Illinois University, Carbondale, IL 62901 Jan Van der Spiegel The Moore School, University of Pennsylvania, Philadelphia, PA 19104 Naomi Takahashi The Moore School, University of Pennsylvania, Philadelphia, PA 19104 Alyssa Apsel Electrical Engineering, California Inst. Technology, Pasadena, CA 91125 Paul Mueller Corticon Inc., 3624 Market Str, Philadelphia, PA 19104 Abstract Two dimensional image motion detection neural networks have been implemented using a general purpose analog neural computer. The neural circuits perform spatiotemporal feature extraction based on the cortical motion detection model of Adelson and Bergen. The neural computer provides the neurons, synapses and synaptic time-constants required to realize the model in VLSI hardware. Results show that visual motion estimation can be implemented with simple sum-andthreshold neural hardware with temporal computational capabilities. The neural circuits compute general 20 visual motion in real-time. 1 INTRODUCTION Visual motion estimation is an area where spatiotemporal computation is of fundamental importance. Each distinct motion vector traces a unique locus in the space-time domain. Hence, the problem of visual motion estimation reduces to a feature extraction task, with each feature extractor tuned to a particular motion vector. Since neural networks are particularly efficient feature extractors, they can be used to implement these visual motion estimators. Such neural circuits have been recorded in area MT of macaque monkeys, where cells are sensitive and selective to 20 velocity (Maunsell and Van Essen, 1983). In this paper, a hardware implementation of 20 visual motion estimation with spatiotemporal feature extractors is presented. A silicon retina with parallel, continuous time edge detection capabilities is the front-end of the system. Motion detection neural networks are implemented on a general purpose analog neural computer which is composed of programmable analog neurons, synapses, axon/dendrites and synaptic time686 R. Etienne-Cummings, 1. van der Spiege~ N. Takahashi, A. Apsel and P. Mueller constants (Van der Spiegel et al., 1994). The additional computational freedom introduced by the synaptic time-constants, which are unique to this neural computer, is required to realize the spatiotemporal motion estimators. The motion detection neural circuits are based on the early ID model of Adelson and Bergen and recent 2D models of David Heeger (Adelson and Bergen, 1985; Heeger et at., 1996). However, since the neurons only computed delayed weighted sum-and-threshold functions, the models must be modified. The original models require division for intensity normalization and a quadratic nonlinearity to extract spatiotemporal energy. In our model, normalization is performed by the silicon retina with a large contrast sensitivity (all edges are normalized to the same output), and rectification replaces the quadratic non-linearity. Despite these modifications, we show that the model works correctly. The visual motion vector is implicitly coded as a distribution of neural activity. Due to its computational complexity, this method of image motion estimation has not been attempted in discrete or VLSI hardware. The general purpose analog neural computer offers a unique avenue for implementing and investigating this method of visual motion estimation. The analysis, implementation and performance of spatiotemporal visual motion estimators are discussed. 2 SPATIOTEMPORAL FEATURE EXTRACTION The technique of estimating motion with spatiotemporal feature extraction was proposed by Adelson and Bergen in 1985 (Adelson and Bergen, 1985). It emerged out of the observation that a point moving with constant velocity traces a line in the space-time domain, shown in figure la. The slope of the line is proportional to the velocity of the point. Hence, the velocity is represented as the orientation of the line. Spatiotemporal orientation detection units, similar to those proposed by Hubel and Wiesel for spatial orientation detection, can be used for detecting motion (Hubel and Wiesel, 1962). In the frequency domain, the motion of the point is also a line where the slope of the line is the velocity of the point. Hence orientation detection filters, shown as circles in figure lb, can be used to measure the motion of the point relative to their tuned velocity. A population of these tuned filters, figure ic, can be used to measure general image motion. m e x -direction :> -Velocity +Velocity (a) (b) (c) Figure 1: (a) ID Motion as Orientation in the Space-Time Domain. (b) and (c) Motion detection with Oriented Spatiotemporal Filters. If the point exhibits 2D motion, the problem is substantially more complicated, as observed by David Heeger (1987). A point executing 2D motion spans a plane in the frequency domain. The spatiotemporal orientation filter tuned to this motion must also span a plane (Heeger et ai., 1987, 1996). Figure 2a shows a filter tuned to 2D motion. Unfortunately, this torus shaped filter is difficult to realize without special mathematical tools. Furthermore, to create a general set of filters for measuring general 2D motion, the filters must cover all the spatiotemporal frequencies and all the possible velocities of the stimuli. The latter requirement is particularly difficult to obtain since there are two degrees of freedom (vX' v.y) to cover. VLSI Implementation of Cortical Vzsual Motion Detection I Ii(a) (b) Figure 2: (a) 20 Motion Detection with 20 Oriented Spatiotemporal Filters. (b) General 20 Motion Detection with 2 Sets of 10 Filters. 687 To circumvent these problems, our model decomposes the image into two orthogonal images, where the perpendicular spatial variation within the receptive field of the filters are eliminated using spatial smoothing. Subsequently, ID spatiotemporal motion detection is used on each image to measure the velocity of the stimuli. This technique places the motion detection filters, shown as the circles in figure 2b, only in the rox-rot and roy-rot planes to extract 20 motion, thereby drastically reducing the complexity of the 20 motion detection model from O(n2) to O(2n). 2.1 CONSTRUCTING THE SPATIOTEMPORAL MOTION FILTERS The filter tuned to a velocity vox (vOy) is centered at ro"x (rooy) and root where vox = roo/roox (voy = roo/roOy)' To create the filters, quadrature pairs (i.e. odd and even pairs) of spatial and temporal band-pass filters centered at the appropriate spatiotemporal frequencies are summed and differenced (Adelson and Bergen, 1985). The 1tI2 phase relationship between the filters allows them to be combined such that they cancel in opposite quadrants, leaving the desired oriented filter, as shown in figure 3a. Equation 1 shows examples of quadrature pairs of spatial and temporal filters implemented. The coefficients of the filters balance the area under their positive and negative lobes. The spatial filters in equation 1 have a 5 x 5 receptive field, where the sampling interval is determined by the silicon retina. Figure 3b shows a contour plot of an oriented filter (a=11 rads/s, 02=201=4Da). S(even) = [0.5 - 0.32Cos(wx )- 0.18Cos(2wx )] (a) S(odd) = [-{).66jSin(wx ) - 0.32jSin(2wx )] (b) -w26 T(even) = t 2 . a« 6 "" 6 (jwt + a)(jwt + 61 )(jwt + 62) , 1 2 . 66 T(odd) = JWt 1 2 . a« 6 "" 6 (jwt +a)(jwt +61)(jwt +62) , 1 2 Left Motion = S(e)T(e) - S(o)T(o) or S(e)T(o) - S(o)T(e) Right Motion = S(e)T(e)+S(o)T(o) or S(e)T(o)+S(o)T(e) (c) (d) (e) (f) (I) To cover a wide range of velocity and stimuli, multiple filters are constructed with various velocity, spatial and temporal frequency selectivity. Nine filters are chosen per dimension to mosaic the rox-rot and roy-rot planes as in figure 2b. The velocity of a stimulus is given by the weighted average of the tuned velocity of the filters, where the weights are the magnitudes of each filter's response. All computations for 20 motion detection based on cortical models have been realized in hardware using a large scale general purpose analog neural computer. 688 R. Etienne-Cummings, J. van der Spiegel, N. Takahashi, A. Apsel and P. Mueller __ 1_ l I I I I S(e)T(e)+S(o)T(o) Right Motion (+vx) ' -~l r .. t~ Wx - l =' ~ = '®;. I I I I (a) S(e)T(e)-S(o)T(o) Left Motion (-vx) 60 30 Wt o [Hz] -30 -60 Tuned Velocity: Vx = 6.3 mmls -3.0 -1.5 0 I OOx [Vmm] (b) Figure 3: (a) Constructing Oriented Spatiotemporal Filters. (b) Contour Plot of One of the Filters Implemented. 3 HARDWARE IMPLEMENTATION 3.1 GENERAL PURPOSE ANALOG NEURAL COMPUTER The computer is intended for fast prototyping of neural network based applications. It offers the flexibility of programming combined with the real-time performance of a hardware system (Mueller, 1995). It is modeled after the biological nervous system, i.e. the cerebral cortex, and consists of electronic analogs of neurons, synapses, synaptic time constants and axon/dendrites. The hardware modules capture the functional and computational aspects of the biological counterparts. The main features of the system are: configurable interconnection architecture, programmable neural elements, modular and expandable architecture, and spatiotemporal processing. These features make the network ideal to implement a wide range of network architectures and applications. The system, shown in part in figure 4, is constructed from three types of modules (chips): (1) neurons, (2) synapses and (3) synaptic time constants and axon/dendrites. The neurons have a piece-wise linear transfer function with programmable (8bit) threshold and minimum output at threshold. The synapses are implemented as a programmable resistance whose values are variable (8 bit) over a logaritnmic range between 5KOhm and lOMohm. The time constant, realized with a load-compensated transconductance amplifier, is selectable between O.5ms and Is with a 5 bit resolution. The axon/dendrites are implemented with an analog cross-point switch matrix. The neural computer has a total of 1024 neurons, distributed over 64 neuron modules, with 96 synaptic inputs per neuron, a total of 98,304 synapses, 6,656 time constants and 196,608 cross point switches. Up to 3,072 parallel buffered analog inputs/outputs and a neuron output analog mulitplexer are available. A graphical interface software, which runs on the host computer, allows the user to symbolically and physically configure the network and display its behavior (Donham, 1994). Once a particular network has been loaded, the neural network runs independently of the digital host and operates in a fully analog, parallel and continuous time fashion. 3.2 NEURAL IMPLEMENTATION OF SPATIOTEMPORAL FILTERS The output of the silicon retina, which transforms a gray scale image into a binary image of edges, is presented to the neural computer to implement the oriented spatiotemporal filters. The first and second derivatives of Gaussian functions are chosen to implement the odd and even spatial filters, respectively. They are realized by feeding the outputs of VLSI Implementation of Cortical VISual Motion Detection Neuron~ ~ Synapses wij "'/ kL.~ ILft~ • • • ( ANALOG INPUTS AND OUTPUTS ~ • c •• Time onstants .::oW itches Figure 4: Block Diagram of the Overall Neural Network Architecture. 689 the retina, with appropriate weights, into a layer of neurons. Three parallel channels with varying spatial scales are implemented for each dimension. The output of the even (odd) spatial filter is subsequently fed to three parallel even (odd) temporal filters, which also have varying temporal tuning. Hence, three non-oriented pairs of spatiotemporal filters are realized for each channel. Six oriented filters are realized by summing arx.l differencing the non-oriented pairs. The oriented filters are rectified, and lateral inhibition is used to accentuate the higher response. Figure 4 shows a schematic of the neural circuitry used to implement the orientation selective filters. The image layer of the network in figure 5 is the direct, parallel output of the silicon retina. A 7 x 7 pixel array from the retina is decomposed into 2, 1 x 7 orthogonal linear images, and the nine motion detection filters are implemented per image. The total number of neurons used to implement this network is 152, the number of synapse is 548 and the number of time-constants is 108. The time-constant values ranges from 0.75 ms to 375 ms. After the networks have been programmed into the VLSI chips of the neural computer, the system operates in full parallel and continuous time analog mode. Consequently, this system realizes a silicon model for biological visual image motion measurement, starting from the retina to the visual cortex. Odd Spatial Filter S(o)"' dG(x)ldx"' 2xExp(-x2) Even Spatial Filter S(e)"' ()2G(x)ldx2= C4x2_2)Exp(-x2) !l\ It\. .ffit \i'V ~u- ~ ooW 0 ° 0?1b210 Velocity Selective Spatiotemporal Filters Figure 5: Neural Network Implementation of the Oriented Spatiotemporal Filters. 690 R. Etienne-Cummings, 1. van der Spiegel, N. Takahashi, A. Apsel and P. Mueller 4 RESULTS The response of the spatiotemporal filters implemented with the neural computer are shown in figure 6. The figure is obtained by sampling the output of the neurons at IMHz using the on-chip analog multiplexers. In figure 6a, the impulse response of the spatial filters are shown as a point moves across their receptive field. Figure 6b shows the outputs of the even and odd temporal filters for the moving point. At the output of the filters, the even and odd signals from the spatial filters are no longer out of phase. This transformation yields to constructive or destructive interference when they are summed and differenced. When the point move in opposite direction, the odd filters changes such that the output of the temporal filters become 1800 out of phase. Subsequent summing and differencing will have the opposite result. Figure 6c shows the output for all nine x-velocity selective filters as a point moves with positive velocity . ..... ,.. .; ...... ;...... [~::o~:':l : ·:~~:·~··11 :·~~,·~'·····I ~ ...... :....... .. + .... ; ... +....·-f--·+·i .... · .... f .... j .... ! .... · liliIt1±JffiIJ 0].0 .. _00 (a)::!«':!: ti.::li~·!1 =·T· .. r .... ' .... 1 .. ; .... ··1 ...... · .. ··: .. 1· .. · .. 1 ........... : .+ .. -+ .. .. , .. , ........ , ~ ", c:~::::::t:~·i·' :; to;:::': I .. -.. ":" ._-:--_ .. _!....... ::: ::: ::: ..... ; .-: ...... -;...... ·····r···-r·····r···-- ····t·o··r····r····· ··-··r····r-··· ·~· ·· · 6 1 .; ... 11 j' I I ... : .. : ... ::;.: .... i.: .. ·1 .. ·: .. ~ .. :·;· .. ··; .. · .. I .. ·'..:: .... :;:· .. ·i. "'1 diJ"i.C~) .~IT.r::l{~)i,;.~.r+: Figure 6: Output of the Neural Circuits for a Moving Point: (a) Spatial Filters, (b) Temporal Filters and (c) Motion Filters. Figure 7 shows the tuning curves for the filters tuned to x-motion. The variations in the responses are due to variations in the analog components of the neural computer. Some aliasing is noticeable in the tuning curves when there is a minor peak in the opposite direction. This results from the discrete properties of the spatial filters, as seen in 1.0 S 08 ~ ~ -= 06 u:: '" ~ :a 0.4 e 0 z 0.2 0.0 Tuning Curves .~. " ·8.0 ·6.0 .4.0 ·2 0 00 20 4.0 Speed [em/sl Figure 7: Tuning Curves for the Nine X-Motion Filters. 6.0 80 ___ +2~ mml, ·._ •• -27mmfs ___ .c)mmls .. _g_. "mmls --+-- .... mmls __ ._ •. 5 mmfs --...- +22 mmls :..::..:..: +l!:::~: __ •• _ -6 mm/s ___ +) .~ mm/:l - -e -- -3.2 mmls --+-- +14.5 mml~ --. -- -Bmm/s --..-- +3 ~ mm/3 --a - --6 mm/s _+3mm1s --9 - - -2.t mmls VLSI Implementation a/Cortical Visual Motion Detection 691 figure 3b. Due to the lateral inhibition employed, the aliasing effects are minimal. Similar curves are obtained for the y-motion tuned filters. For a point moving with v. = 8.66 mmls and Vy = 5 mmls, the output of the motion filters are shown in Table I. Computing a weighted average using equation 2, yields v.m = 8.4 mmls and v ym = 5.14 mmls. This result agrees with the actual motion of the point. v = ~ vtunedo./~ 0 mkJllL.J1 i i (2) Table 1: Filter Responses for a Point Moving at 10 mmls at 30°. X Filters [Speed in mmls] Y Filters [Speed in mm/s] Ifuned Speed 25 9 4 22 7 3.5 14. 3.5 3 26 9.5 5 20 7.8 3.7 15 4. 1 l!iesQonse 0.52 0.95 0.57 0.53 0.9 0.3 0.7 0.9 0.31 0.35 0.67 0.92 0.3 0.85 0.9 0.54 0.9 truned ~peed -27 -7.7 -5 -18 -6 -3.2 -13 -6 -2.1 -25 -8 -4.1 -21 -7 -4 -14 -5 3.5 0.9 -2 Response 0.0 0.05 0.1 0. 1 0.05 0.05 0.0 0.05 0.1 0.1 O.O~ 0.1 0.3 0.05 0.01 0.23 0.05 0.1 5 CONCLUSION 2D image motion estimation based on spatiotemporal feature extraction has been implemented in VLSI hardware using a general purpose analog neural computer. The neural circuits capitalize on the temporal processing capabilities of the neural computer. The spatiotemporal feature extraction approach is based on the 1 D cortical motion detection model proposed by Adelson and Bergen, which was extended to 2D by Heeger et al. To reduce the complexity of the model and to allow realization with simple sum-andthreshold neurons, we further modify the 2D model by placing filters only in the (O.-rot and (Oy-(Ot planes, and by replacing quadratic non-linearities with a rectifiers. The modifications do not affect the performance of the model. While this technique of image motion detection requires too much hardware for focal plane implementation, our results show that it is realizable when a silicon "brain," with large numbers of neurons and synaptic time constant, is available. This is very reminiscent of the biological master. References E. Adelson and J. Bergen, "Spatiotemporal Energy Models for the Perception of Motion," 1. Optical Society of America, Vol. A2, pp. 284-99, 1985 C. Donham, "Real Time Speech Recognition using a General Purpose Analog Neurocomputer," Ph.D. Thesis, Univ. of Pennsylvania, Dept. of Electrical Engineering, Philadelphia, PA, 1995. D. Heeger, E. Simoncelli and J. Movshon, "Computational Models of Cortical Visual Processing," Proc. National Academy of Science, Vol. 92, no. 2, pp. 623, 1996 D. Heeger, "Model for the Extraction of Image Flow," 1. Optical Society of America, Vol. 4, no. 8, pp. 1455-71 , 1987 D. Hubel and T. Wiesel, "Receptive Fields, Binocular Interaction and Functional Architecture in the Cat's Visual Cortex," 1. Physiology, Vol. 160, pp. 106-154, 1962 J. Maunsell and D. Van Essen, "Functional Properties of Neurons in Middle Temporal Visual Area of the Macaque Monkey. I. Selectivity for Stimulus Direction, Speed and Orientation," 1. Neurophysiology, Vol. 49, no. 5, pp. 1127-47, 1983 P. Mueller, 1. Van der Spiegel, D. Blackman, C. Donham and R. Etienne-Cummings, "A Programmable Analog Neural Computer with Applications to Speech Recognition," Proc. Compo & Info. Sci. Symp. (CISS), J. Hopkins, May 1995.
1996
41
1,238
Local Bandit Approximation for Optimal Learning Problems Michael o. Duff Andrew G. Barto Department of Computer Science University of Massachusetts Amherst, MA 01003 {duff.barto}Ccs.umass.edu Abstract In general, procedures for determining Bayes-optimal adaptive controls for Markov decision processes (MDP's) require a prohibitive amount of computation-the optimal learning problem is intractable. This paper proposes an approximate approach in which bandit processes are used to model, in a certain "local" sense, a given MDP. Bandit processes constitute an important subclass of MDP's, and have optimal learning strategies (defined in terms of Gittins indices) that can be computed relatively efficiently. Thus, one scheme for achieving approximately-optimal learning for general MDP's proceeds by taking actions suggested by strategies that are optimal with respect to local bandit models. 1 INTRODUCTION Watkins [1989] has defined optimal learning as:" the process of collecting and using information during learning in an optimal manner, so that the learner makes the best possible decisions at all stages of learning: learning itself is regarded as a mUltistage decision process, and learning is optimal if the learner adopts a strategy that will yield the highest possible return from actions over the whole course of learning." For example, suppose a decision-maker is presented with two biased coins (the decision-maker does not know precisely how the coins are biased) and asked to allocate twenty flips between them so as to maximize the number of observed heads. Although the decision-maker is certainly interested in determining which coin has a higher probability of heads, his principle concern is with optimizing performance en route to this determination. An optimal learning strategy typically intersperses "exploitation" steps, in which the coin currently thought to have the highest proba1020 1 1- P" M. O. Duff and A. G. Barto ~ 1 P • f'r:i) ~-QP22 (a) """~ y i 1;(3,1)(1,1):(1,1)(1,1) ~ 2 / 3 ~2;(3,2)(1,1):(1,1)(1,1) .7 (b) 1 1- P12 aJ 1/3 ~ l-P~, ~1;(2'1)(1'1) :(1'1)(1'1) l~ ~ 1/2 ••• 2 ~~P 2 a P R" R2 :y 22 "" ./ ll~ 1 2 a, ;~ 2:(1,2)(1,1):(1,1)(1,1) ~ 7 a, l-p 2 1;(1,1)(1,1):(1,1)(1.1) > 2;(1.2)(1,1):(1,1)(2,1) <{: 22 a, )fo aj ~ (c) Figure 1: A simple example: dynamics/rewards under (a) action 1 and (b) action 2. (c) The decision problem in hyperstate space. bility of heads is flipped, with "exploration" steps in which, on the basis of observed flips, a coin that would be deemed inferior is flipped anyway to further resolve its true potential for turning up heads. The coin-flip problem is a simple example of a (two-armed) bandit problem. A key feature of these problems, and of adaptive control processes in general, is the so-called "exploration-versus-exploitation trade-off" (or problem of "dual control" [Fel'dbaum, 1965]). As an another example, consider the MDP depicted in Figures l(a) and (b). This is a 2-state/2-action proceSSj transition probabilities label arcs, and quantities within circles denote expected rewards for taking particular actions in particular states. The goal is to assign actions to states so as to maximize, say, the expected infinite horizon discounted sum ofrewards (the value function) over all states. For the case considered in this paper, the transition probabilites are not known. Given that the process is in some state, one action may be optimal with respect to currentlyperceived point-estimates of unknown parameters, while another action may result in greater information gain. Optimal learning is concerned with striking a balance between these two criteria. While reinforcement learning approaches have recognized the dual-effects of control, at least in the sense that one must occasionally deviate from a greedy policy to ensure a search of sufficient breadth, many exploration procedures appear not to be motivated by real notions of optimallearningj rather, they aspire to be practical schemes for avoiding unrealistic levels of sampling and search that would be required if one were to strictly adhere to the theoretical sufficient conditions for convergence-that all state-action pairs must be considered infinitely many times. If one is willing to adopt a Bayesian perspective, then the exploration-versusexploitation issue has already been resolved, in principle. A solution was recognized by Bellman and Kalaba nearly fo rty years ago [Bellman & Kalaba, 1959]j their dynamic programming algorithm for computing Bayes-optimal policies begins by regarding "state" as an ordered pair, or "hyperstate," (:z:,I), where :z: is a point in phase-space (Markov-chain state) and I is the "information pattern," which summarizes past history as it relates to modeling the transitional dynamics of :z:. Computation grows increasingly burdensome with problem size, however, so one is compelled to seek approximate solutions, some of which ignore the effects of information gain entirely. In contrast, the approach suggested in this paper explicitly acknowledges that there is an information-gain component to the optimal learnLocal Bandit Approximation/or Optimal Learning Problems 1021 ing problem; if certain salient aspects of the value of information can be captured, even approximately, then one may be led to a reasonable method for approximating optimal learning policies. Here is the basic idea behind the approach suggested in this paper: First note that there exists a special class of problems, namely multi-armed bandit problems, in which the information pattern is the sole component of the hyperstate. These special problems have the important feature that their optimal policies can be defined concisely in terms of "Gittins indices," and these indices can be computed in a relatively efficient way. This paper is an attempt to make use of the fact that this special subclass of MDP's has tractably-computable optimal learning strategies. Actions for general MDP's are derived by, first, attaching to a given general MDP in a given state a "local" n-armed bandit process that captures some aspect of the value of information gain as well as explicit reward. Indices for the local bandit model can be computed relatively efficiently; the largest index suggests the best action in an optimal-learning sense. The resulting algorithm has a receding-horizon flavor in that a new local-bandit process is constructed after each transition; it makes use of a mean-process model as in some previously-suggested approximation schemes, but here the value of information gain is explicitly taken into account, in part, through index calculations. 2 THE BAYES-BELLMAN APPROACH FOR ADAPTIVE MDP'S Consider the two-state, two-action process shown in Figure 1, and suppose that one is uncertain about the transition probabilities. If the process is in a given state and an action is taken, then the result is that the process either stays in the state it is in or jumps to the other state-one observes a Bernoulli process with unknown parameter-just as in the coin-flip example. But in this case one observes four Bernoulli processes: the result of taking action 1 in state 1, action 1 in state 2, action 2 in state 1, action 2 in state 2. So if the prior probability for staying in the current state, for each of these state-action pairs, is represented by a beta distribution (the appropriate conjugate family of distributions with regard to Bernoulli sampling; I.e., a Bayesian update of a beta prior remains beta), then one may perform dynamic programming in a space of "hyperstates," in which the components are four pairs of parameters specifying the beta distributions describing the uncertainty in the transition probabilities, along with the Markov chain state: (:z:, (aL,Bt), (a~,,B~)(a~,,Bn, (a~,,B~»), where for example (aL,BD denotes the parmeters specifying the beta distribution that represents uncertainty in the transition probability P~l. Figure l(c) shows part ofthe associated decision tree; an optimality equation may be written in terms of the hyperstates. MDP's with more than two states pose no special problem (there exists an appropriate generalization of the beta distribution). What i& a problem is what Bellman calls the "problem of the expanding grid:" the number of hyperstates that must be examined grows exponentially with the horizon. How does one proceed if one is constrained to practical amounts of computation and is willing to settle for an approximate solution? One could truncate the decision tree at some shorter and more manageable horizon, compute approximate terminal values by replacing the distributions with their means, and proceed with a recedinghorizon approach: Starting from the approximate terminal values at the horizon, perform a backward sweep of dynamic programming, computing an optimal policy. Take the initial action of the policy, then shift the entire computational window forward one level and repeat. One can imagine a sort of limiting, degenerate version 1022 M. O. Duff and A. G. Barto of this receding horizon approach in which the horizon is zerOj that is, use the means of the current distributions to calculate an optimal policy, take an "optimal" action, observe a transition, perform a Bayesian modification of the prior, and repeat. This (certainty-equivalence) heuristic was suggested by [Cozzolino et al., 1965], and has recently reappeared in [Dayan & Sejnowski, 1996]. However, as was noted in [Cozzolino et al., 1965] " ... the trade-off between immmediate gain and information does not exist in this heuristic. There is no mechanism which explicitly forces unexplored policies to be observed in early stages. Therefore, if it should happen that there is some very good policy which a priori seemed quite bad, it is entirely possible that this heuristic will never provide the information needed to recognize the policy as being better than originally thought .. .'t This comment and others seem to refer to what is now regarded as a problem of "identifiability" associated with certainty-equivalence controllers in which a closed-loop system evolves identically for both true and false values of the unknown parametersj that is, certainty-equivalence control may make some of the unknown parameters invisible to the identification process and lead one to repeatedly choose the wrong action (see [Borkar & Varaiya, 1979], and also Watkinst discussion of "metastable policiestt in [Watkins, 1989]). 3 BANDIT PROBLEMS AND INDEX COMPUTATION One basic version of the bandit problem may be described as follows: There are some number of statistically independent reward processes-Markov chains with an imposed reward structure associated with the chain's arcs. At each discrete time-step, a decision-maker selects one of these processes to activate. The activated process yields an immediate reward and then changes state. The other processes remain frozen and yield no reward. The goal is to splice together the individual reward streams into one sequence having maximal expected discounted value. The special Cartesian structure of the bandit problem turns out to imply that there are functions that map process-states to scalars (or "indices't), such that optimal policies consist simply of activating the task with the largest index. Consider one of the reward processes, let S be its state space, and let B be the set of all subsets of S. Suppose that :z:(k) is the state of the process at time k and, for B E B, let reB) be the number of transitions until the process first enters the set B. Let v(ij B) be the expected discounted reward per unit of discounted time starting from state i until the stopping time reB): Then the Gittins index of state i for the process under consideration is v(i) = maxv(ijB). BEB (1) [Gittins & Jones, 1979] shows that the indices may be obtained by solving a set of functional equations. Other algorithms that have been suggested include those by Beale (see the discussion section following [Gittins & Jones, 1979]), [Robinsion, 1981], [Varaiya et al., 1985], and [Katehakis & Vein ott , 1987]. [Dufft 1995] provides a reinforcement learning approach that gradually learns indices through online/model-free interaction with bandit processes. The details of these algorithms would require more space than is available here. The algorithm proposed in the next section makes use of the approach of [Varaiya et al., 1985]. Local Bandit Approximation for Optimal Learning Problems 4 LOCAL BANDIT APPROXIMATION AND AN APPROXIMATELY-OPTIMAL LEARNING ALGORITHM 1023 The most obvious difference between the optimal learning problem for an MDP and the multi-armed bandit problem is that the MDP has a phase-space component (Markov chain state) to its hyperstate. A first step in bandit-based approximation, then, proceeds by "removing" this phase-space component. This can be achieved by viewing the process on a time-scale defined by the recurrence time of a given state. That is, suppose the process is in some state, z. In response to some given action, two things can happen: (1) The process can transition, in one time-step, into z again with some immediate reward, or (2) The process can transition into some state that is not z and experience some "sojourn" path of states and rewards before returning to z. On a time-scale defined by sojourn-time, one can view the process in a sort of "state-z-centric" way (if state z never recurs, then the sojourn-time is "infinite" and there is no value-of-information component of the local bandit model to acknowledge). From this perspective, the process appears to have only one state, and is 8em~Markov; that is, the time between transitions is a random variable. Some other action taken in state z would give rise to a different sojourn reward process. For both processes (sojourn-processes initiated by different actions applied to state z), the sojourn path/reward will depend upon the policy for states encountered along sojourn paths, but suppose that this policy is fixed for the moment. By viewing the original process on a time-scale of sojourn-time, one has effectively collapsed the phase-space component of the hyperstate. The new process has one state, z, and the problem of choosing an action, given that one is uncertain about the transition probabilities, presents itself as a semi-Markov bandit problem. The preceding discussion suggests an algorithm for approximately-optimal learning: (0) Given that the uncertainty in transition probabilities is expressed in terms of sufficient statistics < a, Ii >, and the process is currently in state Zt. (1) Compute the optimal policy for the mean process, 7r·[F(a,Ii)]; that is, compute the policy that is optimal for the MDP whose transition probabilities are taken to be the mean values associated with < a, Ii >-this defines a nominal (certainty-equivalent) policy for sojourn states. (2) Construct a local bandit model at state Zt; that is, the decision-maker must choose between some number (the number of admissible actions) of sojourn reward processes-this is a semi-Markov multi-armed bandit problem. (3) Compute the Gittins indices for the local bandit model. ( 4) Take the action with the largest index. (5) Observe a transition to Zt+l in the underlying MDP. (6) Update < a,1i > accordingly (Bayes update). (7) Go to step (1) The local semi-Markov bandit process associated with state 1 / action 1 for the 2-state example MDP of Figure 1 is shown in Figure 2. The sufficient statistics for ptl are denoted by (Q, f3), and Q~.8 and ~ are the expected probabilities for transition into state 1 and state 2, respectively. rand R121 are random variables signifying sojourn time and reward. The goal is to compute the index for the root information-state labeled < Q, f3 > and to compare it with that computed for a similar diagram associated with the bandit 1024 M. O. Duff and A. G. Barto / r ~/ ~,y.~ /,~I u+1, Figure 2: A local semi-Markov bandit process associated with state 1 / action 1 for the 2-state example MDP of Figure 1. process for taking action 2. The approximately-optimal action is suggested by the process having the largest root-node index. Indices for semi-Markov bandits can be obtained by considering the bandits as Markov, but performing the optimization in Equation 1 over a restricted set of stopping times. The algorithm suggested in [Tsitsiklis, 1993], which in turn makes use of methods described in [Varaiya et al., 1985], proceeds by "reducing" the graph through a sequence of node-excisions and modifications of rewards and transition probabilities; [Duff, 1997] details how these steps may be realized for the special semi-Markov processes associated with problems of optimal learning. 5 Discussion In summary, this paper has presented the problem of optimal learning, in which a decision-maker is obliged to enjoy or endure the consequences of its actions in quest of the asymptotically-learned optimal policy. A Bayesian formulation of the problem leads to a clear concept of a solution whose computation, however, appears to entail an examination of an intractably-large number of hyperstates. This paper has suggested extending the Gittins index approach (which applies with great power and elegance to the special class of multi-armed bandit processes) to general adaptive MDP's. The hope has been that if certain salient features of the value of information could be captured, even approximately, then one could be led to a reasonable method for avoiding certain defects of certainty-equivalence approaches (problems with identifiability, "metastability"). Obviously, positive evidence, in the form of empirical results from simulation experiments, would lend support to these ideas- work along these lines is underway. Local bandit approximation is but one approximate computational approach for problems of optimal learning and dual control. Most prominent in the literature of control theory is the "wide-sense" approach of [Bar-Shalom & Tse, 1976], which utilizes local quadratic approximations about nominal state/control trajectories. For certain problems, this method has demonstrated superior performance compared to a certainty-equivalence approach, but it is computationally very intensive and unwieldy, particularly for problems with controller dimension greater than one. One could revert to the view of the bandit problem, or general adaptive MDP, as simply a very large MDP defined over hyperstates, and then consider a someLocal Bandit Approximationfor Optimal Learning Problems 1025 what direct approach in which one performs approximate dynamic programming with function approximation over this domain-details of function-approximation, feature-selection, and "training" all become important design issues. [Duff, 1997] provides further discussion of these topics, as well as a consideration of actionelimination procedures [MacQueen, 1966] that could result in substantial pruning of the hyperstate decision tree. Acknowledgements This research was supported, in part, by the National Science Foundation under grant ECS-9214866 to Andrew G. Barto. References Bar-Shalom, Y. 8£ Tse, E. (1976) Caution, probing and the value of information in the control of uncertain systems, Ann. Econ. Soc. Meas. 5:323-337. R. Bellman 8£ R. Kalaba, (1959) On adaptive control processes. IRE Trans., 4:1-9. Bokar, V. 8£ Varaiya, P.P. (1979) Adaptive control of Markov chains I: finite parameter set. IEEE Trans. Auto. Control 24:953-958. Cozzolino, J.M., Gonzalez-Zubieta, R., 8£ Miller, R.L. (1965) Markov decision processes with uncertain transition probabilities. Tech. Rpt. 11, Operations Research Center, MIT. Dayan, P. 8£ Sejnowski, T. (1996) Exploration Bonuses and Dual Control. Machine Learning (in press). Duff, M.O. (1995) Q-Iearning for bandit problems. in Machine Learning: Proceedings of the Twelfth International Conference on Machine Learning: pp. 209-217. Duff, M.O. (1997) Approximate computational methods for optimal learning and dual control. Technical Report, Deptartment of Computer Science, Univ. of Massachusetts, Amherst. Fel'dbaum, A. (1965) Optimal Control Systems, Academic Press. Gittins, J.C. 8£ Jones, D. (1979) Bandit processes and dynamic allocation indices (with discussion). J. R. Statist. Soc. B 41:148-177. Katehakis, M.H. 8£ Veinott, A.F. (1987) The multi-armed bandit problem: decomposition and computation Math. OR 12: 262-268. MacQueen, J. (1966). A modified dynamic programming method for Markov decision problems, J. Math. Anal. Appl., 14:38-43. Robinsion, D.R. (1981) Algorithms for evaluating the dynamic allocation index. Research Report No. 80/DRR/4, Manchester-Sheffield School of Probability and Statistics. Tsitsiklis, J. (1993) A short proof of the Gittins index theorem. Proc. 3fnd Conf. Dec. and Control: 389-390. Varaiya, P.P., Walrand, J.C., 8£ Buyukkoc, C. (1985) Extensions of the multiarmed bandit problem: the discounted case. IEEE Trans. Auto. Control 30(5):426-439. Watkins, C. (1989) Learning /rom Delayed Rewards Ph.D. Thesis, Cambidge University.
1996
42
1,239
Learning Appearance Based Models: Mixtures of Second Moment Experts Christoph 8regler and Jitendra Malik Computer Science Division University of California at Berkeley Berkeley, CA 94720 email: bregler@cs.berkeley.edu, malik@cs.berkeley.edu Abstract This paper describes a new technique for object recognition based on learning appearance models. The image is decomposed into local regions which are described by a new texture representation called "Generalized Second Moments" that are derived from the output of multiscale, multiorientation filter banks. Class-characteristic local texture features and their global composition is learned by a hierarchical mixture of experts architecture (Jordan & Jacobs). The technique is applied to a vehicle database consisting of 5 general car categories (Sedan, Van with back-doors, Van without back-doors, old Sedan, and Volkswagen Bug). This is a difficult problem with considerable in-class variation. The new technique has a 6.5% misclassification rate, compared to eigen-images which give 17.4% misclassification rate, and nearest neighbors which give 15.7% misclassification rate. 1 Introduction Until a few years ago neural network and other statistical learning techniques were not very popular in computer vision domains. Usually such techniques were only applied to artificial visual data or non-mainstream problems such as handwritten digit recognition. A significant shift has occurred recently with the successful application of appearance-based or viewer-centered techniques for object recognition, supplementing the use of 3D models. Appearance-based schemes rely on collections of images of the object. A principal advantage is that they implicitly capture both shape and photometric information(e.g. surface reflectance variation). They have been most sucessfully applied in the domain of human faces [15, 11, I, 14] though other 3d objects under fixed lighting have also been considered [13]. View-based representations lend themselves very naturally to learning from examples- principal component analysis[15, 13] and radial basis functions[l] have been used. Approaches such as principal component analysis (or "eigen-images") use global representations at the image level. The objective of our research was to develop a representation which would 846 C. Bregler and J. Malik be more 'localist', where representations of different 'parts' of the object would be composed together to form the representation of the object as a whole. This appears to be essential in order to obtain robustness to occlusion and ease of generalization when different objects in a class may have variations in particular parts but not others. A part based view is also more consistent with what is known about human object recognition (Tanaka and collaborators). In this paper, we propose a domain independent part decomposition using a 2D grid representation of overlapping local image regions. The image features of each local patch are represented using a new texture descriptor that we call "Generalized Second Moments". Related representations have already been successfully applied to other early-vision tasks like stereopsis, motion, and texture discrimination. Class-based local texture features and their global relationships are induced using the "Hierarchical Mixtures of Experts" Architecture (HME) [8]. We apply this technique to the domain of vehicle classification. The vehicles are seen from behind by a camera mounted above a freeway(Figure 1). We urge the reader to examine Figure 3 to see examples of the in class variations in the 5 different categories. Our technique could classify five broader categories with an error of as low as 6.5% misclassification, while the best results using eigen-images and nearest neighbor techniques were 17.4% and 15.7% mis-classification error. Figure 1: Typical shot of the freeway segment 2 Representation An appearance based representation should be able to capture features that discriminate the different object categories. It should capture both local textural and global structural information. This corresponds roughly to the notion in 3D object models of (i) parts (ii) relationship between parts. 2.1 Structural Description Objects usually can be decomposed into parts. A face consists of eyes, nose, and mouth. Cars are made out of window screens, taillights, license plates etc. The question is what granularity is appropriate and how much domain knowledge should be exploited. A car could be a single part in a scene, a license plate could be a part, or the letters in the license plate could be the decomposed parts. Eyes, nose, and mouth could be the most important parts of a face for recognition, but maybe other parts are important as well. It would be advantageous if each part could be described in a decoupled way using a representation that was most appropriate for it. Object classification should be based on these local part descriptions and the relationship between the parts. The partitioning reduces the complexity greatly and invariance to the precise relation between the parts could be achieved. For our domain of vehicle classification we don't believe it is appropriate to explicitly code any Learning Appearance Based Models: Mixtures of Second Moment Experts 847 part decomposition. The kind and number of useful parts might vary across different car makes. The resolution of the images (1 OOx 100 pixel) restricts us to a certain degree of granularity. We decided to decompose the image using a 20 grid of overlapping tiles or Gaussian windows but only local classification for each tile region is done. The content of each local tile is represented by a feature vector (next section). The generic grid representation allows the mixture of experts architecture to induce class-based part decomposition, and extract local texture and global shape features. For example the outline of a face could be represented by certain orientation dominances in the local tiles at positions of the face boundary. The eyes are other characteristic features in the tiles. 2.2 Local Features We wanted to extract characteristic features from each local tile. The traditional computer vision approach would be to find edges and junctions. The weakness of these representations is that they do not capture the richness of textured regions, and the hard decision thresholds make the measurement process non-robust. An alternative view is motivated by our understanding of processing in biological vision systems. We start by convolving image regions with a large number of spatial filters, at various orientations, phases, and scales. The response values of such filters contain much more general information about the local neighborhood, a fact that has now been recognized and exploited in a number of early vision tasks like stereopsis, motion and texture analysis [16,9,6, 12, 7]. Although this approach is loosely inspired by the current understanding of processing in the early stages of the primate visual system, the use of spatial filters has many advantages from a pure analytical viewpoint[9, 7]. We use as filter kernels, orientation selective elongated Gaussian derivatives. This enables one to gain the power of orientation specific features, such as edges, without the disadvantage of non-robustness due to hard thresholds. If multiple orientations are present at a single point (e.g. junctions), they are represented in a natural way. Since multiple scales are used for the filters, no ad hoc choices have to be made for the scale parameters of the feature detectors. Interestingly the choices of these filter kernels can also be motivated in a learning paradigm as they provide very useful intermediate layer units in convolutional neural networks [3]. The straightforward approach would then be to characterize each image pixel by such a vector of feature responses. However note that there is considerable redundancy in the filter responsesparticularly at coarse scales, the responses of filters at neighboring pixels are strongly correlated. We would like to compress the representation in some way. One approach might be to subsample at coarse scales, another might be to choose feature locations with local magnitude maxima or high responses across several directions. However there might be many such interesting points in an image region. It is unclear how to pick the right number of points and how to order them. Leaving this issue of compressing the filter response representation aside for the moment, let us study other possible representations of low level image data. One way of representing the texture in a local region is to calculate a windowed second moment matrix [5]. Instead of finding maxima offilterresponses, the second moments of brightness gradients in the local neighborhood are weighted and averaged with a circular Gaussian window. The gradient is a special caSe of Gaussian oriented filter banks. The windowed second moment matrix takes into account the response of all filters in this neighborhood. The disadvantage is that gradients are not very orientation selective and a certain scale has to be selected beforehand. Averaging the gradients "washes" out the detailed orientation information in complex texture regions. Orientation histograms would avoid this effect and have been applied successfully for classification [4]. Elongated families of oriented and scaled kernels could be used to estimate the orientation at each point. But as pointed out already, there might be more than one orientation at each point, and significant information is lost. 848 C. Bregler and 1. Malik Figure 2: Left image: The black rectangle outlines the selected area of interest. Right image: The reconstructed scale and rotation distribution of the Generalized Second Moments. The horizontal axis are angles between 0 and 180 degrees and the vertical axis are different scales. 3 Generalized Second Moments We propose a new way to represent the texture in a local image patch by combining the filter bank approach with the idea of second moment matrices. The goal is to compute a feature vector for a local image patch that contains information about the orientation and scale distribution. We compute for each pixel in the image patch the R basis kernel responses (using X-Y separable steerable scalable approximations of a rich filter family). Given a spatial weighting function of the patch (e.g. Gaussian), we compute the covariance matrix of the weighted set of R dimensional vectors. In [2] we show that this covariance matrix can be used to reconstruct for any desired oriented and scaled version of the filter family the weighted sum of all filter response energies: E(O, CT) = L W(x, y)[F(J,a(x, y)f (1) X,Y Using elongated kernels produces orientation/scale peaks, therefore the sum of all orientation/scale responses doesn't "wash" out high peaks. The height of each individual peak corresponds to the intensity in the image. Little noisy orientations have no high energy responses in the sum. E( 0, CT) is somehow a "soft" orientation/scale histogram of the local image patch. Figure 2 shows an example of such a scale/orientation reconstruction based on the covariance matrix (see [2] for details). Three peaks are seen, representing the edge lines along three directions and scales in the local image patch. This representation greatly reduces the dimensionality without being domain specific or applying any hard decisions. It is shift invariant in the local neighborhood and decouples scale in a nice way. Dividing the R x R covariance matrix by its trace makes this representation also illumination invariant. Using a 10xlO grid and a kernel basis of 5 first Gaussian derivatives and 5 second Gaussian derivatives represents each input image as an 10 . 10 . (5 + 1) . 5 = 3000 dimensional vector (a 5 x 5 covariance matrix has (5 + 1) ·5 independent parameters). Potentially we could represent the full image with one generalized second moment matrix of dimension 20 if we don't care about capturing the part decomposition. 4 Mixtures of Experts Even if we only deal with the restricted domain of man-made object categories (e.g. cars), the extracted features still have a considerable in-class variation. Different car shapes and poses produce nonlinear class subspaces. Hierarchical Mixtures of Experts (HME by Jordan & Jacobs) are able to model such nonlinear decision surfaces with a soft hierarchical decomposition of the Learning Appearance Based Models: Mixtures of Second Moment Experts 849 Sedan Van1 Bug Old Van2 Figure 3: Example images of the five vehicle classes. feature space and local linear classification experts. Potentially different experts are "responsible" for different object poses or sub-categories. The gating functions decompose the feature space into a nested set of regions using a hierarchical structure of soft decision boundaries. Each region is the domain for a specific expert that classifies the feature vectors into object categories. We used generalized linear models (GUM). Given the training data and output labels, the gating functions and expert functions can be estimated using an iterative version of the EM-algorithm. For more detail see [8]. In order to reduce training time and storage requirements, we trained such nonlinear decision surfaces embedded in one global linear subspace. We choose the dimension of this linear subspace to be large enough, so that it captures most of the lower-dimensional nonlinearity (3000 dimensional feature space projected into an 64 dimensional subspace estimated by principal components analysis). 5 Experiments We experimented with a database consisting of images taken from a surveillance camera on a bridge covering normal daylight traffic on a freeway segment (Figure 1). The goal is to classify different types of vehicles. We are able to segment each moving object based on motion cues [10]. We chose the following 5 vehicle classes: Modem Sedan, Old Sedan, Van with back-doors, Van without back-door, and Volkswagen Bug. The images show the rear of the car across a small set of poses (Figure 3). All images are normalized to looxloo pixel using bilinear interpolation. For this reason the size or aspect ratio can not be used as a feature. We ran our experiments using two different image representations: • Generalized Second Moments: A 10 x 10 grid was used. Generalized second moments 850 C. Bregler and J. Malik Miucl ... Classiftcation Error 1.00 0.90 0.80 0.70 0.60 \. O.SO \\ 0.40 \'\ 0.30 -\"', ----'" :~~---~ ---.-.... " ... ::~:::~ ~.,.. --0.20 -~- -----:::: =:::==... 0.10 1"'00.. .-. GSM+HME 0.00 TraiD.Si2e 50.00 100.00 150.00 200.00 Figure 4: The classification errors of four different techniques. The X-axis shows the size of the training set, and the Y-axis shows the percentage of rnisclassified test images. HME stands for Hierarchical Mixtures of Experts, GSM stands for Generalized Second Moments, and I-NN stands for Nearest Neighbors. were computed 1 using a window of (F = 6 pixel, and 5 filter bases of 3: I elongated first and second Gaussian derivatives on a scale range between 0.25 and 1.0. • Principal Components Analysis ("Eigen-Images"): We used no grid decomposition and projected the global gray level vector into a 64 dimensional linear space. Two different classifiers were used: • HME architecture with 8 local experts. • A simple I-Nearest-Neighbor Classifier (1-NN). Figure 4 shows the classification error rate for all 4 combinations as a function of the size of the training set. Each experiment is run 5 times with different sampling of training and test images2 . The database consists of 285 example images. Therefore the number of test images are (285number of training images). Across all experiments the HME architecture based on Generalized Moments was superior to all other techniques. The best performance with a misclassification of 6.5% was achieved using 228 training images. When fewer than 120 training images are used, the HME archi tecture performed worse than nearest neighbors. The most common confusion was between sedans and "old" sedans. The second most confusion was between vans with back-doors, vans without back-doors, and old sedans. IWe experimented also with grid sizes between 6x6 to 16x16, and with 8 filter bases and a rectangle window for the second moment statistics without getting significant improvement 2Por a given training size n we trained 5 classifiers on 5 different training and test sets and computed the average error rate. The training and test set for each classifier was generated by the same database. The n training examples were randomly sampled from the database, and the remaining examples were used for the test set. Learning Appearance Based Models: Mixtures of Second Moment Experts 851 6 Conclusion We have demonstrated a new technique for appearance-based object recognition based on a 20 grid representation, generalized second moments, and hierarchical mixtures of experts. Experiments have shown that this technique has significant better performance than other representation techniques like eigen-images and other classification techniques like nearest neighbors. We believe that learning such appearance-based representations offers a very attractive methodology. Hand-coding features that could discriminant object categories like the different car types in our database seems to be a nearly impossible task. The only choice in such domains is to estimate discriminating features from a set of example images automatically. The proposed technique can be applied to other domains as well. We are planning to experiment with face databases, as well as larger car databases and categories to furtherinvestigate the utility of hierarchical mixtures of experts and generalized second moments. Acknowledgments We would like to thank Leo Breiman, Jerry Feldman, Thomas Leung, Stuart Russell, and Jianbo Shi for helpful discussions and Michael Jordan, Lawrence Saul, and Doug Shy for providing code. References [1] D. Beymer, A. Shashua, and T. Poggio. Example based image analysis and synthesis. M.J. T. A.l. Memo No. 1431, Nov 1993. [2] c. Bregler and J. Malik. Learning Appearance Based Models: Hierarchical Mixtures of Experts Approach based on Generalized Second Moments. Technical Report UCBIICSD-96-897, Compo Sci. Dep., U.c. Berkeley, http://www.cs/ breglerlsoft.html, 1996. [3] Y. Le Cun, B. Boser, J.S. Denker, S. Solla, R. Howard, and L. Jackel. Back-propagation applied to handwritten zipcode recognition. Neural Computation, 1(4), 1990. [4] W. Freeman and M. Roth. Orientation histograms for hand gesture recognition. In International Workshop on Automatic Face- and Gesture-Recognition, 1995. [5] J. Garding and T. Lindeberg. Direct computation of shape cues using scale-adapted spatial derivative operators. Int. J. of Computer Vision, 17, February 1996. [6] OJ. Heeger. Optical flow using spatiotemporal filters. Int. J. of Computer Vision, 1, 1988. [7] D. Jones and 1. Malik. Computational framework for determining stereo correspondence from a set of linear spatial filters. Image and Vision Computing, 10(10), 1992. [8] M.1. Jordan and R. A. Jacobs. Hierarchical mixtures of experts and the em algorithm. Neural Computation, 6(2), March 1994. [9] J.J. Koenderink. Operational significance of receptive field assemblies. BioI. Cybern., 58: 163-171, 1988. [10] D. Koller, J. Weber, and J. Malik. Robust multiple car tracking with occlusion reasoning. In Proc. Third European Conference on Computer Vision, pages 189-196, May 1994. [11] M. Lades,J.C. Vorbrueggen, J. Buhmann, J. Lange, C. von der Malsburg, and R.P. Wuertz. Distortion invariant object recognition in the dynamic link architecure. In IEEE Transactions on Computers, volume 42, 1993. [12] J. Malik and P. Perona. Preattentive texture discrimination with early vision mechanisms. J. Opt. Soc. Am. A, 7(5):923-932, 1990. [13] H. Murase and S.K. Nayar. Visual learning and recognition of 3-d objects from appearance. Int. J. Computer Vision, 14(1):5-24, January 1995. [14] H.A. Rowley, S. Baluja, and T. Kanade. Human face detection in visual scenes. In NIPS, volume 8, 1996. [15] M. Turk and A. Pentland. Eigenfaces for recognition. Journal of Cognitive Neuroscience, 3(1 ):71-86, 1991. [16] R.A. Young. The gaussian derivative theory of spatial vision: Analysis of cortical cell receptive field line-weighting profiles. Technical Report GMR-4920, General Motors Research, 1985.
1996
43
1,240
Interpreting images by propagating Bayesian beliefs Yair Weiss Dept. of Brain and Cognitive Sciences Massachusetts Institute of Technology E10-120, Cambridge, MA 02139, USA yweiss<opsyche.mit.edu Abstract A central theme of computational vision research has been the realization that reliable estimation of local scene properties requires propagating measurements across the image. Many authors have therefore suggested solving vision problems using architectures of locally connected units updating their activity in parallel. Unfortunately, the convergence of traditional relaxation methods on such architectures has proven to be excruciatingly slow and in general they do not guarantee that the stable point will be a global minimum. In this paper we show that an architecture in which Bayesian Beliefs about image properties are propagated between neighboring units yields convergence times which are several orders of magnitude faster than traditional methods and avoids local minima. In particular our architecture is non-iterative in the sense of Marr [5]: at every time step, the local estimates at a given location are optimal given the information which has already been propagated to that location. We illustrate the algorithm's performance on real images and compare it to several existing methods. 1 Theory The essence of our approach is shown in figure 1. Figure 1a shows the prototypical ill-posed problem: interpolation of a function from sparse data. Figure 1b shows a traditional relaxation approach to the problem: a dense array of units represents the value of the interpolated function at discretely sampled points. The activity of a unit is updated based on the local data (in those points where data is available) and the activity of the neighboring points. As discussed below, the local update rule can Interpreting Images by Propagating Bayesian Beliefs 909 /. r···_· .. ·_-_· __ · __ · __ ···_· .. ···_-_·_···_-_····· __ ··_-_··" '''--'''1 i y*? 0 9 I I y 0-0-6--0-0-0 ! i .. ___ ._. __ . __ . __ ._. __ ..... _. __ . __ . _____ ..... _. ___ ._. ________ j r···--·-- _ ····_···_-_···· __ ·_···_·_·······_····_·····-···_····_······ .. ··_··_···'1 ! y*O 0 91 I I 6 . ! I1U,cfX ! i .......... ! ! ll,crO .. O.. ..0 .. 0 .. 0 i i I1P,~ ! : ! ;, ___ • __ • _ _ __ ._ ......... ___ •••••• ____ • __ ............. _ ...... ....... __ ._1 a b c Figure 1: a. a prototypical ill-posed problem h. Traditional relaxation approach: dense array of units represent the value of the interpolated function. Units update their activity based on local information and the activity of neighboring units. c. The Bayesian Belief Propagation (BBP) approach. Units transmit probabilities and combine them according to probability calculus in two non-interacting streams. be defined such that the network converges to a state in which the activity of each unit corresponds to the value of the globally optimal interpolating function. Figure lc shows the Bayesian Belief Propagation (BBP) approach to the problem. As in the traditional approach the function is represented by the activity of a dense array of units. However the units transmit probabilities rather than single estimates to their neighbors and combine the probabilities according to the probability calculus. To formalize the above discussion, let Yk represent the activity of a unit at location k, and let Yk be noisy samples from the true function. A typical interpolation problem would be to minimize: J(Y) = L Wk(Yk - yk)2 + A L(Yi - Yi+l)2 (1) k Where we have defined Wk = a for grid points with no data, and Wk = 1 for points with data. Since J is quadratic, any local update in the direction of the gradient will converge to the optimal estimate. This yields updates of the sort: ( ( Yk-l + YHl ) ( '" » Yk +- Yk + TJk A 2 - Yk + Wk Yk - Yk (2) Relaxation algorithms differ in their choice of TJ: TJ = Ij(A + Wk) corresponds to Gauss-Seidel relaxation and TJ = 1.9j(A + Wk) corresponds to successive over relaxation (SOR) which is the method of choice for such problems [10]. To derive a BBP update rule for this problem, note that that minimizing J(Y) is equivalent to maximizing the posterior probability of Y given Y'" assuming the following generative model: Yi+l = Yi + 11 (3) Y; = WiYi + TJ (4) Where 11 "'" N(O, (TR), TJ "'" N(O , (TD) . The ratio of (TD to (TR plays a role similar to that of A in the original cost functional. The advantage of considering the cost functional as a posterior is that it enables us to use the methods of Hidden Markov Models, Bayesian Belief Nets and Optimal Estimation to derive local update rules (cf. [6, 7, 1 D. Denote the posterior by Pi(u) = P(ri = uIY"'), the Markovian property allows us to factor Pi(U) into three terms: one depending on the local data, another depending on data to the left of i and a third depending on data to the right of i. Thus: Pi(U) = cai(u)Li(u).Bi(u) (5) where ai(u) = P(Yi = uIYl~ i -l)' .Bi(U) = P(Yi = uIYi+l,N)' Li(U) = P(Yi"'lri = u) and c denotes a normalizing constant. Now, denoting the conditional Ci(u, v) = 910 y. Weiss P(Yi = ulYi-l = v), O'i(U) can be written in terms of O'i-l(V): O'j(u) = C 1 O'i-l(V)Ci(u, v)Li-l(V) (6) where c denotes another normalizing constant. A symmetric equation can be written for f3i( u). This suggests a propagation scheme where units represent the probabilities given in the left hand side of equations 5-6 and updates are based on the right hand side, i.e. on the activities of neighboring units. Specifically, for a Gaussian generating process the probabilities can be represented by their mean and variance. Thus denote Pi I'V N(pi' Ui), and similarly O'j """ N(pf, un and f3i I'V N(p~, uf). Performing the integration in 6 gives a Kalman-Filter like update for the parameters: Pi +.!!!..... y.. + ..LIL~ + 1 ILl? aD' ai rl -;;trrl . _1_ + W._l O~_l UD ( 1 Wi-l )-1 +UR + -a- + -Ui _ l UD (the update rules for the parameters of f3 are analogous) (7) (8) (9) So far we have considered continuous estimation problems but identical issues arise in labeling problems, where the task is to estimate a label L" which can take on M discrete values. We will denote L,,(m) = 1 if the label takes on value m and zero otherwise. Typically one minimizes functionals of the form: J(L) = L L V,,(m)L,,(m) - ALL L,,(m)Lk+l(m) (10) m " m " Traditional relaxation labeling algorithms minimize this cost functional with updates of the form: (11) Again different relaxation labeling algorithms differ in their choice of f. A linear sum followed by a threshold gives the discrete Hopfield network updates, a linear sum followed by a "soft" threshold gives the continuous or mean-field Hopfield up dates and yet another form gives the relaxation labeling algorithm of Rosenfeld et al. (see [3] for a review of relaxation labeling methods ). To derive a BBP algorithm for this case one can again rewrite J as the posterior of a Markov generating process, and calculate P(L,,(m) = 1) for this process.1. This gives the same expressions as in equations 5-6 with the integral replaced by a linear sum. Since the probabilities here are not Gaussian, the O'i, f3i, Pi will not be represented by their mean and variances, but rather by a vector of length M. Thus the update rule for O'i will be: (12) (and similarly for f3.) IFor certain special cases, knowing P(Lk(m) = 1) is not sufficient for choosing the sequence of labels that minimizes J. In those cases one should do belief revision rather than propagation [6] Interpreting I11Ulges by Propagating Bayesian Beliefs 911 a. b. Figure 2: R. the first frame of a sequence. The hand is translated to the left. b. contour extracted using standard methods 1.1 Convergence Equations 5-6 are mathematical identities. Hence, it is possible to show [6] that after N iterations the activity of units Pi will converge to the correct posteriors, where N is the maximal distance between any two units in the architecture, and an iteration refers to one update of all units. Furthermore, we have been able to show that after n < N iterations, the activity of unit Pi is guaranteed to represent the probability of the hidden state at location i given all data within distance n. This guarantee is significant in the light of a distinction made by Marr (1982) regarding local propagation rules. In a scheme where units only communicate with their neighbors, there is an obvious limit on how fast the information can reach a given unit: i.e. after n iterations the unit can only know about information within distance n. Thus there is a minimal number of iterations required for all data to reach all units. Marr distinguished between two types of iterations - those that are needed to allow the information to reach the units, versus those that are used to refine an estimate based on information that has already arrived. The significance of the guarantee on Pi is that it shows that BBP only uses the first type of iteration - iterations are used only to allow more information to reach the units. Once the information has arrived, Pi represents the correct posterior given that information and no further iterations are needed to refine the estimate. Moreover, we have been able to show that propagations schemes that do not propagate probabilities (such as those in equations 2) will in general not represent the optimal estimate given information that has already arrived. To summarize, both traditional relaxation updates as in equation 2 and BBP updates as in equations 7-9 give simple rules for updating a unit's activity based on local data and activities of neighboring units. However, the fact that BBP updates are based on the probability calculus guarantees that a unit's activity will be optimal given information that has already arrived and gives rise to a qualitative difference between the convergence of these two types of schemes. In the next section, we will demonstrate this difference in image interpretation problems. 2 Results Figure 2a shows the first frame of a sequence in which the hand is translated to the left. Figure 2b shows the bounding contour of the hand extracted using standard techniques. 2.1 Motion propagation along contours Local measurements along the contour are insufficient to determine the motion. Hildreth [2] suggested to overcome the local ambiguity by minimizing the following 912 , a. c. , ... , : . . . . . . . " 0 • • , - .:. '... ... . , b. d. .'~" ~:. ~ .-.-. ------... - - •.. --..••........ " . . - ----- ~ -- -- -- - - _":. -- -- -- - -- y. Weiss Figure 3: R. Local estimate of velocity along the contour. h. Performance of SOR, gradient descent and BBP as a function of time. BBP converges orders of magnitude faster than SOR. c. Motion estimate of SOR after 500 iterations. d. Motion estimate of BBP after 3 iterations. cost functional: J(V) = I)dx~v1: + dt1:)2 +..\ L IlvUl v1:11 2 (13) 1: 1: where dx, dt denote the spatial and temporal image derivatives and V1: denotes the velocity at point k along the contour. This functional is analogous to the interpolation functional (eq. 1) and the derivation of the relaxation and BBP updates are also analogous. Figure 3a shows the estimate of motion based solely on local information. The estimates are wrong due to the aperture problem. Figure 3b shows the performance of three propagation schemes: gradient descent, SOR and BBP. Gradient descent converges so slowly that the improvement in its estimate can not be discerned in the plot. SOR converges much faster than gradient descent but still has significant error after 500 iterations. BBP gets the correct estimate after 3 iterations! (Here and in all subsequent plots an iteration refers to one update of all units in the network). This is due to the fact that after 3 iterations, the estimate at location k is the optimal one given data in the interval [k - 3, k + 3]. In this case, there is enough data in every such interval along the contour to correctly estimate the motion. Figure 3c shows the estimate produced by SOR after 500 iterations. Even with simple visual inspection it is evident that the estimate is quite wrong. Figure 3d shows the (correct) estimate produced by BBP after 3 iterations. 2.2 Direction of figure propagation The extracted contour in figure 2 bounds a dark and a light region. Direction of figure (DOF) (e.g. [9]) refers to which of these two regions is figure and which is ground. A local cue for DOF is convexity - given three neighboring points along the contour we prefer the DOF that makes the angle defined by those points acute Interpreting Images by Propagating Bayesian Beliefs a. b. c. d. .. ~ - . _ .. - -- - - _.--_. --.- --- -_.- - --4Il : \ " " " 20 " I ", " I, o ~ - ~ -- -- - - - -- --- --- o 10 100 160 :lOD 250)00 ~ - ----- . ANII-Or--.,.. 9/3 Figure 4: a. Local estimate of DOF along the contour. b. Performance of Hopfield,gradient descent, relaxation labeling and BBP as a function of time. BBP is the only method that converges to the global minimum. c. DOF estimate of Hopfield net after convergence. d. DOF estimate of BBP after convergence. rather than obtuse. Figure 4a shows the results of using this local cue on the hand contour. The local cue is not sufficient. We can overcome the local ambiguity by minimizing a cost functional that takes into account the DOF at neighboring points in addition to the local convexity. Denote by Lk(m) the DOF at point k along the contour and define (14) m k m k with Vk(m) determined by the acuteness of the angle at location k. Figure 4b shows the performance of four propagation algorithms on this task: three traditional relaxation labeling algorithms (MF Hopfield, Rosenfeld et aI, constrained gradient descent) and BBP. All three traditional algorithms converge to a local minimum, while the BBP converges to the global minimum. Figure 4c shows the local minimum reached by the Hopfield network and figure 4d shows the correct solution reached by the BBP algorithm. Recall (section 1.1) that BBP is guaranteed to converge to the correct posterior given all the data. 2.3 Extensions to 2D In the previous two examples ambiguity was reduced by combining information from other points on the same contour. There exist, however, cases when information should be propagated to all points in the image. Unfortunately, such propagation problems correspond to Markov Random Field (MRF) generative models, for which calculation of the posterior cannot be done efficiently. However, Willsky and his 914 y. Weiss colleagues [4] have recently shown that MRFs can be approximated with hierarchical or multi-resolution models. In current work, we have been using the multi-resolution generative model to derive local BBP rules. In this case, the Bayesian beliefs are propagated between neighboring units in a pyramidal representation of the image. Although this work is still in preliminary stages, we find encouraging results in comparison with traditional 2D relaxation schemes. 3 Discussion The update rules in equations 5-6 differ slightly from those derived by Pearl [6] in that the quantities a, {3 are conditional probabilities and hence are constantly normalized to sum to unity. Using Pearl's original algorithm for sequences as long as the ones we are considering will lead to messages that become vanishingly small. Likewise our update rules differ slightly from the forward-backward algorithm for HMMs [7] in that ours are based on the assumption that all states are equally likely apriore and hence the updates are symmetric in a and {3. Finally, equation 9 can be seen as a variant of a Riccati equation [1]. In addition to these minor notational differences, the context in which we use the update rules is different. While in HMMs and Kalman Filters, the updates are seen as interim calculations toward calculating the posterior, we use these updates in a parallel network of local units and are interested in how the estimates of units in this network improve as a function of iteration. We have shown that an architecture that propagates Bayesian beliefs according to the probability calculus yields orders of magnitude improvements in convergence over traditional schemes that do not propagate probabilities. Thus image interpretation provides an important example of a task where it pays to be a Bayesian. Acknowledgments I thank E. Adelson, P. Dayan, J. Tenenbaum and G. Galperin for comments on versions ofthis manuscript; M.1. Jordan for stimulating discussions and for introducing me to Bayesian nets. Supported by a training grant from NIGMS. References [1] Arthur Gelb, editor. Applied Optimal Estimation. MIT Press, 1974. [2] E. C. Hildreth. The Measurement of Visual Motion. MIT Press, 1983. [3] S.Z. Li. Markov Random Field Modeling in Computer Vision. Springer-Verlag, 1995. [4] Mark R. Luettgen, W. Clem Karl, and Allan S. Willsky. Efficient multiscale regularization with application to the computation of optical flow. IEEE Transactions on image processing, 3(1):41-64, 1994. [5] D. Marr. Vision. H. Freeman and Co., 1982. [6] Judea Pearl. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann, 1988. [7] Lawrence Rabiner and Biing-Hwang Juang. Fundamentals of Speech recognition. PTR Prentice Hall, 1993. [8] A. Rosenfeld, R. Hummel, and S. Zucker. Scene labeling by relaxation operations. IEEE Transactions on Systems, Man and Cybernetics, 6:420-433, 1976. [9] P. Sajda and L. H. Finkel. Intermediate-level visual representations and the construction of surface perception. Journal of Cognitive Neuroscience, 1994. [10] Gilbert Strang. Introduction to Applied Mathematics. Wellesley-Cambridge, 1986.
1996
44
1,241
Why did TD-Gammon Work? Jordan B. Pollack & Alan D. Blair Computer Science Department Brandeis University Waltham, MA 02254 {pollack,blair} @cs.brandeis.edu Abstract Although TD-Gammon is one of the major successes in machine learning, it has not led to similar impressive breakthroughs in temporal difference learning for other applications or even other games. We were able to replicate some of the success of TD-Gammon, developing a competitive evaluation function on a 4000 parameter feed-forward neural network, without using back-propagation, reinforcement or temporal difference learning methods. Instead we apply simple hill-climbing in a relative fitness environment. These results and further analysis suggest that the surprising success of Tesauro's program had more to do with the co-evolutionary structure of the learning task and the dynamics of the backgammon game itself. 1 INTRODUCTION It took great chutzpah for Gerald Tesauro to start wasting computer cycles on temporal difference learning in the game of Backgammon (Tesauro, 1992). Letting a machine learning program play itself in the hopes of becoming an expert, indeed! After all, the dream of computers mastering a domain by self-play or "introspection" had been around since the early days of AI, forming part of Samuel's checker player (Samuel, 1959) and used in Donald Michie's MENACE tic-tac-toe learner (Michie, 1961). However such self-conditioning systems, with weak or non-existent internal representations, had generally been fraught with problems of scale and abandoned by the field of AI. Moreover, self-playing learners usually develop eccentric and brittle strategies which allow them to draw each other, yet play poorly against humans and other programs. Yet Tesauro's 1992 result showed that this self-play approach could be powerful, and after some refinement and millions of iterations of self-play, his TD-Gammon program has become one of the best backgammon players in the world (Tesauro, 1995). His derived weights are viewed by his corporation as significant enough intellectual property to keep as a trade secret, except to leverage sales of their minority operating system (International Business Machines, 1995). Others have replicated this TD result both for research purposes (Boyan, 1992) and commercial purposes. Why did TD-Gammon Work? 11 With respect to the goal of a self-organizing learning machine which starts from a minimal specification and rises to great sophistication, TD-Gammon stands alone. How is its success to be understood, explained, and replicated in other domains? Is TD-Gammon unbridled good news about the reinforcement learning method? Our hypothesis is that the success ofTD-gammon is not due to the back-propagation, reinforcement, or temporal-difference technologies, but to an inherent bias from the dynamics of the game of backgammon, and the co-evolutionary setup of the training, by which the task dynamically changes as the learning progresses. We test this hypothesis by using a much simpler co-evolutionary learning method for backgammon - namely hill-climbing. 2 SETUP We use a standard feedforward neural network with two layers and the sigmoid function, set up in the same fashion as Tesauro with 4 units to represent the number of each player's pieces on each of the 24 points, plus 2 units each to indicate how many are on the bar and off the board. In addition, we added one more unit which reports whether or not the game has reached the endgame or "race" situation, making a total of 197 input units. These are fully connected to 20 hidden units, which are then connected to one output unit that judges the position. Including bias on the hidden units, this makes a total of 3980 weights. The game is played by generating all legal moves, converting them into the proper network input, and picking the position judged as best by the network. We started with all weights set to zero. Our initial algorithm was hillclimbing: 1. add gaussian noise to the weights 2. play the network against the mutant for a number of games 3. if the mutant wins more than half the games, select it for the next generation. The noise was set so each step would have a 0.05 RMS distance (which is the euclidean distance divided by J3980). Surprisingly, this worked reasonably well! The networks so evolved improved rapidly at first, but then sank into mediocrity. The problem we perceived is that comparing two close backgammon players is like tossing a biased coin repeatedly: it may take dozens or even hundreds of games to find out for sure which of them is better. Replacing a well-tested champion is dangerous without enough information to prove the challenger is really a better player and not just a lucky novice. Rather than burden the system with so much computation, we instead introduced the following modifications to the algorithm to avoid this "Buster Douglas Effect": Firstly, the games are played in pairs, with the order of play reversed and the same random seed used to generate the dice rolls for both games. This washes out some of the unfairness due to the dice rolls when the two networks are very close - in particular, if they were identical, the result would always be one win each. Secondly, when the challenger wins the contest, rather than just replacing the champion by the challenger, we instead make only a small adjustment in that direction: champion := 0.95*champion + 0.05*challenger This idea, similar to the "inertia" term in back-propagation, was introduced on the assumption that small changes in weights would lead to small changes in decision-making by the evaluation function. So, by preserving most of the current champion's decisions, we would be less likely to have a catastrophic replacement of the champion by a lucky novice challenger. In the initial stages of evolution, two pairs of parallel games were played and the challenger was required to win 3 out of 4 of these games. 12 1. B. Pollack and A. D. Blair ~ o 50 ~ 25 Generation (x 103) 5 10 15 20 25 30 35 Figure 1. Percentage of losses of our first 35,000 generation players against PUBEVAL. Each match consisted of 200 games. Figure 1 shows the first 35,000 players rated against PUBEVAL, a strong public-domain player trained by Tesauro using human expert preferences. There are three things to note: (1) the percentage of losses against PUBEVAL falls from 100% to about 67% by 20,000 generations, (2) the frequency of successful challengers increases over time as the player improves, and (3) there are epochs (e.g. starting at 20,000) where the performance against PUBEVAL begins to falter. The first fact shows that our simple self-playing hill-climber is capable of learning. The second fact is quite counter-intuitive - we expected that as the player improved, it would be harder to challenge it! This is true with respect to a uniform sampling of the 4000 dimensional weight space, but not true for a sampling in the neighborhood of a given player: once the player is in a good part of weight space, small changes in weights can lead to mostly similar strategies, ones which make mostly the same moves in the same situations. However, because of the few games we were using to determine relative fitness, this increased frequency of change allows the system to drift, which may account for the subsequent degrading of performance. To counteract the drift, we decided to change the rules of engagement as the evolution proceeds according to the following "annealing schedule": after 10,000 generations, the number of games that the challenger is required to win was increased from 3 out of 4 to 5 out of 6; after 70,000 generations, it was further increased to 7 out of 8. The numbers 10,000 and 70,000 were chosen on an ad hoc basis from observing the frequency of successful challenges. After 100,000 games, we have developed a surprisingly strong player, capable of winning 40% of the games against PUBEV AL. The networks were sampled every 100 generations in order to test their performance. Networks at generation 1,000, 10,000 and 100,000 were extracted and used as benchmarks. Figure 2 shows the percentage of losses of the sampled players against the three benchmark networks. Note that the three curves cross the 50% line at 1, 10, and 100, respectively and show a general improvement over time. The end-game of backgammon, called the "bear-off," can be used as another yardstick of the progress of learning. The bear-off occurs when all of a player's pieces are in the player's home, or first 6 points, and then the dice rolls can be used to remove pieces. We set up a racing board with two pieces on each player's 1 through 7 point and one piece on the 8 point, and played a player against itself 200 games, averaging the number of rolls. We found a monotonic improvement, from 22 to less then 19 rolls, over the lOOk generations. PUBEV AL scored 16.6 on this task. Why did TD-Gammon Work? l00.-------r-----~~----~------~------~ 20 40 60 80 100 Generation (x 103) Figure 2. Percentage of losses against benchmark networks at generation 1,000 [lower), 10,000 [middle) and 100,000 [upper). 3 DISCUSSION 3.1 Machine Learning and Evolution 13 We believe that our evidence of success in learning backgammon using simple hillclimbing indicates that the reinforcement and temporal difference methodology used by Tesauro in TO-gammon was non-essential for its success. Rather, the success came from the setup of co-evolutionary self-play biased by the dynamics of backgammon. Our result is thus similar to the bias found by Mitchell, Crutchfield & Graber in Packard's evolution of cellular automata to the "edge of chaos"(Packard, 1988, Mitchell et aI., 1993). TO-Gammon is a major milestone for a kind of evolutionary machine learning in which the initial specification of model is far simpler than expected because the learning environment is specified implicitly, and emerges as a result of the co-evolution between a learning system and its training environment: The learner is embedded in an environment which responds to its own improvements in a never-ending spiral. While this effect has been seen in population models, it is completely unexpected for a "1 + 1" hillclimbing evolution. Co-evolution was explored by Hillis (Hillis, 1992) on the sorting problem, by Angeline & Pollack (Angeline and Pollack, 1994) on genetically programmed tic-tac-toe players, on predator/prey games, e.g. (Cliff and Miller, 1995, Reynolds, 1994), and by Juille & Pollack on the intertwined spirals problem (Juille and Pollack, 1995). Rosin & Belew applied competitive fitness to several games (Rosin and Belew, 1995). However, besides Tesauro's TO-Gammon, which has not to date been viewed as an instance of co-evolutionary learning, Sims' artificial robot game (Sims, 1994) is the only other domain as complex as Backgammon to have had substantial success. 3.2 Leamability and Unleamability Learnability can be formally defined as a time constraint over a search space. How hard is it to randomly pick 4000 floating-point weights to make a good backgammon evaluator? It is simply impossible. How hard is it to find weights better than the current set? Initially, when all weights are random, it is quite easy. As the playing improves, we would expect it to get harder and harder, perhaps similar to the probability of a tornado constructing a 747 out of a junkyard. However, if we search in the neighborhood of the current weights, we will find many players which make mostly the same moves but which can capitalize on each other's slightly different choices and exposed weaknesses in a tournament. 14 J B. Pollack and A. D. Blair Although the setting of parameters in our initial runs involved some guesswork, now that we have a large set of "players" to examine, we can try to understand the phenomenon. Taking the 1000th, 1O,000th, and loo,OOOth champions from our run, we sampled random players in their neighborhoods at different RMS distances to find out how likely is it to find a winning challenger. We took 1000 random neighbors at each of 11 different RMS distances, and played them 8 games against the corresponding champion. Figure 3 plots l00r---------------~----------------_, ~ 75 f u 50~~ ... =_:_::~--__ "' :,:,::. :~ ..... .... .. .......... 10k """' ,:,:,:::::~,::,::, ",. RMS distance from champion O~--------------~----------------~ o 0.05 0.1 Figure 3. Distance versus probability of random challenger winning against champions at generation 1,000, 10,000 and 100,000. the average number of games won against the three champions in the range of neighborhoods. This graph demonstrates that as the players improve over time, the probability of finding good challengers in their neighborhood increases. This accounts for why the frequency of successful challenges goes up. Each successive challenger is only required to take the small step of changing a few moves of the champion in order to beat it. Therefore, under co-evolution what was apparently unlearnable becomes learnable as we convert from a single question to a continuous stream of questions, each one dependent on the previous answer. 3.3 Avoiding Mediocre Stable States In general, the problem with learning through self-play is that the player could keep playing the same kinds of games over and over, only exploring some narrow region of the strategy space, missing out on critical areas of the game where it could then be vulnerable to other programs or human experts. Such a learning system might declare success when in reality it has simply converged to a "mediocre stable state" of continual draws or a long term cooperation which merely mimics competition. Such a state can arise in human education systems, where the student gets all the answers right and rewards the teacher with positive feedback for not asking harder questions. The problem is particularly prevalent in self-play for deterministic games such as chess or tic-tac-toe. We have worked on using a population to get around it (Angeline and Pollack, 1994). Schraudolph et al., 1994 added non-determinism to the game of Go by choosing moves according to the Boltzmann distribution of statistical mechanics. Others, such as Fogel, 1993, expanded exploration by forcing initial moves. Epstein, 1994, has studied a mix of training using self-play, random testing, and playing against an expert in order to better understand this phenomenon. We are not suggesting that 1 + 1 hillclimbing is an advanced machine learning technique which others should bring to many tasks. Without internal cognition about an opponent's behavior, co-evolution usually requires a population. Therefore, there must be something about the dynamics of backgammon itself which is helpful because it permitted both TD Why did TD-Gammon Work? 15 learning and hill-climbing to succeed where they would clearly fail on other tasks and in other games of this scale. If we can understand why the backgammon domain led to successful acquisition of expert strategies from random initial conditions, we might be able to re-cast other domains in its image. Tesauro, 1992 pointed out some of the features of Backgammon that make it suitable for approaches involving self-play and random initial conditions. Unlike chess, a draw is impossible and a game played by an untrained network making random moves will eventually terminate (though it may take much longer than a game between competent players). Moreover the randomness of the dice rolls leads self-play into a much larger part of the search space than it would be likely to explore in a deterministic game. We believe it is not simply the dice rolls which overcome the problems of self-learning. Others have tried to add randomness to deterministic games and have not generally met with success. There is something critical about the dynamics of backgammon which sets its apart from other games with random elements like Monopoly. Namely, that the outcome of the game continues to be uncertain until all contact is broken and one side has a clear advantage. What many observers find exciting about backgammon, and what helps a novice sometimes overcome an expert, is the number of situations where one dice roll, or an improbable sequence, can dramatically reverse which player is expected to win. A learning system can be viewed as a meta-game between teacher and student, which are identical in a self-play situation. The teacher's goal is to expose the student's mistakes, while the student's goal is to placate the teacher and avoid such exposure. A mediocre stable state for a self-learning system can be seen as an equilibrium situation in this metagame. A player which learns to repeatedly draw itself will have found a meta-game equilibrium and stop learning. If draws are not allowed, it may still be possible for a self-playing learner to collude with itself - to simulate competition while actually cooperating (Angeline, 1994). For example, if slightly suboptimal moves would allow a player to "throw" a game, a player under self-play could find a meta-game equilibrium by alternately throwing games to itself! Our hypothesis is that the dynamics of backgammon discussed above actively prevent this sort of collusion from forming in the meta-game of self-learning. 4 CONCLUSIONS Tesauro's 1992 result beat Sun's Gammontool and achieved parity against his own Neurogammon 1.0, trained on expert knowledge. Neither of these is available. Following the 1992 paper on TO-learning, he incorporated a number of hand-crafted expert-knowledge features, eventually producing a network which achieved world master level play (Tesauro, 1995). These features included concepts like existence of a prime, probability of blots being hit, and probability of escape from behind the opponent's barrier. Our best players win about 45% against PUBEVAL which was trained using "comparison training"(Tesauro, 1989). Therefore we believe our players achieve approximately the same power as Tesauro's 1992 results, without any advanced learning algorithms. We do not claim that our 100,000 generation player is as good as TO-Gammon, ready to challenge the best humans, just that it is surprisingly good considering its humble origins from hill-climbing with a relative fitness measure. Tuning our parameters or adding more input features would make more powerful players, but that is not the point of this study. TO-Gammon remains a tremendous success in Machine Learning, but the causes for its success have not been well understood. Replicating some ofTO-Gammon's success under a much simpler learning paradigm, we find that the primary cause for success must be the dynamics of backgammon combined with the power of co-evolutionary learning. If we can isolate the features of the backgammon domain which enable evolutionary learning to work so well, it may lead to a better understanding of the conditions necessary, in general, for complex self-organization. 16 1. B. Pollack and A. D. Blair Acknowledgments This work is supported by ONR grant NOOOI4-96-1-0418 and a Krasnow Foundation Postdoctoral fellowship. Thanks to Gerry Tesauro for providing PUBEVAL and subsequent means to calibrate it, Jack Laurence and Pablo Funes for development of the WWW front end to our evolved player. Interested players can challenge our evolved network using a web browser through our home page at: http://www.demo.cs.brandeis.edu References Angeline, P. J. (1994). An alternate interpretation of the iterated prisoner's dilemma and the evolution of non-mutual cooperation. In Brooks, R. and Maes, P., editors, Proceedings 4th Artificial Life Conference, pages 353-358. MIT Press. Angeline, P. J. and Pollack, J. B. (1994). Competitive environments evolve better solutions for complex tasks. In Forrest, S., editor, Genetic Algorithms: Proceedings of the Fifth Inter national Conference. Boyan, J. A. (1992). Modular neural networks for learning context-dependent game strategies. Master's thesis, Computer Speech and Language Processing, Cambridge University. Cliff, D. and Miller, G. (1995). Tracking the red queen: Measurements of adaptive progress in co-evolutionary simulations. In Third European Conference on Artificial Life, pages 200-218. Hillis, D. (1992). Co-evolving parasites improves simulated evolution as an optimization procedure. In C. Langton, C. Taylor, J. E and Rasmussen, S., editors, Artificial Life II. Addison-Wesley, Reading, MA. International Business Machines (Sept. 12, 1995). IBM's family funpak for osl2 warp hits retail shelves. Juille, H. and Pollack, J. (1995). Massively parallel genetic programming. In Angeline, P. and Kinnear, K., editors, Advances in Genetic Programming II. MIT Press, Cambridge. Michie, D. (1961). Trial and error. In Science Survey, part 2, pages 129-145. Penguin. Mitchell, M., Hraber, P. T., and Crutchfield, J. P. (1993). Revisiting the edge of chaos: Evolving cellular automata to perform computations. Complex Systems, 7. Packard, N. (1988). Adaptation towards the edge of chaos. In Kelso, J. A. S., Mandell, A. J., and Shlesinger, M. E, editors, Dynamic patterns in complex systems, pages 293301. World Scientific. Reynolds, C. (1994). Competition, coevolution, and the game of tag. In Proceedings 4th Artificial Life Conference. MIT Press. Rosin, C. D. and Belew, R. K. (1995). Methods for competitive co-evolution: finding opponents worth beating. In Proceedings of the 6th international conference on Genetic Algorithms, pages 373-380. Morgan Kaufman. Samuel, A. L. (1959). some studies of machine learning using the game of checkers. IBM ]oural of Research and Development. Sims, K. (1994). Evolving 3d morphology and behavior by competition. In Brooks, R. and Maes, P., editors, Proceedings 4th Artificial Life Conference. MIT Press. Tesauro, G. (1989). Connectionist learning of expert preferences by comparison training. In Touretzky, D., editor, Advances in Neural Information Processing Systems, volume 1, pages 99-106, Denver 1988. Morgan Kaufmann, San Mateo. Tesauro, G. (1992). Practical issues in temporal difference learning. Machine Learning, 8:257-277. Tesauro, G. (1995). Temporal difference learning and td-gammon. Communications of the ACM, 38(3):58-68.
1996
45
1,242
NeuroScale: Novel Topographic Feature Extraction using RBF Networks David Lowe D.LoweOaston.ac.uk Michael E. Tipping H.E.TippingOaston.ac.uk Neural Computing Research Group Aston University, Aston Triangle, Birmingam B4 7ET1 UK http://www.ncrg.aston.ac.uk/ . Abstract Dimension-reducing feature extraction neural network techniques which also preserve neighbourhood relationships in data have traditionally been the exclusive domain of Kohonen self organising maps. Recently, we introduced a novel dimension-reducing feature extraction process, which is also topographic, based upon a Radial Basis Function architecture. It has been observed that the generalisation performance of the system is broadly insensitive to model order complexity and other smoothing factors such as the kernel widths, contrary to intuition derived from supervised neural network models. In this paper we provide an effective demonstration of this property and give a theoretical justification for the apparent 'self-regularising' behaviour of the 'NEUROSCALE' architecture. 1 'NeuroScale': A Feed-forward Neural Network Topographic Transformation Recently an important class of topographic neural network based feature extraction approaches, which can be related to the traditional statistical methods of Sammon Mappings (Sammon, 1969) and Multidimensional Scaling (Kruskal, 1964), have been introduced (Mao and Jain, 1995; Lowe, 1993; Webb, 1995; Lowe and Tipping, 1996). These novel alternatives to Kohonen-like approaches for topographic feature extraction possess several interesting properties. For instance, the NEuROSCALE architecture has the empirically observed property that the generalisation perfor544 D. Lowe and M. E. Tipping mance does not seem to depend critically on model order complexity, contrary to intuition based upon knowledge of its supervised counterparts. This paper presents evidence for their 'self-regularising' behaviour and provides an explanation in terms of the curvature of the trained models. We now provide a brief introduction to the NEUROSCALE philosophy of nonlinear topographic feature extraction. Further details may be found in (Lowe, 1993; Lowe and Tipping, 1996). We seek a dimension-reducing, topographic transformation of data for the purposes of visualisation and analysis. By 'topographic', we imply that the geometric structure of the data be optimally preserved in the transformation, and the embodiment of this constraint is that the inter-point distances in the feature space should correspond as closely as possible to those distances in the data space. The implementation of this principle by a neural network is very simple. A Radial Basis Function (RBF) neural network is utilised to predict the coordinates of the data point in the transformed feature space. The locations of the feature points are indirectly determined by adjusting the weights of the network. The transformation is determined by optimising the network parameters in order to minimise a suitable error measure that embodies the topographic principle. The specific details of this alternative approach are as follows. Given an mdimensional input space of N data points x q , an n-dimensional feature space of points Yq is generated such that the relative positions of the feature space points minimise the error, or 'STRESS', term: N E = 2: 2:(d~p - dqp)2, (1) p q>p where the d~p are the inter-point Euclidean distances in the data space: d~p = J(xq - Xp)T(Xq - xp), and the dqp are the corresponding distances in the feature space: dqp = J(Yq - Yp)T(Yq - Yp)· The points yare generated by the RBF, given the data points as input. That is, Yq = f(xq;W), where f is the nonlinear transformation effected by the RBF with parameters (weights and any kernel smoothing factors) W. The distances in the feature space may thus be given by dqp =11 f(xq) - f(xp) " and so more explicitly by (2) where ¢k 0 are the basis functions, JLk are the centres of those functions, which are fixed, and Wlk are the weights from the basis functions to the output. The topographic nature of the transformation is imposed by the STRESS term which attempts to match the inter-point Euclidean distances in the feature space with those in the input space. This mapping is relatively supervised because there is no specific target for each Y q; only a relative measure of target separation between each Yq, Yp pair is provided. In this form it does not take account of any additional information (for example, class labels) that might be associated with the data points, but is determined strictly by their spatial distribution. However, the approach may be extended to incorporate the use of extra 'subjective' information which may be NeuroScale: Novel Topographic Feature Extraction using RBF Networks 545 used to influence the transformation and permits the extraction of 'enhanced', more informative, feature spaces (Lowe and Tipping, 1996). Combining equations (1) and (2) and differentiating with respect to the weights in the network allows the partial derivatives of the STRESS {)E/{)WZk to be derived for each pattern pair. These may be accumulated over the entire pattern set and the weights adjusted by an iterative procedure to minimise the STRESS term E. Note that the objective function for the RBF is no longer quadratic, and so a standard analytic matrix-inversion method for fixing the final layer weights cannot be employed. We refer to this overall procedure as 'NEUROSCALE'. Although any universal approximator may be exploited within NEUROSCALE, using a Radial Basis Function network allows more theoretical analysis of the resulting behaviour, despite the fact that we have lost the usual linearity advantages of the RBF because of the STRESS measure. A schematic ofthe NEUROSCALE model is given in figure 1, and illustrates the role of the RBF in transforming the data space to the feature space. ERROR MEASURE ~j FEATURE SPACE DATA SPACE RBF Figure 1: The NEUROSCALE architecture. 2 Generalisation In a supervised learning context, generalisation performance deteriorates for overcomplex networks as 'overfitting' occurs. By contrast, it is an interesting empirical observation that the generalisation performance of NEUROSCALE, and related models, is largely insensitive to excessive model complexity. This applies both to the number of centres used in the RBF and in the kernel smoothing factors which themselves may be viewed as regularising hyperparameters in a feed-forward supervised situation. This insensitivity may be illustrated by Figure 2, which shows the training and test set performances on the IRIS data (for 5-45 basis functions trained and tested on 45 separate samples). To within acceptable deviations, the training and test set 546 D. Lowe and M. E. Tipping STRESS values are approximately constant. This behaviour is counter-intuitive when compared with research on feed forward networks trained according to supervised approaches. We have observed this general trend on a variety of diverse real world problems, and it is not peculiar to the IRIS data. 4.5 X 10-' Training and Test Errors 4 3.5 3 0.5 o o 5 10 15 20 25 30 35 40 45 Number of Basis Functions Figure 2: Training and test errors for NEUROSCALE Radial Basis Functions with various numbers of basis functions. Training errors are on the left, test errors are on the right. There are two fundamental causes of this observed behaviour. Firstly, we may derive significant insight into the necessary form of the functional transformation independent of the data. Secondly, given this prior functional knowledge, there is an appropriate regularising component implicitly incorporated in the training algorithm outlined in the previous section. 2.1 Smoothness and Topographic Transformations For a supervised problem, in the absence of any explicit prior information, the smoothness of the network function must be determined by the data, typically necessitating the setting of regularising hyperparameters to counter overfitting behaviour. In the case of the distance-preserving transformation effected by NEUROSCALE, an understanding of the necessary smoothness may be deduced a priori. Consider a point Xq in input space and a nearby test point xp = Xq + Epq , where Epq is an arbitrary displacement vector. Optimum generalisation demands that the distance between the corresponding image points Yq and Yp should thus be I/Epq 1/. Considering the Taylor expansions around the point Y q we find n Ilyp - Yq 112 = 2)E;qgq,)2 + O(E4), 1=1 = E;q (t gqlg~) Epq + O(E4), 1=1 (3) = E;qGqEpq + O(E4), NeuroScale: Novel Topographic Feature Extraction using RBF Networks 547 where the matrix Gq L~=l gqlg~ and gql is the gradient vector (8YI(q)/f)xl, ... ,8YI(q)/8xn )T evaluated at x = x q . For structure preservation the corresponding distances in input and output spaces need to be retained for all values of €opq: IIYp - Yq 112= €oT €o, and so G q = I with the requirement that secondand higher-order terms must vanish. In particular note that measures of curvature proportional to (82YI(Q)/8xn2 should vanish. In general, for dimension reduction, we cannot ensure that exact structure preservation is obtained since the rank of G q is necessarily less than n and hence can never equate to the identity matrix. However, when minimising STRESS we are locally attempting to minimise the residual II I - G q II, which is achieved when all the vectors €opq of interest lie within the range of G q• 2.2 The Training Mechanism An important feature of this class of topographic transformations is that the STRESS measure is invariant under arbitrary rotations and transformations of the output configuration. The algorithm outlined previously tends towards those configurations that generally reduce the sum-of-squared weight values (Tipping, 1996). This is achieved without any explicit addition of regularisation, but rather it is a feature of the relative supervision algorithm. The effect of this reduction in weight magnitudes on the smoothness of the network transformation may be observed by monitoring an explicit quantitative measure of total curvature: (4) where Q ranges over the patterns, i over the input dimensions and lover the output dimensions. Figure 3 depicts the total curvature of NEUROSCALE as a function of the training iterations on the IRIS subset data for a variety of model complexities. As predicted, curvature generally decreases during the training process, with the final value independent of the model complexity. Theoretical insight into this phenomenon is given in (Tipping, 1996). This behaviour is highly relevant, given the analysis of the previous subsection. That the training algorithm implicitly reduces the sum-of-squares weight values implies that there is a weight decay process occurring with an associated smoothing effect. While there is no control over the magnitude of this element, it was shown that for good generalisation, the optimal transformation should be maximally smooth. This self-regularisation operates differently to regularisers normally introduced to stabilise the ill-posed problems of supervised neural network models. In the latter case the regulariser acts to oppose the effect of reducing the error on the training set. In NEUROSCALE the implicit weight decay operates with the minimisation of STRESS since the aim is to 'fit' the relative input positions exactly. That there are many RBF networks which satisfy a given STRESS level may be seen by training a network a posteriori on a predetermined Sammon mapping of a data set by a supervised approach (since then the targets are known explicitly). In general, such a posteriori trained networks do not have a low curvature and hence 548 D. Lowe and M. E. Tipping Curvature against Time for NeuroScale 2~~ . -----------.------------~----------, 2000 I o I ~ 1500 I :l '" ::; -- 15 Basis Functions - 30 Basis Functions .. 45 Basis Functions ~L-----------~50------------1~OO----------~150 Epoch Number Figure 3: Curvature against time during the training of a NEUROSCALE mapping on the Iris data, for networks with 15, 30 and 45 basis functions. do not show as good a generalisation behaviour as networks trained according to the relative supervision approach. The method by which NEUROSCALE reduces curvature, is to select, automatically, RBF networks with minimum norm weights. This is an inherent property of the training algorithm to reduce the STRESS criterion. 2.3 An example An effective example of the ease of production of good generalising transformations is given by the following experiment. A synthetic data set comprised four Gaussian clusters, each with spherical variance of 0.5, located in four dimensions with centres at (xc, 0, 0, 0) : Xc E {I, 2, 3, 4}. A NEUROSCALE transformation to two dimensions was trained using the relative supervision approach, using the three clusters at Xc = 1,3 and 4. The network was then tested on the entire dataset, with the fourth cluster included, and the projections are given in Figure 4 below. The apparently excellent generalisation to test data not sampled from the same distribution as the training data is a function of the inherent smoothing within the training process and also reflects the fact that the test data lay approximately within the range of the matrices G q determined during training. 3 Conclusion We have described NEUROSCALE, a parameterised RBF Sammon mapping approach for topographic feature extraction. The NEUROSCALE method may be viewed as a technique which is closely related to Sammon mappings and nonlinear metric MDS, with the added flexibility of producing a generalising transformation. A theoretical justification has been provided for the empirical observation that the generalisation performance is not affected by model order complexity issues. This counter-intuitive result is based on arguments of necessary transformation smoothNeuroScale: Novel Topographic Feature Extraction using RBF Networks 549 NeuroSca1. trained on 3 linear clusters NeuroScaJe 'asled on 4 "near clulters 1,5 1.5 a 0 • O.S • a,. \. : ;: .. t:. ~ o .' : ... . . , ..... -O.S :- • -, -, 1.S 2S 3S 4S ,.S 2S 3S 4.S Figure 4: Training and test projections of the four clusters. Training STRESS was 0.00515 and test STRESS 0.00532. ness coupled with the apparent self-regularising aspects of NEUROSCALE. The relative supervision training algorithm implicitly minimises a measure of curvature by incorporating an automatic 'weight decay' effect which favours solutions generated by networks with small overall weights. Acknowledgements This work was supported in part under the EPSRC contract GR/J75425, "Novel Developments in Learning Theory for Neural Networks". References Kruskal, J. B. (1964). Multidimensional scaling by optimising goodness of fit to a nonmetric hypothesis. Psychometrika, 29(1}:1-27. Lowe, D. {1993}. Novel 'topographic' nonlinear feature extraction using radial basis functions for concentration coding in the 'artificial nose'. In 3rd lEE International Conference on Artificial Neural Networks. London: lEE. Lowe, D. and Tipping, M. E. {1996}. Feed-forward neural networks and topographic mappings for exploratory data analysis. Neural Computing and Applications, 4:83-95. Mao, J. and Jain, A. K. {1995}. Artificial neural networks for feature extraction and multivariate data projection. IEEE Transactions on Neural Networks, 6{2}:296-317. Sammon, J. W. {1969}. A nonlinear mapping for data structure analysis. IEEE Transactions on Computers, C-18{5}:401-409. Tipping, M. E. {1996}. Topographic Mappings and Feed-Forward Neural Networks. PhD thesis, Aston University, Aston Street, Birmingham B4 7ET, UK. Available from http://www.ncrg.aston.ac . uk/. Webb, A. R. (1995). Multidimensional scaling by iterative majorisation using radial basis functions. Pattern Recognition, 28{5}:753-759.
1996
46
1,243
Time Series Prediction Using Mixtures of Experts Ron Meir Assaf J. Zeevi Information Systems Lab Department of Electrical Engineering Stanford University Stanford, CA. 94305 azeevi~isl.stanford.edu Department of Electrical Engineering Technion Haifa 32000, Israel rmeir~ee.technion.ac.il Robert J. Adler Department of Statistics University of North Carolina Chapel Hill, NC. 27599 adler~stat.unc.edu Abstract We consider the problem of prediction of stationary time series, using the architecture known as mixtures of experts (MEM). Here we suggest a mixture which blends several autoregressive models. This study focuses on some theoretical foundations of the prediction problem in this context. More precisely, it is demonstrated that this model is a universal approximator, with respect to learning the unknown prediction function. This statement is strengthened as upper bounds on the mean squared error are established. Based on these results it is possible to compare the MEM to other families of models (e.g., neural networks and state dependent models). It is shown that a degenerate version of the MEM is in fact equivalent to a neural network, and the number of experts in the architecture plays a similar role to the number of hidden units in the latter model. 310 A. 1. Zeevi, R. Meir and R. 1. Adler 1 Introduction In this work we pursue a new family of models for time series, substantially extending, but strongly related to and based on the classic linear autoregressive moving average (ARMA) family. We wish to exploit the linear autoregressive technique in a manner that will enable a substantial increase in modeling power, in a framework which is non-linear and yet mathematically tractable. The novel model, whose main building blocks are linear AR models, deviates from linearity in the integration process, that is, the way these blocks are combined. This model was first formulated in the context of a regression problem, and an extension to a hierarchical structure was also given [2]. It was termed the mixture of experts model (MEM). Variants of this model have recently been used in prediction problems both in economics and engineering. Recently, some theoretical aspects of the MEM , in the context of non-linear regression, were studied by Zeevi et al. [8], and an equivalence to a class of neural network models has been noted. The purpose of this paper is to extend the previous work regarding the MEM in the context of regression, to the problem of prediction of time series. We shall demonstrate that the MEM is a universal approximator, and establish upper bounds on the approximation error, as well as the mean squared error, in the setting of estimation of the predictor function. It is shown that the MEM is intimately related to several existing, state of the art, statistical non-linear models encompassing Tong's TAR (threshold autoregressive) model [7], and a certain version of Priestley's [6] state dependent models (SDM). In addition, it is demonstrated that the MEM is equivalent (in a sense that will be made precise) to the class of feedforward, sigmoidal, neural networks. 2 Model Description The MEM [2] is an architecture composed of n expert networks, each being an AR( d) linear model. The experts are combined via a gating network, which partitions the input space accordingly. Considering a scalar time series {xt}, we associate with each expert a probabilistic model (density function) relating input vectors x!=~ == [Xt-l,Xt-2, ... ,Xt-d] to an output scalar Xt E]R and denote these probabilistic models by p(xtl x!=~;Oj,O"j) j = 1,2, ... ,n where (OJ,O"j) is the expert parameter vector, taking values in a compact subset of ]Rd+l. In what follows we will use upper case X t to denote random variables, and lower case Xt to denote values taken by those r.v.'s. Letting the parameters of each expert network be denoted by ( OJ, 0" j ), j = 1,2, ... , n, those of the gating network by Og and letting 8 = ({OJ, O"j }j=l, Og) represent the complete set of parameters specifying the model, we may express the conditional distribution of the model, p(xtlx:=~, 8), as n p(Xtlx!=~; 8) = L gj (x!=~; Og)p(Xtlx!=~; OJ, O"j), (1) j=l Tune Series Prediction using Mixtures of Experts 311 o Vx~=~. We assume that the parameter vector e E n, a compact subset of JR2n(d+1) . Following the work of Jordan and Jacobs [2] we take the probability density functions to be Gaussian with mean Of X:=~ + OJ,O and variance (7j (representative of the underlying, local AR{d) model). The function 9j{X; Og) == exp{O~x + Og; .O}/(E~1 exp{O~x + Ogi.O}' thus implementing a multiple output logistic regression function. The underlying non-linear mapping (i.e., the conditional expectation, or L2 prediction function) characterizing the MEM, is described by using (1) to obtain the conditional expectation of Xt, n f~ = E[XtIX:=~; Mn] = L 9j{X:=~; Og)[Of X:=~ + OJ,o]' (2) j=1 where Mn denotes the MEM model. Here the subscript n stands for the number ~ () _ t-l d ~ of experts. Thus, we have Xt = fn = fn(Xt_d; e) where fn : JR x n ---+ JR, and Xt denotes the projection of X t on the 'relevant past', given the model, thus defining the model predictor function. We will use the notation MEM(n; d) where n is the number of experts in the model (proportional to the complexity, or number of parameters in the model), and d the lag size. In this work we assume that d is known and given. 3 Main results 3.1 Background We consider a stationary time series, more precisely a discrete time stochastic process {Xd which is assumed to be strictly stationary. We define the L2 predictor function f = E[X IXt- 1] = E[X IXt- 1] t -00 t t-d a.s. for some fixed lag size d. Markov chains are perhaps the most widely encountered class of probability models exhibiting this dependence. The NAR(d), that is nonlinear AR{ d), model is another example, widely studied in the context of time series (see [4] for details). Assuming additive noise, the NAR(d) model may be expressed as (3) We note that in this formulation {cd plays the role of the innovation process for Xt, and the function fe) describes the information on Xt contained within its past history. In what follows, we restrict the discussion to stochastic processes satisfying certain constraints on the memory decay, more precisely we are assuming that {Xd is an exponentially a-mixing process. Loosely stated, this assumption enables the process to have a law of large numbers associated with it, as well as a certain version of the central limit theorem. These results are the basis for analyzing the asymptotic behavior of certain parameter estimators (see, [9] for further details), but other than that this assumption is merely stated here for the sake of completeness. We note in 312 A. 1. Zeevi, R. Meir and R. 1. Adler passing that this assumption may be substantially weakened, and still allow similar results to hold, but requires more background and notation to be introduced, and therefore is not pursued in what follows (the reader is referred to [1] for further details). 3.2 Objectives Knowing the L2 predictor function, f, allows optimal prediction of future samples, where optimal is meant in the sense that the predicted value is the closest to the true value of the next sample point, in the mean squared error sense. It therefore seems a reasonable strategy, to try and learn the optimal predictor function, based on some finite realization of the stochastic process, which we will denote VN = {Xt}t!t'H. Note that for N » d, the number of sample points is approximately N. We therefore define our objective as follows. Based on the data V N , we seek the 'best' approximation to f, the L2 predictor function, using the MEM(n, d) predictor f~ E Mn as the approximator model. More precisely, define the least squares (LS) parameter estimator for the MEM(n, d) as N A • " [ t-I]2 9n,N = arg~l~ L.t Xt - fn(Xt _ d , 9) t=d+l where fn(X:=~, 9) is f~ evaluated at the point X:=~, and define the LS functional estimator as A _ (J fn,N = fnl(J=on,N where 9n ,N is the LS parameter estimator. Now, define the functional estimator risk as MSE[f, in,N] == Ev [/ If - in'NI2dv] where v is the d fold probability measure of the process {Xt}. In this work we maintain that the integration is over some compact domain Id C Rd, though recent work [3] has shown that the results can be extended to Rd, at the price of slightly slower convergence rates. It is reasonable, and quite customary, to expect a 'good' estimator to be one that is asymptotically unbiased. However, growth of the sample size itself need not, and in general does not, mean that the estimator is 'becoming' unbiased. Consequently, as a figure of merit, we may restrict attention to the approximation capacity of the model. That is we ask, what is the error in approximating a given class of predictor functions, using the MEM(n, d) (Le., {Mn}) as the approximator class. To measure this figure, we define the optimal risk as where f~ == f!I(J=(J* and MSE[j, f~] == / If - f~12dv, 9~ = argmin/ If - f~12dv, (JE9 Time Series Prediction using Mixtures of Experts 313 that is, 9~ is the parameter minimizing the expected L2 loss function. One may think of f~ as the 'best' predictor function in the class of approximators, i.e., the closest approximation to the optimal predictor, given the finite complexity, n, (Le., finite number of parameters) of the model. Here n is simply the number of experts (AR models) in the architecture. 3.3 Upper Bounds on the Mean Squared Error and Universal Approximation Results Consider first the case where we are simply interested in approximating the function f, assuming it belongs to some class of functions. The question then arises as to how well one may approximate f by a MEM architecture comprising n experts. The answer to this question is given in the following proposition, the proof of which can be found in [8]. Proposition 3.1 (Optimal risk bounds) Consider the class of functions Mn defined in (2) and assume that the optimal predictor f belongs to a Sobolev class containing r continuous derivatives in L 2 • Then the following bound holds: MSE[f, f~] ::; n2~jd where c is a constant independent of n . (4) PROOF SKETCH: The proof proceeds by first approximating the normalized gating function gj 0 by polynomials of finite degree, and then using the fact that polynomials can approximate functions in Sobolev space to within a known degree of approximation. The following main theorem, establishing upper bounds on the functional estimator risk, constitutes the main result of this paper. The proof is given in [9]. Theorem 3.1 (Upper bounds on the estimator risk) Suppose the stochastic process obeys the conditions set forth in the previous section. Assume also that the optimal predictor function, f, possesses r smooth derivatives in L 2 . Then for N sufficiently large we have ~ c m* ( 1 ) MSE[j, fn,N] ::; n2rjd + 2; + 0 N ' (5) where r is the number of continuous derivatives in L2 that f is assumed to possess, d is the lag size, and N is the size of the data set 'DN · PROOF SKETCH: The proof proceeds by a standard stochastic Taylor expansion of the loss around the point 9~. Making common regularity assumptions [1] and using the assumption on the a-mixing nature of the process allows one to establish the usual asymptotic normality results, from which the result follows. We use the notation m~ to denote the effective number of parameters. More precisely, m~ = Tr{B~(A~)-l} and the matrices A* and B* are related to the Fisher information matrix in the case of misspecified estimation (see [1] for further discussion). The upper bound presented in Theorem 3.1 is related to the classic bias - variance decomposition in statistics and the obvious tradeoffs are evident by inspection. 314 A. 1 Zeevi, R. Meir and R. 1 Adler 3.4 Comments It follows from Proposition 3.1 that the class of mixtures of experts is a universal approximator, w.r.t. the class of target functions defined for the optimal predictor. Moreover, Proposition 3.1 establishes the rate of convergence of the approximator, and therefore relates the approximation error to the number of experts used in the architecture (n) . Theorem 3.1 enhances this result, as it relates the sample complexity and model complexity, for this class of models. The upper bounds may be used in defining model selection criteria, based on upper bound minimization. In this setting, we may use an estimator of the stochastic error bound (i.e., the estimation error), to penalize the complexity of the model, in the spirit of Ale, MDL etc (see [8] for further discussion). At a first glance it may seem surprising to find that a combination of linear models is a universal function approximator. However, one must keep in mind that the global model is nonlinear, due to the gating network. Nevertheless, this result does imply, at least on a theoretical ground, that one may restrict the MEM{n, d) to be locally linear, without loss of generality. Thus, taking a simple local model, enabling efficient and tractable learning algorithms (see [2]), still results in a rich global model. 3.5 Comparison Recently, Mhaskar [5] proved upper bounds on a feedforward sigmoidal neural network, for target functions in the same class as we consider herein, i.e., the Sobolev class. The bound we have obtained in Proposition 3.1, and its extension in [8], demonstrate that w.r.t. to this particular target class, neural networks and mixtures of experts are equivalent. That is, both models attain optimal precision in the degree of approximation results (see [5] for details of this argument). Keeping in mind the advantages of the MEM with respect to learning and generalization [2], we believe that our results lend further credence to the emerging view as to the superiority of modular architectures over the more standard feed forward neural networks. Moreover, the detailed proof of Proposition 3.1 (see [8]) actually takes the MEM(n, d) to be made up of local constants. That is, the linear experts are degenerated to constant functions. Thus, one may conjecture that mixtures of experts are in fact a more general class than feedforward neural networks, though we have no proof of this as of yet. Two nonlinear alternatives, generalizing standard statistical linear models, have been pointed out in the introductory section. These are Tong's TAR (threshold autoregressive) model [7], and the more general SDM (state dependent models) introduced by Priestley. The latter models can be reduced to a TAR model by imposing a more restrictive structure (for further details see [6]) . We have shown, based on the results described above (see [9]), that the MEM may be viewed as a generalization of the SDM (and consequently of the TAR model). The relation to the state dependent models is of particular interest, as the mixtures of experts is structured on state dependence as well. Exact statement and proofs of these facts can be found in [9]. Time Series Prediction using Mixtures of Experts 315 We should also note that we have conducted several numerical experiments, comparing the performance of the MEM with other approaches. We tested the model on both synthetic as well as real-world data. Without any fine-tuning of parameters we found the performance of the MEM, with linear expert functions, to compare very favorably with other approaches (such as TAR, ARMA and neural networks). Details of the numerical results may be found in [9]. Moreover, the model also provided a very natural and intuitive segmentation of the process. 4 Discussion In this work we have pursued a novel non-linear model for prediction in stationary time series. The mixture of experts model (MEM) has been demonstrated to be a rich model, endowed with a sound theoretical basis, and compares favorably with other, state of the art, nonlinear models. We hope that the results of this study will aid in establishing the MEM as, yet another, powerful tool for the study of time-series applicable to the fields of statistics, economics, and signal processing. References [1] Domowitz, I. and White, H. "Misspecified Models with Dependent Observations", Journal of Econometrics, vol. 20: 35-58, 1982. [2] Jordan, M. and Jacobs, R. "Hierarchical Mixtures of Experts and the EM Algorithm" , Neural Computation, vol. 6, pp. 181-214, 1994. [3] Maiorov, V. and Meir. V. "Approximation Bounds for Smooth Functions in C(1Rd) by Neural and Mixture Networks", submitted for publication, December 1996. [4] Meyn, S.P. and Tweedie, R.L. (1993) Markov Chains and Stochastic Stability, Springer-Verlag, London. [5] Mhaskar, H. (1996) "Neural Networks for Optimal Approximation of Smooth and Analytic Functions", Neural Computation vol. 8(1), pp. 164-177. [6] Priestley M.B. Non-linear and Non-stationary Time Series Analysis, Academic Press, New York, 1988. [7] Tong, H. Threshold Models in Non-linear Time Series Analysis, Springer Verlag, New York, 1983. [8] Zeevi, A.J., Meir, R. and Maiorov, V. "Error Bounds for Functional Approximation and Estimation Using Mixtures of Experts", EE Pub. CC-132., Electrical Engineerin g Department, Technion, 1995. [9] Zeevi, A.J., Meir, R. and Adler, R.J. "Non-linear Models for Time Series Using Mixtures of Experts", EE Pub. CC-150, Electrical Engineering Department, Technion, 1996. PART IV ALGORITHMS AND ARCHITECTURE
1996
47
1,244
MIMIC: Finding Optima by Estimating Probability Densities Jeremy S. De Bonet, Charles L. Isbell, Jr., Paul Viola Artificial Intelligence Laboratory Massachusetts Institute of Technology Cambridge, MA 02139 Abstract In many optimization problems, the structure of solutions reflects complex relationships between the different input parameters. For example, experience may tell us that certain parameters are closely related and should not be explored independently. Similarly, experience may establish that a subset of parameters must take on particular values. Any search of the cost landscape should take advantage of these relationships. We present MIMIC, a framework in which we analyze the global structure of the optimization landscape. A novel and efficient algorithm for the estimation of this structure is derived. We use knowledge of this structure to guide a randomized search through the solution space and, in turn, to refine our estimate ofthe structure. Our technique obtains significant speed gains over other randomized optimization procedures. 1 Introduction Given some cost function C(x) with local minima, we may search for the optimal x in many ways. Variations of gradient descent are perhaps the most popular. When most of the minima are far from optimal, the search must either include a brute-force component or incorporate randomization. Classical examples include Simulated Annealing (SA) and Genetic Algorithms (GAs) (Kirkpatrick, Gelatt and Vecchi, 1983; Holland, 1975). In all cases, in the process of optimizing C(x) many thousands or perhaps millions of samples of C( x) are evaluated. Most optimization algorithms take these millions of pieces of information, and compress them into a single point x-the current estimate of the solution (one notable exception are GAs to which we will return shortly). Imagine splitting the search process into two parts, both taking t/2 time steps. Both parts are structurally identical: taking a description of CO, they start their search from some initial point. The sole benefit enjoyed by the second part of the search over the first is that the initial MIMIC: Finding Optima by Estimating Probability Densities 425 point is perhaps closer to the optimum. Intuitively, there must be some additional information that could be learned from the first half of the search, if only to warn the second half about avoidable mistakes and pitfalls. We present an optimization algorithm called Mutual-Information-Maximizing Input Clustering (MIMIC). It attempts to communicate information about the cost function obtained from one iteration of the search to later iterations of the search directly. It does this in an efficient and principled way. There are two main components of MIMIC: first, a randomized optimization algorithm that samples from those regions of the input space most likely to contain the minimum for CO; second, an effective density estimator that can be used to capture a wide variety of structure on the input space, yet is computable from simple second order statistics on the data. MIMIC's results on simple cost functions indicate an order of magnitude improvement in performance over related approaches. Further experiments on a k-color map coloring problem yield similar improvements. 2 Related Work Many well known optimization procedures neither represent nor utilize the structure of the optimization landscape. In contrast, Genetic Algorithms (GA) attempt to capture this structure by an ad hoc embedding of the parameters onto a line (the chromosome). The intent of the crossover operation in standard genetic algorithms is to preserve and propagate a group of parameters that might be partially responsible for generating a favorable evaluation. Even when such groups exist, many of the offspring generated do not preserve the structure of these groups because the choice of crossover point is random. In problems where the benefit of a parameter is completely independent of the value of all other parameters, the population simply encodes information about the probability distribution over each parameter. In this case, the crossover operation is equivalent to sampling from this distribution; the more crossovers the better the sample. Even in problems where fitness is obtained through the combined effects of clusters of inputs, the GA crossover operation is beneficial only when its randomly chosen clusters happen to closely match the underlying structure of the problem. Because of the rarity of such a fortuitous occurrence, the benefit of the crossover operation is greatly diminished. As as result, GAs have a checkered history in function optimization (Baum, Boneh and Garrett, 1995; Lang, 1995). One of our goals is to incorporate insights from GAs in a principled optimization framework. There have been other attempts to capture the advantages of GAs. Population Based Incremental Learning (PBIL) attempts to incorporate the notion of a candidate population by replacing it with a single probability vector (Baluja and Caruana, 1995). Each element of the vector is the probability that a particular bit in a solution is on. During the learning process, the probability vector can be thought of as a simple model of the optimization landscape. Bits whose values are firmly established have probabilities that are close to lor O. Those that are still unknown have probabilities close to 0.5. When it is the structure of the components of a candidate rather than the particular values of the components that determines how it fares, it can be difficult to move PBIL's representation towards a viable solution. Nevertheless, even in these sorts of problems PBIL often out-performs genetic algorithms because those algorithms are hindered by the fact that random crossovers are infrequently beneficial. A very distinct, but related technique was proposed by Sabes and Jordan for a 426 J. S. de Bonet, C. L. Isbell and P. WoLa reinforcement learning task (Sabes and Jordan, 1995). In their framework, the learner must generate actions so that a reinforcement function can be completely explored. Simultaneously, the learner must exploit what it has learned so as to optimize the long-term reward. Sabes and Jordan chose to construct a Boltzmann distribution from the reinforcement function: p(x) = exp~~) where R(x) is the reinforcement function for action X, T is the temperature, and ZT is a normalization factor. They use this distribution to generate actions. At high temperatures this distribution approaches the uniform distribution, and results in random exploration of RO. At low temperatures only those actions which garner large reinforcement are generated. By reducing T, the learner progresses from an initially randomized search to a more directed search about the true optimal action. Interestingly, their estimate for p( x) is to some extent a model of the optimization landscape which is constructed during the learning process. To our knowledge, Sabes and Jordan have neither attempted optimization over high dimensional spaces, nor attempted to fit p( x) with a complex model. 3 MIMIC Knowing nothing else about C(x) it might not be unreasonable to search for its minimum by generating points from a uniform distribution over the inputs p( x). Such a search allows none of the information generated by previous samples to effect the generation of subsequent samples. Not surprisingly, much less work might be necessary if samples were generated from a distribution, p8(x), that is uniformly distributed over those x's where C(x) ~ 0 and has a probability of 0 elsewhere. For example, if we had access to p8 M (x) for OM = minx C( x) a single sample would be sufficient to find an optimum. This insight suggests a process of successive approximation: given a collection of points for which C( x) ~ 00 a density estimator for p/Jo (x) is constructed. From this density estimator additional samples are generated, a new threshold established, 01 = 00 f, and a new density estimator created. The process is repeated until the values of C( x) cease to improve. The MIMIC algorithm begins by generating a random population of candidates choosen uniformly from the input space. From this population the median fitness is extracted and is denoted 00 . The algorithm then proceeds: 1. Update the parameters of the density estimator of p/J·(x) from a sample. 2. Generate more samples from the distribution p/J·(x). 3. Set 0i+l equal to the Nth percentile of the data. Retain only the points less than Oi+1 ' The validity of this approach is dependent on two critical assumptions: p(\x) can be successfully approximated with a finite amount of data; and D(pl1-f(X)llp (x)) is small enough so that samples from p8(x) are also likely to be samples from p/J-f(X) (where D(pllq) is the Kullback-Liebler divergence between p and q). Bounds on these conditions can be used to prove convergence in a finite number of successive approximation steps. The performance of this approach is dependent on the nature of the density approximator used. We have chosen to estimate the conditional distributions for every pair of parameters in the representation, a total of O( n2 ) numbers. In the next section we will show how we use these conditionals distributions to construct a joint distribution which is closest in the KL sense to the true joint distribution. Such an MIMIC: Finding Optima by Estimating Probability Densities 427 approximator is capable of representing clusters of highly related parameters. While this might seem similar to the intuitive behavior of crossover, this representation is strictly more powerful. More importantly, our clusters are learned from the data, and are not pre-defined by the programmer. 4 Generating Events from Conditional Probabilities The joint probability distribution over a set of random variables, X = {Xi}, is: Given only pairwise conditional probabilities, p(Xi IXj) and unconditional probabilities, p(Xi), we are faced with the task of generating samples that match as closely as possible the true joint distribution, p(X). It is not possible to capture all possible joint distributions of n variables using only the unconditional and pairwise conditional probabilities; however, we would like to describe the true joint distribution as closely as possible. Below, we derive an algorithm for choosing such a description. Given a permutation of the numbers between 1 and n, 7r = i1 i2 ... in, we define a class of probability distributions, P1l"(X): (2) The distribution P1l"(X) uses 7r as an ordering for the pairwise conditional probabilities. Our goal is to choose the permutation 7r that maximizes the agreement between P1l"(X) and the true distribution p(X). The agreement between two distributions can be measured by the Kullback-Liebler divergence: D(pllp1l") = l p[logp - logp1l" ]dX = Ep[logp] - Ep[logp1l"] = -h(p) - Ep[logp(XilIXh)P(Xi2IXi3) .. . p(Xin_lIXi,,)p(Xin)] = -h(p) + h(Xi1IXi2) + h(Xh IXi3) + .. . + h(Xin_1IXiJ + h(XiJ. This divergence is always non-negative, with equality only in the case where p(7r) and p(X) are identical distributions. The optimal 7r is defined as the one that minimizes this divergence. For a distribution that can be completely described by pairwise conditional probabilities, the optimal 7r will generate a distribution that will be identical to the true distribution. Insofar as the true distribution cannot be captured this way, the optimal P1l"(X) will diverge from that distribution. The first term in the divergence does not depend on 7r. Therefore, the cost function, J1l"(X), we wish to minimize is: The optimal 7r is the one that produces the lowest pairwise entropy with respect to the true distribution. By searching over all n! permutations, it is possible to determine the optimal 7r. In the interests of computational efficiency, we employ a straightforward greedy algorithm to pick a permutation: 428 J. S. de Bonet, C. L. Isbell and P. Viola 1. in =:: arg minj h(Xj). 2. ik =:: arg minj h( Xj IXik+J, where j t= ik+1 ... in and k =:: n - 1, n - 2, ... ,2,1. where hO is the empirical entropy. Once a distribution is chosen, generating samples is also straightforward: 1. Choose a value for Xin based on its empirical probability P(Xin). 2. for k =:: n - 1, n - 2, ... ,2,1, choose element Xik based on the empirical conditional probability P(Xik jXik+1 )· The first algorithm runs in time O(n2 ) and the second in time O(n2 ). 5 Experiments To measure the performance of MIMIC, we performed three benchmark experiments and compared our results with those obtained using several standard optimization algorithms. We will use four algorithms in our comparisons: 1. MIMIC - the algorithm above with 200 samples per iteration 2. PBIL - standard population based incremental learning 3. RHC - randomized hill climbing 4. GA - a standard genetic algorithm with single crossover and 10% mutation rate 5.1 Four Peaks The Four Peaks problem is taken from (Baluja and Caruana, 1995). Given an N -dimensional input vector X, the four peaks evaluation function is defined as: where I(X, T) =:: max [tail(O, X), head(l, X)] + R(X, T) tai/(b, X) =:: number of trailing b's in X head(b, X) =:: number of leading b's in X R(X T) = {N iftail(?,X) > T and head(l,X) > T , 0 otherWIse (4) (5) (6) (7) There are two global maxima for this function. They are achieved either when there are T + 1 leading l's followed by all O's or when there are T + 1 trailing O's preceded by all 1 'so There are also two suboptimal local maxima that occur with a string of all l's or all O's. For large values of T, this problem becomes increasingly more difficult because the basin of attraction for the inferior local maxima become larger. Results for running the algorithms are shown in figure 1. In all trials, T was set to be 10% of N, the total number of inputs. The MIMIC algorithm consistently maximizes the function with approximately one tenth the number of evaluations required by the second best algorithm. MIMIC: Finding Optima by Estimating Probability Densities Function Evaluations Required to Maximize 4 Peaks 1200.--~-~--~-~----, ~ I ()()() • MIMIC o .~ 0 PBIL ~ 800 x RHC & • GA ~ 600 '------" il ~400 5 ~ 200l::=;;~~~~~:::::=J o 40 50 60 70 80 Inputs 429 Figure 1: Number of evaluations of the Four-Peak cost function for different algorithms plotted for a variety of problems sizes. 5.2 Six Peaks The Six Peaks problem is a slight variation on Four Peaks where R(X,T) = { ; if tai/(O,x) > T and head(l, x) > Tor tai/(l, x) > T and head(O, x) > T otherwise (8) This function has two additional global maxima where there are T + 1 leading O's followed by all 1 's or when there are T + 1 trailing 1 's preceded by all O's. In this case, it is not the values of the candidates that is important, but their structure: the first T + 1 positions should take on the same value, the last T + 1 positions should take on the same value, these two groups should take on different values, and the middle positions should take on all the same value. Results for this problem are shown in figure 2. As might be expected, PBIL performed worse than on the Four Peak problem because it tends to oscillate in the middle of the space while contradictory signals pull it back and forth. The random crossover operation of the G A occasionally was able to capture some of the underlying structure, resulting in an improved relative performance of the GA. As we expected, the MIMIC algorithm was able to capture the underlying structure of the problem, and combine information from all the maxima. Thus MIMIC consistently maximizes the Six Peaks function with approximately one fiftieth the number of evaluations required by the other algorithms. 5.3 Max K-Coloring A graph is K-Colorable if it is possible to assign one of k colors to each of the nodes of the graph such that no adjacent nodes have the same color. Determining whether a graph is K-Colorable is known to be NP-Complete. Here, we define Max K-Coloring to be the task of finding a coloring that minimizes the number of adjacent pairs colored the same. Results for this problem are shown in figure 2. We used a subset of graphs with a single solution (up to permutations of color) so that the optimal solution is dependent only on the structure of the parameters. Because of this, PBIL performs poorly. GA's perform better because any crossover point is representative of some of the underlying structure of the graphs used. Finally, MIMIC performs best because 430 Function Evaluations Required to Maximize 6 Peaks 1300r---~--~-~--~---' ",1200 e o ·~1000 .a '" W 800 • MIMIC o PBIL x RHC + GA 'Cl L--_--' ~600 § ~ 400 o ~ 200 o 20 30 40 50 60 Inputs J. S. de Bonet, C. L. Isbell and P. Vwla Function Evaluations Required to Maximize K-Coloring 1200r---~-~-~-~-~---.. :gIOOO o .;:1 '" ~ 800 W 'Cl 600 ~ ~400 g ~ 200 • MIMIC o PBIL x RHC + GA 40 Figure 2: Number of evaluations of the Six-Peak cost function (left) and the K-Color cost function (right) for a variety of problem sizes. it is able to capture all of the structural regularity within the inputs. 6 Conclusions We have described MIMIC, a novel optimization algorithm that converges faster and more reliably than several other existing algorithms. MIMIC accomplishes this in two ways. First, it performs optimization by successively approximating the conditional distribution of the inputs given a bound on the cost function. Throughout this process, the optimum of the cost function becomes gradually more likely. As a result, MIMIC directly communicates information about the cost function from the early stages to the later stages of the search. Second, MIMIC attempts to discover common underlying structure about optima by computing second-order statistics and sampling from a distribution consistent with those statistics. Acknowledgments In this research, Jeremy De Bonet is supported by the DOD Multidisciplinary Research Program of the University Research Initiative, Charles Isbell by a fellowship granted by AT&T Labs-Research, and Paul Viola by Office of Naval Research Grant No. N00014-96-1-0311. Greg Galperin helped in the preparation of this paper. References Baluja, S. and Caruana, R. (1995). Removing the genetics from the standard genetic algorithm. Technical report, Carnegie Mellon Univerisity. Baum, E. B., Boneh, D., and Garrett, C. (1995). Where genetic algorithms excel. In Proceedings of the Conference on Computational Learning Theory, New York. Association for Computing Machinery. Holland, J. H. (1975). Adaptation in Natural and Artificial Systems. The Michigan University Press. Kirkpatrick, S., Gelatt, C., and Vecchi, M. (1983). Optimization by Simulated Annealing. Science, 220(4598):671-680. Lang, K. (1995). Hill climbing beats genetic search on a boolean circuit synthesis problem of koza's. In Twelfth International Conference on Machine Learning. Sabes, P. N. and Jordan, M. 1. (1995). Reinforcement learning by probability matching. In David S. Touretzky, M. M. and Perrone, M., editors, Advances in Neural Information Processing, volume 8, Denver 1995. MIT Press, Cambridge.
1996
48
1,245
Neural network models of chemotaxis the nematode Caenorhabditis elegans Thomas C. Ferree, Ben A. Marcotte, Shawn R. Lockery Institute of Neuroscience, University of Oregon, Eugene, Oregon 97403 Abstract We train recurrent networks to control chemotaxis in a computer model of the nematode C. elegans. The model presented is based closely on the body mechanics, behavioral analyses, neuroanatomy and neurophysiology of C. elegans, each imposing constraints relevant for information processing. Simulated worms moving autonomously in simulated chemical environments display a variety of chemotaxis strategies similar to those of biological worms. 1 INTRODUCTION • In The nematode C. elegans provides a unique opportunity to study the neuronal basis of neural computation in an animal capable of complex goal-oriented behaviors. The adult hermaphrodite is only 1 mm long, and has exactly 302 neurons and 95 muscle cells. The morphology of every cell and the location of most electrical and chemical synapses are known precisely (White et al., 1986), making C. elegans especially attractive for study. Whole-cell recordings are now being made on identified neurons in the nerve ring of C. elegans to determine electrophysiological properties which underly information processing in this animal (Lockery and Goodman, unpublished). However, the strengths and polarities of synaptic connections are not known, so we use neural network optimization to find sets of synaptic strengths which reproduce actual nematode behavior in a simulated worm. We focus on chemotaxis, the ability to move up (or down) a gradient of chemical attractants (or repellants). In the laboratory, flat Petri dishes (radius = 4.25 cm) are prepared with a Gaussian-shaped field of attractant at the center, and worms are allowed to move freely about. Worms propel themselves forward by generating an undulatory body wave, which produces sinusoidal movement. In chemotaxis, the nervous system generates motor commands which bias this movement and direct 56 T. C. Ferree, B. A. Marcotte and S. R. Lockery the animal toward higher attractant concentration. Anatomical constraints pose important problems for C. elegans during chemotaxis. In particular, the animal detects the presence of chemicals with a pair of sensory organs (amphids) at the tip of the nose, each containing the processes of multiple chemosensory neurons. During normal locomotion, however, the animal moves on its side so that the two amphids are perpendicular to the Petri dish. C. elegans cannot, therefore, sense the gradient directly. One possible strategy for chemotaxis, which has been suggested previously (Ward, 1973), is that the animal computes a temporal derivative of the local concentration during a single head sweep, and combines this with some form of proprioceptive feedback indicating muscle contraction and the direction of head sweep, to compute the spatial gradient for chemotaxis. The existence of this and other strategies is discussed later. In Section 2, we derive a simple model of the nematode body which produces realistic sinusoidal trajectories in response to motor commands from the nervous system. In Section 3, we give a simple model of the C. elegans nervous system based on preliminary physiological data. In Section 4, we use a stochastic optimization algorithm to determine sets of synaptic weights which control chemotaxis, and discuss solutions. 2 BIOMECHANICS OF NEMATODE ORIENTATION Nematode locomotion has been studied in detail (Niebur and Erdos, 1991; Niebur and Erdos, 1993). These authors derived Newtonian force equations for each muscular segment of the body, which can be solved numerically to generate forward sinusoidal movement. Unfortunately, such a thorough treatment is computationally intensive and not practical to use with network optimization. To simplify the problem we first recognize that chemotaxis is a behavior more of orientation than of locomotion. We therefore derive a set of biomechanical equations which direct the head to generate sinusoidal movement, which can be biased by the network toward higher chemical concentrations. We focus our attention on the point (x,1/) at the tip of the nose, since that is where the animal senses the chemical environment. As shown in Figure 1 ( a), we assign a velocity vector fJ directed along the midline of the first body segment, i.e., the head. Assuming that the worm moves forward at constant speed v, we can write the velocity vector as fJ(t) = (~;, ~~) = (vcos(}(t),vsin(}(t)) (1) where x, y and (} are measured relative to fixed coordinates in the Petri dish. Assuming that the worm moves without lateral slipping and that the undulatory wave of muscular contraction initiated in the neck travels posteriorally without modification, then each body segment simply follows the one previous (anterior) to it. In this way, the head directs the movement and the rest of the body simply follows. Figure l(b) shows an expanded view of the neck segment. As the worm moves forward, the posterior boundary of that segment assumes the position held by its anterior neighbor at a slightly earlier time. If L is the total body length and N is Neural Network Models of Chemotaxis 57 the number of body segments, then this time delay is 6t ~ L/Nv. (For L = 1 mm, v = 0.22 mm/s and N = 10 we have 6t ~ 0.45 s, roughly an order of magnitude smaller than the relevant behavioral time scale: the head-sweep period T ~ 4.2 s.) If we define the neck angle aCt) == (h(t) - 92(t), then the above arguments imply dOl aCt) = 91 (t) - 91 (t - 6t) ~ dt 6t (2) where the second relation is essentially a backward-Euler algorithm for dOl/dt. Since 9 == 91, we have reached the intuitive result that the neck angle a determines the rate of turning dO/dt. Note that while 91 and 92 are defined relative to the fixed laboratory coordinates, their difference a is invariant under rotations of these coordinates, and can therefore be viewed as intrinsic to the body. This allows us to derive an expression for a in terms of muscle cell contraction, or motor neuron depolarization, as follows. (a) v (b) Neck T Iv 1 Figure 1: Nematode body mechanics. (a) Segmented model of the nematode body, showing the direction of motion v. (b) Expanded view of the neck segment, showing dorsal (D) and ventral (V) neck muscles. Nematodes maintain nearly constant volume during movement. To incorporate this constraint, albeit approximately, we assume that at all times the geometry of each segment is such that (ID -10) = -(Iv -10), where 10 == L/N is the equilibrium length of a relaxed segment. For small angles a, we have a ~ (Iv -ID)/d, where d is the body diameter. The dashed lines in Figure 1(b) indicate dorsal and ventral muscles, which are believed to develop tension nearly independent of length (Toida et al., 1975). When contracting, these muscles must work against the elasticity of the cuticle, internal fluid pressure, and elasticity and developed tension of the opposing muscles. If these elastic forces act linearly, then TD-TV ~ k (Iv-ID), where TD and Tv are dorsal and ventral muscle tensions, and k is an effective force constant. For simplicity, we further assume that each muscle develops tension linearly in response to the voltage of its corresponding motor neuron, i.e., TD,v = E VD,V, where E is a positive constant, and VD and Vv are dorsal and ventral motor neuron voltages. Combining these results, we have finally ~~ = "1 (VD(t) - Vv(t») (3) 58 T. C. Ferree, B. A. Marcotte and S. R. Lockery where"Y = (Nv/L)· (E/kd). With appropriate motor commands, equations (I) and (3) can be integrated numerically to generate sinusoidal worm trajectories like those of biological worms. This model embodies the main anatomical features that are likely to be important in C. elegans chemotaxis, yet is sufficiently compact to be embedded in a network optimization procedure. 3 CHEMOTAXIS CONTROL CIRCUIT C. elegans neurons are tiny and have very simple morpologies: a typical neuron in the head has a spherical soma 1-2 pm in diameter, and a single cylindrical process 60-80 pm in length and 0.1-0.2 pm in diameter. Compartmental models, based on this morphology and preliminary physiological recordings, indicate that C. elegans neurons are effectively isopotential (Lockery, 1995). Furthermore, C. elegans neurons do not fire classical all-or-none action potentials, but appear to rely primarily on graded signal propagation (Lockery and Goodman, unpublished). Thus, a reasonable starting point for a network model is to represent each neuron by a single isopotential compartment, in which voltage is the state variable, and the membrane conductance in purely ohmic. Anatomical data indicate that the C. elegans nervous system has both electrical and chemical synapses, but the synaptic transfer functions are not known. However, steady-state synaptic transfer functions for chemical synapses have been measured in Ascaris s'U'Um, a related species of nematode, where it was found that postsynaptic voltage is a graded function of presynaptic voltage, due to tonic neurotransmitter release (Davis and Stretton, 1989). This voltage dependence is sigmoidal, i.e., VpOlt "-J tanh(Vpre). A simple network model which captures all of these features is T~ = -Vi + Vmax tanh (f3 t Wij (Vj - Vj») + ~Iilm{t) (4) 3=1 where Vi is the voltage of the ith neuron. Here all voltages are measured relative to a common resting potential, Vmax is an arbitrary voltage scale which sets the operational range of the neurons, and f3 sets the voltage sensitivity of the synaptic transfer function. The weight Wij represents the net strength and polarity of all synaptic connections from neuron j to neuron i, and the constants Vj determine the center of each transfer function. The membane time constant T is assumed to be the same for all cells, and will be discussed further later. Note that in (4), synaptic transmission occurs instantaneously: the time constant T arises from capacitive current through the cell membrane, and is unrelated to synaptic transmission. Note also that the way in which (4) sums multiple inputs is not unique, i.e., other sigmoidal models which sum inputs differently are equally plausible, since no data on synaptic summation exists for either C. elegans or Ascaris 8'U'Uffl. . The stimulus term ~8t1m(t) is used to introduce chemosensation and sinusoidal locomotion to the network in (4). We use i = 1 to label a single chemosensory neuron at the tip of the nose, and i = n - 1 == D and i = n == V to label dorsal and ventral motor neurons. For simplicity we assume that the chemosensory neuron voltage responds linearly to the local chemical concentration: (5) Neural Network Models of Chemotaxis 59 where Vchem is a positive constant, and the local concentration C(x, y) is always evaluated at the instantaneous nose position. In the previous section, we emphasized that locomotion is effectively independent of orientation. We therefore assume the existence of a central pattern generator (CPG) which is outside the chemotaxis control circuit (4). Thus, in addition to synaptic input from other neurons, each motor neuron receives a sinusoidal stimulus (6) where VCPG and w = 211" /T are positive constants. 4 RESULTS AND DISCUSSION Equations (1), (3) and (4), together with (5) and (6), comprise a set of n + 3 first-order nonlinear differential equations, which can be solved numerically given initial conditions and a specification of the chemical environment. We use a fourthorder Runge-Kutta algorithm and find favorable stability and convergence. The necessary body parameters have been measured by observing actual worms (Pierce and Lockery, unpublished): v = 0.022 cm/s, T = 4.2 s and '"Y = 0.8/(2VcPG). The chemical environment is also chosen to agree roughly with experimental values: C(x,y) = Coexp(-(x2 + y2)/-Xb), with Co = 0.052 p.mol/cm3 and -Xc = 2.3 cm. To optimize networks to control chemotaxis, we use a simple simulated annealing algorithm which searches over the (n2 + 3)-dimensional space of parameters Wij, {3, Vchem and VCPG. In the results shown here, we used n = 12, and set V; = O. Each set of the resulting parameters represents a different nervous system for the model worm. At the beginning of each run, the worm is initialized by choosing an initial position (xo, Yo), an initial angle 00 , and by setting V. = O. Upon numerically integrating, simulated worms move autonomously in their environment for a predetermined amount of time, typically the real-time equivalent of 10-15 minutes. We quantify the performance, or fitness, of each worm during chemotaxis by computing the average chemical concentation at the tip of its nose over the duration of each run. To avoid lucky scores, the actual score for each worm is obtained by averaging over several initial conditions. In Figure 2, we show a comparison of tracks produced by (a) biological and (b) simulated worms during chemotaxis. In each case, three worms were placed in a dish with a radial gradient and allowed to move freely for the real-time equivalent of 15 minutes. In (b), the three worms have the same neural parameters (Wij, (3, Vchem, VCPG), but different initial angles 00 • In both (a) and (b), all three worms make initial movements, then move toward the center of the dish and remain there. In other optimizations, rather than orbit the center, the simulated worms may approach the center asymptotically from one side, make simple geometric patterns which pass through the center, or exhibit a variety of other distinct strategies for chemotaxis. This is similar to the situation with biological worms, which also have considerable variation in the details of their tracks. The behavior shown in Figure 2 was produced using T = 500 ms. However, preliminary electrophysiological recordings from C. elegans neurons suggest that the actual value may be as much as an order of magnitude smaller, but not bigger (Lockery and Goodman, unpublished). This presents a potential problem for chemotaxis com60 T. C. Ferree, B. A. Marcotte and S. R. Lockery putation, since shorter time constants require greater sensitivity to small changes in O(x, y) in order to compute a temporal derivative, which is believed to be required. During optimization, we have seen that for a fixed number of neurons n, finding optimal solutions becomes more difficult as T is decreased. This observation is very difficult to quantify, however, due to the existence of local maxima in the fitness function. Nevertheless, this suggests that additional mechanisms may need to be included to understand neural computation in C. elegana. First, time- and voltage-dependent conductances will modify the effective membrane time constant, and may increase the effective time scale for computation by individual neurons. Second, more neurons and synaptic delays will also move the effective neuronal time scale closer to that of the behavior. Either of these will allow comparisons of O(z, II) across greater distances, thereby requiring less sensitivity to compute the gradient, and potentially improving the ability of these networks to control chemotaxis. 2em Figure 2: Nematodes performing chemotaxis: (a) biological (Pierce and Lockery, unpublished), and (b) simulated. We also note, based on a variety of other results, not shown here, that the headsweep strategy, described in the introduction, is by no means the only strategy for chemotaxis in this system. In particular, we have optimized networks without a CPG, i.e., with VCPG = 0 in (6), and found parameter sets that successfully control chemotaxis. This presents the possibility that even worms with a CPG do not necessarily compute the gradient based on lateral movement of the head, but may instead respond only to changes in concentration along their mean trajectory. Similar results have been reported previously, although based on a somewhat different biomechanical model (Beer and Gallagher, 1992). Finally, we have also optimized discrete-time networks, obtained by setting T = 0 in (4) and updating all units synchronously. As is well-known, on relatively short time scales (- T) such a system tends to "overshoot" at each successive time step, leading to sporadic behavior of the network and the body. Knowing this, it is interesting that simulated worms with such a nervous system are capable of reliable behavior over longer time scales, i.e., they successfully perform chemotaxis. Neural Network Models of Chemotaxis 61 5 CONCLUSIONS AND FUTURE WORK The main result of this paper is that a small nervous system, based on gradedpotential neurons, is capable of controlling chemotaxis in a worm-like physical body with the dimensions of C. elegans. The model presented is based closely on the body mechanics, behavioral analyses, neuroanatomy and neurophysiology of C. elegans, and is a reliable starting point for more realistic models to follow. Furthermore, we have established the existence of chemotaxis strategies that had not been anticipated based on behavioral experiments with real worms. Future work will involve both improvement of the model and analysis of the resulting solutions. Improvements will include introducing voltage- and time-dependent membrane conductances, as this data becomes available, and more realistic models of synaptic transmission. Also, laser ablation experiments have been performed that suggest which interneurons and motor neurons in C. elegans may be important for chemotaxis (Bargmann, unpublished), and these data can be used to constrain the synaptic connections during optimization. Analyses will be aimed at determining the role of individual physiological and anatomical features, and how they function together to govern the collective properties of the network as a whole during chemotaxis. Acknowledgements The authors would like to thank Miriam Goodman and Jon Pierce for helpful discussions. This work has been supported by NIMH MH11373, NIMH MH51383, NSF IBN 9458102, ONR N00014-94-1-0642, the Sloan Foundation, and The Searle Scholars Program. References Beer, R. D. and J. C. Gallagher (1992). Evolving dynamical neural networks for adaptive behavior, Adaptive Behavior 1(1):91-122. Davis, R. E. and A. O. W. Stretton {1989}. Signaling properties of Ascaris matorneurons: Graded active responses, graded synaptic transmission, and tonic transmitter release, J. Neurosci. 9:415-425. Lockery, S. R. (1995). Signal propagation in the nerve ring of C. elegans, Soc. Neurosd. Abstr. 569.1:1454. Niebur, E. and P. Erdos (1991). Theory of the locomotion of nematodes: Dynamics of undulatory progression on a surface, Biophys. J. 60:1132-1146. Niebur, E. and P. Erdos (1993). Theory of the locomotion of nematodes: Control of the somatic motor neurons by interneurons, Math. Biosci. 118:51-82. Toida, N., H. Kuriyama, N. Tashiro and Y. Ito {1975}. Obliquely striated muscle, Physiol. Rev. 55:700-756. Ward, S. (1973). Chemotaxis by the nematode Caenorhabditis elegans: Identification of attractants and analysis of the response by use of mutants, Proc. Nat. Acad. Sci. USA 10:817-821. White, J. G., E. Southgate, J. N. Thompson and S. Brenner (1986). The structure of the nervous system of C. elegans, Phil. Trans. R. Soc. London 314:1-340.
1996
49
1,246
A Silicon Model of Amplitude Modulation Detection in the Auditory Brainstem And~ van Schaik, Eric Fragniere, Eric Vittoz MANIRA Center for Neuromimetic Systems Swiss Federal Institute of Technology CH-lOlS Lausanne email: Andre.van_Schaik@di.epfl.ch Abstract Detectim of the periodicity of amplitude modulatim is a major step in the determinatim of the pitch of a SOODd. In this article we will present a silicm model that uses synchrroicity of spiking neurms to extract the fundamental frequency of a SOODd. It is based m the observatim that the so called 'Choppers' in the mammalian Cochlear Nucleus synchrmize well for certain rates of amplitude modulatim, depending m the cell's intrinsic chopping frequency. Our silicm model uses three different circuits, i.e., an artificial cochlea, an Inner Hair Cell circuit, and a spiking neuron circuit 1. INTRODUCTION Over the last few years, we have developed and implemented several analog VLSI building blocks that allow us to model parts of the auditory pathway [1], [2], [3]. This paper presents me experiment using these building blocks to create a model for the detection of the fundamental frequency of a harmroic complex. The estimatim of this fundamental frequency by the model shows some important similarities with psychoacoustic experiments in pitch estimation in humans [4]. A good model of pitch estimation will give us valuable insights in the way the brain processes sounds. Furthermore, a practical application to speech recognition can be expected, either by using the pitch estimate as an element in the acoustic vector fed to the recognizer, or by normalizing the acoustic vector to the pitch. 742 A. van Schaik, E. Fragniere and E. Wttoz Although the model doesn't yield a complete model of pitch estimatim, and explains probably mly one of a few different mechanisms the brain uses for pitch estimatim, it can give us a better understanding of the physiological background of psycho-acoustic results. An electrmic model can be especially helpful, when the parameters of the model can be easily controlled, and when the model will operate in real time. 2. THE MODEL The model was originally developed by Hewitt and Meddis [4], and was based m the observatim that Chopper cells in the Cochlear Nucleus synchronize when the stimulus is modulated in amplitude within a particular modulation frequency range [5]. Coehl ........ ecIuudeaI Ilierlug lDaer IbIr n111 AN A A til ~ Hair e4illllCldoi Inferior ... Beahu ecJInddmee nil Fig. 1. Diagram of the AM detection model. BMF=Best Modulation Frequency. The diagram shown in figure 1 shows the elements of the model. The cochlea filters the incoming sound signal. Since the width of the pass-band of a cochlear band-pass filter is proportimal to its cut-off frequency, the filters will not be able to resolve the individual harmooics of a high frequency carrier (>3kHz) amplitude modulated at a low rate «500Hz). The outputs of the cochlear filters that have their cut-off frequency slightly above the carrier frequency of the signal will therefm-e still be modulated in amplitude at the mginal modulatim frequency. This modulatim compment will therefm-e synchronize a certain group of Chopper cells. The synchronizatim of this group of Chopper cells can be detected using a coincidence detecting neurm, and signals the presence of a particular amplitude modulation frequency. This model is biologically plausible, because it is known that the choppers synchronize to a particular amplitude modulatim frequency and that they project their output towards the Inferim- Colliculus (ammgst others). Furthermm-e, neurms that can functim as coincidence detectm-s are shown to be present in the Inferim- Colliculus and the rate of firing of these neurms is a A Silicon Model of Amplitude Modulation Detection 743 band-pass functiro of the amplitude modulatiro rate. It is not known to date however if the choppers actually project to these coincidence detectoc neurons. The actual mechanism that synchrooizes the chopper cells will be discussed with the measurements in sectim 4. In the next sectim. we will first present the circuits that allowed us to build the VLSI implementation of this model. 3. THE CIRCUITS All of the circuits used in our model have already been presented in m(X'e detail elsewhere. but we will present them briefly f(X' completeness. Our silicro cochlea has been presented in detail at NlPS'95 [1]. and m(X'e details about the Inner Hair Cell circuit and the spiking neuron circuit can be found in [2]. 3.1 THE SILICON COCHLEA The silicro cochlea crosists of a cascade of secmd (X'der low-pass filters. Each filter sectiro is biased using Compatible Lateral Bipolar Transist(X's (Q.BTs) which crotrol the cut-off frequency and the quality fact(X' of each sectiro. A single resistive line is used to bias all Q.BTs. Because of the exponential relatiro between the Base-Emitter Voltage and the Collectoc current of the Q.BTs. the linear voltage gradient introduced by the resistive line will yield a filter cascade with an exponentially decreasing cut-off frequency of the filters. The output voltage of each filter Vout then represents the displacement of a basilar membrane sectiro. In (X'der to obtain a representatim of the basilar membrane velocity. we take the difference between Vout and the voltage m the internal node of the second order filter. We have integrated this silicro cochlea using 104 filter stages. and the output of every second stage is connected to an output pin. 3.2 THE INNER HAIR CELL MODEL The inner hair cell circuit is used to half-wave rectify the basilar membrane velocity signal and to perf(I'm some f(I'm of temp<X'al adaptatiro. as can be seen in figure 2b. The differential pair at the input is used to crovert the input voltage into a current with a compressive relatim between input amplitude and the actual amplitude of the current. ~Or---------------------~ ~1.5 11.0 ~ 0.5 0.0 .. _NIl ... ..., ...... ..,.,.. ................. o 5 10 15 20 25 30 35 tlma (me) Fig. 2. a) The Inner Hair Cell circuit. b) measured output current We have integrated a small chip containing 4 independent inner hair cells. 744 A. van Schaik, E. Fragniere and E. Vzttoz 3.3 THE SPIKING NEURON MODEL The spiking neuron circuit is given in figure 3. The membrane of a biological neuron is modeled by a capacitance. Cmem. and the membrane leakage current is coo.trolled by the gate voltage. VIeako of an NMOS transiskX'. In the absence of any input O"ex=O). the membrane voltage will be drawn to its resting potential (coo.trolled by V rest). by this leakage current. Excitatory inputs simply add charge to the membrane capacitance. whereas inhibitory inputs are simply modeled by a negative lex. If an excitatory current larger than the leakage current of the membrane is injected. the membrane potential will increase from its resting potential. This membrane potential. Vroom, is COOlpared with a coo.trollable threshold voltage V three. using a basic transconductance amplifier driving a high impedance load. If V mem exceeds V threa. an action potential will be generated. Fig. 3. The Spiking Neuron circuit The generation of the actioo potential happens in a similar way as in the biological neuron. where an increased sodium coo.ductance creates the upswing of the spike. and a delayed increase of the potassium coo.ductance creates the downswing. In the circuit this is modeled as follows. H V mam rises above Vthrea• the output voltage of the COOlparat<X" will rise to the positive power supply. The output of the following inverter will thus go low. thereby allowing the "sodium current" INa to pull Up the membrane potential. At the same time however. a second inverter will allow the capacitance CK to be charged at a speed which can be coo.trolled by the current ~p. As soon as the voltage on CK is high enough to allow coo.ductioo of the NMOS to which it is connected. the "potassium current" IK will be able to discharge the membrane capacitance. H V mam now drops below V threat the output of the first inverter will become high. cutting off the current INa. Furtherm<X"e. the second inverter will then allow CK to be discharged by the current IKdown. If IKdown is small, the voltage on CK will decrease only slowly. and. as loog as this voltage stays high enough to allow IK to discharge the membrane. it will be impossible to stimulate the neuron if lex is smaller than IK • Theref<X"e ~own can be said to control the 'refractory period' of the neuron. We have integrated a chip. coo.taining a group of 32 neurons. each having the same bias voltages and currents. The COOlponent mismatch and the noise ensure that we actually have 32 similar. but not completely equal neurons. 4. TEST RESULTS Most neuro-physiological data coo.cerning low frequency amplitude modulation of high frequency carriers exists f<X" carriers at about 5kHz and a modulation depth of about 50%. We theref<X"e used a 5 kHz sinusoid in our tests and a 50% modulatioo depth at frequencies below 550Hz. A Silicon Model of Amplitude Modulation Detection 745 250 250 200 1-'-1 200 i 150 j 150 100 1 100 50 50 0 0 0 10 20 30 40 50 0 10 20 30 40 50 Itn.~.) Itn.~.) Fig. 4. PSTH of the chopper chip for 2 different sound intensities First step in the elaboration of the model is to test if the group of spiking neurons on a single chip is capable of performing like a group of similar Choppers. Neurons in the auditory brainstem are often characterized with a Post Stimulus Time Histogram (pSTH), which is a histogram of spikes in response to repeated stimulatien with a pure tone of short duratien. If the choppers en the chip are really similar, the PSTH of this group of choppers will be very similar to the PSTH of a single chopper. In figure 4 the PSTH of the circuit is shown. It is the result of the summed response of the 32 neurens en chip to 20 repeated stimulatiens with a 5kHz tene burst. This figure shows that the response of the Choppers yields a PSTH typical of chopping neurens, and that the chopping frequency, keeping all other parameters constant, increases with increasing sound intensity. The chopping rate for an input signal of given intensity can be controlled by setting the refractory period of the spiking neurons, and can thus be used to create the different groups of choppers shown in figure 1. The chopping rate of the choppers in figure 4 is about 300Hz for a 29dB input signal. 1..,..···········,·_0.5 0 ••• -Merrbrane potential Stirrulation o < ::> (5 ::>(» 10 time (rrs) threshoki refractory charging spkt period period w iclth Fig. 5. Spike generation for a Chopper cell. To understand why the Choppers will synchronize fix' a certain amplitude modulatien frequency, one has to look at the signal envelope, which CCIltains tempocal information en a time scale that can influence the spiking neurens. The 5kHz carrier itself will not contain any tempocal informatien that influences the spiking neuren in an impoctant way. Consider the case when the modulation frequency is similar to the chopping frequency (figure 5). If a Chopper then spikes during the rising flank of the envelope, it will come out of its refractory period just before the next rising flank of the envelope. If the driven chopping frequency is a bit too low, the Chopper will come out of its refractory period a bit later, therefix'e it receives a higher average stimulation and it spikes a little higher en the rising flank of the envelope. This in turn increases the chopping frequency and thus provides a form of negative feedback on the chopping frequency. This theref(X'e makes spiking en a certain point en the rising flank of the envelope a stable situatien. With the same reasening one can show that spiking en the falling flank is theref(X'e an unstable situatien. Furthermore, it is not possible to stabilize a cell driven above its maximum chopping rate, n(X' is it possible to stabilize a cell that fires m(X'e than ence per modulatien period. Since a group of similar choppers will 746 A. van Schaik, E. Fragniere and E. Vittoz stabilize at about the same point on the rising flank. their spikes will thus coincide when the modulation frequency allows them to. 2SO ..... ···• ........... •• ...... • .. ·•• .. •• .. ·•••·• ... ·· .... • .. I==·[! 250200 ........................................ m ·· ... •· .. ··· .... i~4&;i3"l i:· 200 l o 100 200 300 400 500 600 modllllllDn .... (HI) modulldon .... (HI) Fig. 6. AM sensitivity of the coincidence detecting neuron. Another free parameter of the model is the threshold of the coincidence detecting neuron. If this parameter is set so that at least 60% of the choppers must spike within Ims to be considered a coincidence. we obtain the output of figure 6. We can see that this yields the expected band-pass Modulation Transfer Function (MTF). and that the best modulation frequency f(X' the 29dB input signal caTesponds to the intrinsic chopping rate of the group of neuroo.s. Figure 6 also shows that the best modulatioo. frequency (BMF). just as the chopping rate. increases with increasing sound intensity. but that the maximum number of spikes per second actually decreases. This second effect is caused by the fact that the stabilizing effect of the positive flank of the signal envelope only influences the time during which the neuron is being charged. which becomes a smaller part of the total spiking period at higher intensities. The negative feedback thus has less influence on the total chopping period and therefore synchronizes the choppers less. 25O~ 200~ 250 .••...•••...•..••...........•.........•.•..••.•••.••••.•.•..•..•.........•• ---.15.5c1! 200 f------' o 100 200 300 400 500 600 100 200 300 400 500 600 moduWlon .... (HI) modllllllDn .... (HE) a b Fig. 7. AM sensitivity of the coincidence detecting neuron. When the coincidence threshold is lowered to 50%. we can see in figure 7a that the maximum number of spikes goes up. because this threshold is m(X'e easily reached. Furthermore. a second pass-band shows up at double the BMF. This is because the choppers fire only every second amplitude modulation period. and part of the group of choppers will synchronize during the odd periods. whereas others during the even periods. The division of the group of choppers will typically be close to. but hardly ever exactly 50-50. so that either during the odd or during the even modulation period the 50% coincidence threshold is exceeded. The 60% threshold of figure 6 will only rarely be exceeded. explaining the weak second peak seen around 500Hz in this figure. A Silicon Model of Amplitude Modulation Detection 747 Figure 7b. shows the MIF f<X" low intensity signals with a SO% coincidence threshold. At low intensities the effect of an additional non-linearity, the stimulation threshold, shows up. Whenever the instantaneous value of the envelope is lower than the stimulatioo threshold, the spiking neuroo will not be stimulated because its input current will be lower than the cell's leakage current. At these low intensities the activity during the valleys of the modulatioo envelope will thus not be enough to stimulate the Choppers (see figure 5). F<X" stimuli with a lower modulatioo frequency than the group's chopping frequency, the Choppers will come out of their refract<X"y period in such a valley. These choppers theref<X"e will have to wait f<X" the envelope amplitude to increase above a certain value, bef<X"e they receive anew a stimulatioo. This waiting period nullifies the effect of the variatioo of the refractory period of the Choppers, and thus synchronizes the Choppers for low modulatioo frequencies. A secoo.d effect of this waiting period is that in this case the firing rate of the Choppers matches the modulation frequency. When the modulation frequency becomes higher than the maximum chopping frequency, the Choppers will fire only every secood period, but will still be synchronized, as can be seen between 300Hz and 500Hz in figure 7b. 5. CONCLUSIONS In this article we have shown that it is possible to use our building blocks to build a multi-chip system that models part of the audit(X"y pathway. Furtherm<X"e, the fact that the spiking neuron chip can be easily biased to function as a group of similar Choppers, combined with the relative simplicity of the spike generation mechanism of a single neuroo 00 chip, allowed us to gain insight in the process by which chopping neurons in the mammalian Cochlear Nucleus synchronize to a particular range of amplitude modulation frequencies. References [1] A. van Schaik, E. Fragniere, & E. Vittoz, "Improved silicm cochlea using compatible lateral bipolar transist<X"s," Advances in Neural Information Processing Systems 8, MIT Press, Cambridge, 1996. [2] A. van Schaik, E. Fragniere, & E. Vittoz, "An analoge electronic model of ventral cochlear nucleus neurons," Proc. Fifth Int. Con! on Microelectronics for Neural Networks and Fuzzy Systems, IEEE Computer Society Press, Los Alamitos, 1996, pp. S2-59. [3] A. van Schaik and R Meddis, "The electronic ear; towards a blueprint," Neurobiology, NATO ASI series, Plenum Press, New York, 1996. [4] MJ. Hewitt and R Meddis, "A computer model of amplitude-modulatioo sensitivity of single units in the inferior colliculus," 1. Acoust. Soc. Am., 9S, 1994, pp. I-IS. [S] MJ. Hewitt, R Meddis, & T.M. Shackletoo, "A computer model of a cochlearnucleus stellate cell: respooses to amplitude-modulated and pure tone stimuli," 1. Acoust. Soc. Am., 91. 1992, pp. 2096-2109. PART VI SPEECH, HANDWRITING AND SIGNAL PROCESSING
1996
5
1,247
GTM: A Principled Alternative to the Self-Organizing Map Christopher M. Bishop C.M .Bishop@aston.ac.uk Markus Svensen Christopher K. I. Williams svensjfm@aston.ac.uk C.K.r. Williams@aston.ac.uk Neural Computing Research Group Aston University, Birmingham, B4 7ET, UK http://www.ncrg.aston.ac.uk/ Abstract The Self-Organizing Map (SOM) algorithm has been extensively studied and has been applied with considerable success to a wide variety of problems. However, the algorithm is derived from heuristic ideas and this leads to a number of significant limitations. In this paper, we consider the problem of modelling the probability density of data in a space of several dimensions in terms of a smaller number of latent, or hidden, variables. We introduce a novel form of latent variable model, which we call the GTM algorithm (for Generative Topographic Mapping), which allows general non-linear transformations from latent space to data space, and which is trained using the EM (expectation-maximization) algorithm. Our approach overcomes the limitations of the SOM, while introducing no significant disadvantages. We demonstrate the performance of the GTM algorithm on simulated data from flow diagnostics for a multi-phase oil pipeline. 1 Introduction The Self-Organizing Map (SOM) algorithm of Kohonen (1982) represents a form of unsupervised learning in which a set of unlabelled data vectors tn (n = 1, ... , N) in a D-dimensional data space is summarized in terms of a set of reference vectors having a spatial organization corresponding (generally) to a two-dimensional sheet1 • IBiological metaphor is sometimes invoked when motivating the SOM procedure. It should be stressed that our goal here is not neuro-biological modelling, but rather the development of effective algorithms for data analysis. GTM: A PrincipledAlternative to the Self-Organizing Map 355 While this algorithm has achieved many successes in practical applications, it also suffers from some major deficiencies, many of which are highlighted in Kohonen (1995) and reviewed in this paper. From the perspective of statistical pattern recognition, a fundamental goal in unsupervised learning is to develop a representation of the distribution p(t) from which the data were generated. In this paper we consider the problem of modelling p( t) in terms of a number (usually two) of latent or hidden variables. By considering a particular class of such models we arrive at a formulation in terms of a constrained Gaussian mixture which can be trained using the EM (expectation-maximization) algorithm. The topographic nature of the representation is an intrinsic feature of the model and is not dependent on the details of the learning process. Our model defines a generative distribution p(t) and will be referred to as the GTM (Generative Topographic Mapping) algorithm (Bishop et al., 1996a). 2 Latent Variables The goal of a latent variable model is to find a representation for the distribution p( t) of data in a D-dimensional space t = (t 1 , ... , t D) in terms of a number L of latent variables x = (Xl, ... , XL). This is achieved by first considering a non-linear function y(x; W), governed by a set of parameters W, which maps points x in the latent space into corresponding points y(x; W) in the data space. Typically we are interested in the situation in which the dimensionality L of the latent space is less than the dimensionality D of the data space, since our premise is that the data itself has an intrinsic dimensionality which is less than D. The transformation y(x; W) then maps the latent space into an L-dimensional non-Euclidean manifold embedded within the data space. IT we define a probability distribution p(x) on the latent space, this will induce a corresponding distribution p(YIW) in the data space. We shall refer to p(x) as the prior distribution of x for reasons which will become clear shortly. Since L < D, the distribution in t-space would be confined to a manifold of dimension L and hence would be singular. Since in reality the data will only approximately live on a lower-dimensional manifold, it is appropriate to include a noise model for the t vector. We therefore define the distribution of t, for given x and W, to be a spherical Gaussian centred on y(x; W) having variance {3-1 so that p(tlx, W, {3) f"V N(tly(x; W),{3-1 I). The distribution in t-space, for a given value of W, is then obtained by integration over the x-distribution p(tIW,{3) = I p(tlx, W,{3)p(x) dx. (1) For a given a data set 1) = (t1 , •.. , t N) of N data points, we can determine the parameter matrix W, and the inverse variance {3, using maximum likelihood, where the log likelihood function is given by N L(W,{3) = L Inp(tnIW,{3). (2) n=l In principle we can now seek the maximum likelihood solution for the weight matrix, once we have specified the prior distribution p(x) and the functional form of the 356 C. M. Bishop, M. Svensen and C. K. I. Williams y(x;W) t) Q • • • ~ ~ • • • • • • t2 x) Figure 1: We consider a prior distribution p(x) consisting of a superposition of delta functions, located at the nodes of a regular grid in latent space. Each node XI is mapped to a point Y(XI; W) in data space, which forms the centre of the corresponding Gaussian distribution. mapping y(x; W), by maximizing L(W, f3). The latent variable model can be related to the Kohonen SOM algorithm by choosingp(x) to be a sum of delta functions centred on the nodes of a regular grid in latent space p(x) = 1/ K ~~l l5(x Xl). This form of p(x) allows the integral in (1) to be performed analytically. Each point Xl is then mapped to a corresponding point Y(XI; W) in data space, which forms the centre of a Gaussian density function, as illustrated in Figure 1. Thus the distribution function in data space takes the form of a Gaussian mixture model p(tIW, f3) = 1/ K ~~l p(tIXl, W, f3) and the log likelihood function (2) becomes N {I K } L(W,f3) = ~ In K ~P(tnIXI' W,f3) . (3) This distribution is a constrained Gaussian mixture since the centres of the Gaussians cannot move independently but are related through the function y(x; W). Note that, provided the mapping function y(x; W) is smooth and continuous, the projected points Y(XI; W) will necessarily have a topographic ordering. 2.1 The EM Algorithm IT we choose a particular parametrized form for y(x; W) which is a differentiable function of W we can use standard techniques for non-linear optimization, such as conjugate gradients or quasi-Newton methods, to find a weight matrix W*, and inverse variance f3*, which maximize L (W , f3). However, our model consists of a mixture distribution which suggests that we might seek an EM algorithm (Dempster et al., 1977). By making a careful choice of model y(x; W) we will see that the M-step can be solved exactly. In particular we shall choose y(x; W) to be given by a generalized linear network model of the form y(x; W) = W <fJ(x) (4) where the elements of <fJ(x) consist of M fixed basis functions <Pj (x), and W is a D x M matrix with elements Wkj. Generalized linear networks possess the same universal approximation capabilities as multi-layer adaptive networks, provided the basis functions <P j (x) are chosen appropriately. GTM: A Principled Alternative to the Self-Organizing Map 357 By setting the derivatives of (3) with respect to Wkj to zero, we obtain (5) where ~ is a K x M matrix with elements <P/j = <Pj(X/), T is a N x D matrix with elements tkn, and R is a K x N matrix with elements R'n given by (6) which represents the posterior probability, or responsibility, of the mixture components I for the data point n. Finally, G is a K x K diagonal matrix, with elements Gil = 'E:=l R'n(W,!3). Equation (5) can be solved for W using standard matrix inversion techniques. Similarly, optimizing with respect to {3 we obtain (7) Here (6) corresponds to the E-step, while (5) and (7) correspond to the M-step. Typically the EM algorithm gives satisfactory convergence after a few tens of cycles. An on-line version of this algorithm can be obtained by using the Robbins-Monro procedure to find a zero of the objective function gradient, or by using an on-line version of the EM algorithm. 3 Relation to the Self-Organizing Map The list below describes some of the problems with the SOM procedure and how the GTM algorithm solves them. 1. The SOM algorithm is not derived by optimizing an objective function, unlike GTM. Indeed it has been proven (Erwin et al., 1992) that such an objective function cannot exist for the SOM algorithm. 2. In GTM the neighbourhood-preserving nature of the mapping is an automatic consequence of the choice of a smooth, continuous function y(x; W). Neighbourhood-preservation is not guaranteed by the SOM procedure. 3. There is no assurance that the code-book vectors will converge using SOM. Convergence of the batch GTM algorithm is guaranteed by the EM algorithm, and the Robbins-Monro theorem provides a convergence proof for the on-line version. 4. GTM defines an explicit probability density function in data space. In contrast, SOM does not define a density model. Attempts have been made to interpret the density of codebook vectors as a model of the data distribution but with limited success. The advantages of having a density model include the ability to deal with missing data in a principled way, and the straightforward possibility of using a mixture of such models, again trained using EM. 358 C. M. Bishop, M. Svensen and C. K. I. Williams Figure 2: Examples of the posterior probabilities (responsibilities) of the latent space points at an early stage (left) and late stage (right) during the convergence of the GTM algorithm, evaluated for a single data point from the training set in the oil-flow problem discussed in Section 4. Note how the probabilities form a localized 'bubble' whose size shrinks automatically during training, in contrast to the hand-crafted shrinkage of the neighbourhood function in the SOM. 5. For SOM the choice of how the neighbourhood function should shrink over time during training is arbitrary, and so this must be optimized empirically. There is no neighbourhood function to select for GTM. 6. It is difficult to know by what criteria to compare different runs of the SOM procedure. For GTM one simply compares the likelihood of the data under the model, and standard statistical tests can be used for model comparison. Notwithstanding these key differences, there are very close similarities between the SOM and GTM techniques. Figure 2 shows the posterior probabilities (responsibilities) corresponding to the oil flow problem considered in Section 4. At an early stage of training the responsibility for representing a particular data point is spread over a relatively large region of the map. As the EM algorithm proceeds so this responsibility 'bubble' shrinks automatically. The responsibilities (computed in the E-step) govern the updating of Wand {3 in the M-step and, together with the smoothing effect of the basis functions <Pi (x), play an analogous role to the neighbourhood function in the SOM procedure. While the SOM neighbourhood function is arbitrary, however, the shrinking responsibility bubble in GTM arises directly from the EM algorithm. 4 Experimental Results We present results from the application of this algorithm to a problem involving 12-dimensional data arising from diagnostic measurements of oil flows along multiphase pipelines (Bishop and James, 1993). The three phases in the pipe (oil, water and gas) can belong to one of three different geometrical configurations, corresponding to stratified, homogeneous, and annular flows, and the data set consists of 1000 points drawn with equal probability from the 3 classes. We take the latent variable space to be two-dimensional, since our goal in this application is data visualization. Each data point tn induces a posterior distribution p(xltn, W,(3) in x-space. However, it is often convenient to project each data point down to a unique point in x-space, which can be done by finding the mean of the posterior distribution. GTM: A Principled Alternative to the Self-Organizing Map ........... x "">CI~* ...... )( • * :. · . . . ... \ i/r"', . . . . . .... )( .. .. .. .. .. .. .. .. x. .. .. .. .. .. .. .. .. .. .. .. .. )( .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ••••• x .'.;", . • •• . }; .. ''0.;;: · ........... • • .- t • ... . x ••• x ••••••• ., ••••• ~. .. • ••••••••••••• • • •• )( >U! + •• r .. . +.,', • · ........ +: + ...... )( I\. 1l ,. .....;I'.,-it: "':/ ..• .. ;. .... • • )( •••• ~+.,.~+ ~ .+. ~ .. • • • • •• • + + .... ~"", oIit ... ··· "' :t, x· •. ++~ r~",+:""",d'-"'+"·.' • • •• +*+ . ..... i+.t~ • .".. .. >,..: x ••• * Ii ...... _.;.#'''' .+-"'" ...... *ot++ . ...... .j., •••• *.*+ .~ .. • •• t: • r.; ....... t '0* *+ + • .;. ... )( •• i+iI' ..... **" ..... ~.Iji.; ••• • • + .. ++; ... <:1:.+ •••• 0 •• tt • •• ........ ' .... \ti"' ....... .,... ** ..... + ++ + .... ~ + l ++e • • • GOne t . t.. + x .. _ • . ;.>;:.,.: > •••••• o :CD 0 +::c:+* '* + • + + •• ~)4IpO •• ~ CIa # ... . * ~. CID •• qD. : .•• •• ~ •• W •• ~ i; ... + + • • • " • • • • .,. • • 6 . 1:+++ •• .. " "'~~_ClD. " ••• o •• p •• lb ;, .....•.. °.0. ~ • ~ . . . ...... • 0 " . Q)<P •• ~· .~. .:::91l • .... . ~~ ~ .... !:~ ~~" ......... ' .'~:.' o::~;~ . ,. : : ~; · J: ~'!;,· i • cf& •• ... 0. r# •... )(.)()( . . 0. 0: 0 '''0 ii .• • '.;': ··..:-.. '00:. 359 Figure 3: The left plot shows the posterior-mean projection of the oil flow data in the latent space of the non-linear model. The plot on the right shows the same data set visualized using the batch SOM procedure, in which each data point is assigned to the point on the feature map corresponding to the codebook vector to which it is nearest. In both plots, crosses, circles and plus-signs represent the three different oil-flow configurations. Figure 3 shows the oil data visualized with GTM and SaM. The CPU times taken for the GTM, SaM with a Gaussian neighbourhood, and SaM with a 'top-hat' neighbourhood were 644, 1116 and 355 seconds respectively. In each case the algorithms were run for 25 complete passes through the data set. 5 Discussion In the fifteen years since the SaM procedure was first proposed, it has been used with great success in a wide variety of applications. It is, however, based on heuristic concepts, rather than statistical principles, and this leads to a number of serious deficiencies in the algorithm. There have been several attempts to provide algorithms which are similar in spirit to the SaM but which overcome its limitations, and it is useful to compare these to the GTM algorithm. The formulation of the elastic net algorithm described by Durbin et al. (1989) also constitutes a Gaussian mixture model in which the Gaussian centres acquire a spatial ordering during training. The principal difference compared with GTM is that in the elastic net the centres are independent parameters but are encouraged to be spatially close by the use of a quadratic regularization term, whereas in GTM there is no regularizer on the centres and instead the centres are constrained to lie on a manifold given by the non-linear projection of the latent-variable space. The existence of a well-defined manifold means that the local magnification factors can be evaluated explicitly as continuous functions of the latent variables (Bishop et al., 1996b). By contrast, in algorithms such as SaM and the elastic net, the embedded manifold is defined only indirectly as a discrete approximation by the locations of the code-book vectors or Gaussian centres. One version of the principal curves algorithm (Tibshirani, 1992) introduces a generative distribution based on a mixture of Gaussians, with a well-defined likelihood function, which is trained by the EM algorithm. However, the number of Gaussian components is equal to the number of data points, and the algorithm has been 360 C. M. Bishop, M. Svensen and C. K. I. Williams formulated for one-dimensional manifolds, making this algorithm further removed from the SOM. MacKay (1995) considers convolutional models of the form (1) using multi-layer network models in which a discrete sample from the latent space is interpreted as a Monte Carlo approximation to the integration over a continuous distribution. Although an EM approach could be applied to such a model, the M-step of the corresponding EM algorithm would itself require a non-linear optimization. In conclusion, we have provided an alternative algorithm to the SOM which overcomes its principal deficiencies while retaining its general characteristics. We know of no significant disadvantage in using the GTM algorithm in place of the SOM. While we believe the SOM procedure is superseded by the GTM algorithm, is should be noted that the SOM has provided much of the inspiration for developing GTM. A web site for GTM is provided at: http://www.ncrg.aston.ac.uk/GTM/ which includes postscript files of relevant papers, software implementations in Matlab and C, and example data sets used in the development of the GTM algorithm. Acknowledgements This work was supported by EPSRC grant GR/K51808: Neural Networks for Visualisation of High-Dimensional Data. Markus Svensen would like to thank the staff of the SANS group in Stockholm for their hospitality during part of this project. References Bishop, C. M. and G. D. James (1993). Analysis of multiphase flows using dual-energy gamma densitometry and neural networks. Nuclear Instruments and Methods in Physics Research A327, 580-593. Bishop, C. M., M. Svensen, and C. K. I. Williams (1996a). Gtm: The generative topographic mapping. Technical Report NCRG/96/015, Neural Computing Research Group, Aston University, Birmingham, UK. Submitted to Neural Computation. Bishop, C. M., M. Svensen, and C. K. I. Williams (1996b). Magnification factors for the GTM algorithm. In preparation. Dempster, A. P., N. M. Laird, and D. B. Rubin (1977). Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, B 39 (1), 1-38. Durbin, R., R. Szeliski, and A. Yuille {1989}. An analysis of the elastic net approach to the travelling salesman problem. Neural Computation 1 (3),348-358. Erwin, E., K. Obermayer, and K. Schulten (1992). Self-organizing maps: ordering, convergence properties and energy functions. Biological Cybernetics 67,47-55. Kohonen, T. (1982). Self-organized formation of topologically correct feature maps. Biological Cybernetics 43, 59-69. Kohonen, T. (1995). Self-Organizing Maps. Berlin: Springer-Verlag. MacKay, D. J. C. (1995). Bayesian neural networks and density networks. Nuclear Instruments and Methods in Physics Research, A 354 (1), 73-80. Tibshirani, R. (1992). Principal curves revisited. Statistics and Computing 2, 183-190.
1996
50
1,248
Support Vector Method for Function Approximation, Regression Estimation, and Signal Processing· Vladimir Vapnik AT&T Research 101 Crawfords Corner Holmdel, N J 07733 vlad@research.att.com Steven E. Golowich Bell Laboratories 700 Mountain Ave. Murray Hill, NJ 07974 golowich@bell-Iabs.com Abstract Alex Smola· GMD First Rudower Shausee 5 12489 Berlin asm@big.att.com The Support Vector (SV) method was recently proposed for estimating regressions, constructing multidimensional splines, and solving linear operator equations [Vapnik, 1995]. In this presentation we report results of applying the SV method to these problems. 1 Introduction The Support Vector method is a universal tool for solving multidimensional function estimation problems. Initially it was designed to solve pattern recognition problems, where in order to find a decision rule with good generalization ability one selects some (small) subset of the training data, called the Support Vectors (SVs). Optimal separation of the SV s is equivalent to optimal separation the entire data. This led to a new method of representing decision functions where the decision functions are a linear expansion on a basis whose elements are nonlinear functions parameterized by the SVs (we need one SV for each element of the basis). This type of function representation is especially useful for high dimensional input space: the number of free parameters in this representation is equal to the number of SVs but does not depend on the dimensionality of the space. Later the SV method was extended to real-valued functions. This allows us to expand high-dimensional functions using a small basis constructed from SVs. This ·smola@prosun.first.gmd.de 282 v. Vapnik, S. E. Golowich and A. Smola novel type of function representation opens new opportunities for solving various problems of function approximation and estimation. In this paper we demonstrate that using the SV technique one can solve problems that in classical techniques would require estimating a large number of free parameters. In particular we construct one and two dimensional splines with an arbitrary number of grid points. Using linear splines we approximate non-linear functions. We show that by reducing requirements on the accuracy of approximation, one decreases the number of SVs which leads to data compression. We also show that the SV technique is a useful tool for regression estimation. Lastly we demonstrate that using the SV function representation for solving inverse ill-posed problems provides an additional opportunity for regularization. 2 SV method for estimation of real functions Let x E Rn and Y E Rl. Consider the following set of real functions: a vector x is mapped into some a priori chosen Hilbert space, where we define functions that are linear in their parameters 00 Y = I(x,w) = L Wi<Pi(X), W = (WI, ... ,WN, ... ) E n (1) i=1 In [Vapnik, 1995] the following method for estimating functions in the set (1) based on training data (Xl, Yd, ... , (Xl, Yl) was suggested: find the function that minimizes the following functional: 1 l R(w) = £ L IYi - I(Xi, w)lt: + I(w, w), (2) i=1 where { 0 if Iy - I(x, w)1 < £, Iy - I(x, w)lt: = Iy - I(x, w)l- £ otherwise, (3) (w, w) is the inner product of two vectors, and I is some constant. It was shown that the function minimizing this functional has a form: l I(x, a, a*) = L(a; - ai)(<I>(xi), <I>(x)) + b (4) ;=1 where ai, ai 2:: 0 with aiai = 0 and (<I>(Xi), <I>(x» is the inner product of two elements of Hilbert space. To find the coefficients a; and ai one has to solve the following quadratic optimization problem: maximize the functional i l l W(a*, a) = -£ L(a; +ai)+ Ly(a; -ai)-~ L (a; -ai)(aj -aj )(<I>(Xi), <I>(Xj)), i=1 ;=1 i,j=1 subject to constraints l L(ai-ai)=O, O~ai,a;~C, i=l, ... ,f. i=1 (5) (6) SV Method for Function Approximation and Regression Estimation 283 The important feature of the solution (4) of this optimization problem is that only some of the coefficients (a; - ai) differ from zero. The corresponding vectors Xi are called Support Vectors (SVs). Therefore (4) describes an expansion on SVs. It was shown in [Vapnik, 1995] that to evaluate the inner products (<1>(Xi)' <1>(x)) both in expansion (4) and in the objective function (5) one can use the general form of the inner product in Hilbert space. According to Hilbert space theory, to guarantee that a symmetric function K ( u, v) has an expansion 00 K(u, v) = L ak1fJk(u)tPk(V) k=l with positive coefficients ak > 0, i.e. to guarantee that K (u, v) is an inner product in some feature space <1>, it is necessary and sufficient that the conditions J K(u, v)g(u)g(v) du dv > 0 be valid for any non-zero function 9 on the Hilbert space (Mercer's theorem). Therefore, in the SV method, one can replace (4) with l I(x, a, a*) = L(a; - ai)K(x, Xi) + b i=l (7) (8) where the inner product (<1>( Xi), <1>( x» is defined through a kernel K (Xi, x). To find coefficients ai and ai one has to maximize the function l l l W(a*, a) = -[ L(a; +ai)+ Ly(a; -ai)- ~ L (a; -ai)(aj -aj)K(xi, Xj) (9) i=l i=l i,j=l subject to constraints (6). 3 Constructing kernels for inner products To define a set of approximating functions one has to define a kernel K (Xi, X) that generates the inner product in some feature space and solve the corresponding quadratic optimization problem. 3.1 Kernels generating splines We start with the spline functions. According to their definition, splines are piecewise polynomial functions, which we will consider on the set [0,1]. Splines of order n have the following representation n N In(x) = L arxr + L Wj(x t~r~. (10) r=O ~=l where (x - t)+ = max{(x - t), O}, tl, ... , tN E [0,1] are the nodes, and ar , Wj are real values. One can consider the spline function (10) as a linear function in the n + N + 1 dimensional feature space spanned by 1, x, ... , xn, (x - tdf., ... , (x - tN)f.. 284 V. Vapnik, S. E. Golowich and A. Smola Therefore the inner product that generates splines of order n in one dimension is n N I«Xi,Xj) = Lx;xj + L(Xi -t3)~(Xj -t3)~' (11) r=O 3=1 Two dimensional splines are linear functions in the (N + n + 1)2 dimensional space 1, x, ... , xn, y, ... , yn, ... , (x td~(y t~)~, ... , (x - tN )~(y - tN )~. (12) Let us denote by Ui = (Xi ,Yi), Uj = (Xi,Yj) two two-dimensional vectors. Then the generating kernel for two dimensional spline functions of order n is It is easy to check that the generating kernel for the m-dimensional splines is the product of m one-dimensional generating kernels. In applications of the SV method the number of nodes does not play an important role. Therefore, we introduce splines of order d with an infinite number of nodes S~oo). To do this in the R1 case, we map any real value Xi to the element 1, Xi, ... , xi, (Xi - t)+ of the Hilbert space. The inner product becomes n 1 I«Xi,Xj) = Lx;xj+ 1 (Xi-t)~(Xj -t)~dt r=O 0 (13) For linear splines S~oo) we therefore have the following generating kernel: In many applications expansions in Bn-splines [Unser & Aldroubi, 1992] are used, where Bn(x) = E (-~y ( n + 1 ) (X + n + 1 _ r)n . n. r 2 + r=O One may use Bn-splines to perform a construction similar to the above, yielding the kernel 3.2 Kernels generating Fourier expansions Lastly, Fourier expansion can be considered as a hyperplane in following 2N + 1 dimensional feature space 1 . N' N V2 ' cos x, sln x, ... , cos x, sln x. The inner product in this space is defined by the Dirichlet formula: SV Method for Function Approximation and Regression Estimation 285 4 Function estimation and data compression In this section we approximate functions on the basis of observations at f points (16) We demonstrate that to construct an approximation within an accuracy of ±c at the data points, one can use only the subsequence of the data containing the SVs. We consider approximating the one and two dimensional functions . smlxl f(x) = smclxl = -I-xl(17) on the basis of a sequence of measurements (without noise) on the uniform lattice (100 for the one dimensional case and 2,500 for the two-dimensional case). For different c we approximate this function by linear splines from si 00) . Figure 1: Approximations with different levels of accuracy require different numbers ofSV: 31 SV for c = 0.02 (left) and 9 SV for c = 0.1. Large dots indicate SVs . os ..,.. .. -.. D • +;. .... .. . . -+ ••••• : ."'.4\ , •• ~. • • -.: ••••• + •• .. .... " ...... . •• ,.... +: :. .+ .+ .... ,+ ::, •••• Figure 2: Approximation of f( x, y) sinc vi x2 + y2 by two dimensional linear splines with accuracy c = 0.01 (left) required 157 SV (right) .. o o 0 of> • ~ • • .0 • 0 • • ~ . • • • Figure 3: sincx function corrupted by different levels of noise «(7 = 0.2 left, 0.5 right) and its regression. Black dots indicate SV, circles non-SV data. 286 V. Vapnik, S. E. Golowich and A. Smola 5 Solution of the linear operator equations In this section we consider the problem of solving linear equations in the set of functions defined by SVs. Consider the problem of solving a linear operator equation Af(t) = F(x), f(t) E 2, F(x) E W, (18) where we are given measurements of the right hand side (Xl, FI ), ... , (Xl, Fl). (19) Consider the set of functions f(t, w) E 2 linear in some feature space {<I>(t) = (¢>o(t), ... , ¢>N(t), ... )}: 00 f(t, w) = L wr¢>r(t) = (W, <I>(t» . (20) r=O The operator A maps this set of functions into 00 00 F(x, w) = Af(t, w) = L wrA¢>r(t) = L wrtPr(x) = (W, w(x» (21) r=O r=O where tPr(x) = A¢>r(t), w(x) = (tPl(X), ... , tPN(X), ... ). Let us define the generating kernel in image space 00 K(Xi, Xj) = L tPr(Xi)tPr(Xj) = (W(Xi)' W(Xj» (22) r=O and the corresponding cross-kernel function 00 K,(Xi' t) = L tPr(xd¢>r(t) = (W(Xi), <I>(t». (23) r=O The problem of solving (18) in the set of functions f(t, w) E 2 (finding the vector W) is equivalent to the problem of regression estimation (21) using data (19). To estimate the regression on the basis of the kernel K(Xi, Xj) one can use the methods described in Section 1. The obtained parameters (a; - ai, i = 1, ... f) define the approximation to the solution of equation (18) based on data (19): l f(t, a) = L(ai - ai)K,(xi, t). i=l We have applied this method to solution of the Radon equation j aCm) f( m cos tt + u sin tt, m sin tt - u cos tt )du = p( m, tt), -a(m) -1 ~ m ~ 1, 0 < tt < 11", a(m) = -/1 - m 2 (24) using noisy observations (ml' ttl, pd, ... , (ml' ttl, Pi)' where Pi = p( mi, tti) + ~i and {ed are independent with Eei = 0, Eel < 00. sv Method for Function Approximation and Regression Estimation 287 For two-dimensional linear splines S~ 00) we obtained analytical expressions for the 'kernel (22) and cross-kernel (23). We have used these kernels for solving the corresponding regression problem and reconstructing images based on data that is similar to what one might get from a Positron Emission Tomography scan [Shepp, Vardi & Kaufman, 1985]. A remarkable feature of this solution is that it aVOIds a pixel representation of the function which would require the estimation of 10,000 to 60,000 parameters. The spline approximation shown here required only 172 SVs. Figure 4: Original image (dashed line) and its reconstruction (solid line) from 2,048 observations (left). 172 SVs (support lines) were used in the reconstruction (right). 6 Conclusion In this article we present a new method of function estimation that is especially useful for solving multi-dimensional problems. The complexity of the solution of the function estimation problem using the SV representation depends on the complexity of the desired solution (i.e. on the required number of SVs for a reasonable approximation of the desired function) rather than on the dimensionality of the space. Using the SV method one can solve various problems of function estimation both in statistics and in applied mathematics. Acknowledgments We would like to thank Chris Burges (Lucent Technologies) and Bernhard Scholkopf (MPIK Tiibingen) for help with the code and useful discussions. This work was supported in part by NSF grant PHY 95-12729 (Steven Golowich) and by ARPA grant N00014-94-C-0186 and the German National Scholarship Foundation (Alex Smola). References 1. Vladimir Vapnik, "The Nature of Statistical Learning Theory", 1995, Springer Verlag N.Y., 189 p. 2. Michael Unser and Akram Aldroubi, "Polynomial Splines and Wevelets - A Signal Perspectives", In the book: "Wavelets -A tutorial in Theory and Applications" , C.K. Chui (ed) pp. 91 - 122, 1992 Academic Press, Inc. 3. 1. Shepp, Y. Vardi, and L. Kaufman, "A statistical model for Positron Emission Tomography," J. Amer. Stat. Assoc. 80:389 pp. 8-37 1985.
1996
51
1,249
Smoothing Regularizers for Projective Basis Function Networks John E. Moody and Thorsteinn S. Rognvaldsson * Department of Computer Science, Oregon Graduate Institute PO Box 91000, Portland, OR 97291 moody@cse.ogi.edu denni@cca.hh.se Abstract Smoothing regularizers for radial basis functions have been studied extensively, but no general smoothing regularizers for projective basis junctions (PBFs), such as the widely-used sigmoidal PBFs, have heretofore been proposed. We derive new classes of algebraically-simple mH'-order smoothing regularizers for networks of the form f(W, x) = L7=1 Ujg [xT Vj + Vjol + uo, with general projective basis functions g[.]. These regularizers are: N Ra(W,m) = LU;lIvjIl2m-1 GlobalForm j=1 N RdW,m) = LU;lIvjll2m Local Form j=1 These regularizers bound the corresponding m t" -order smoothing integral where W denotes all the network weights {Uj, uo, vj, vo}, and O(x) is a weighting function on the D-dimensional input space. The global and local cases are distinguished by different choices of O( x). The simple algebraic forms R(W, m) enable the direct enforcement of smoothness without the need for costly Monte-Carlo integrations of S (W, m). The new regularizers are shown to yield better generalization errors than weight decay when the implicit assumptions in the latter are wrong. Unlike weight decay, the new regularizers distinguish between the roles of the input and output weights and capture the interactions between them. • Address as of September I. 1996: Centre for Computer Architecture. University of Halmstad, P.O.Box 823, S-301 18 Halmstad, Sweden 586 1. E. Moody and T. S. Rognvaldsson 1 Introduction: What are the right biases? Regularization is a technique for reducing prediction risk by balancing model bias and model variance. A regularizer R(W) imposes prior constraints on the network parameters W. Using squared error as the most common example, the objective functional that is minimized during training is M E = 2~ ?=[y(i) - I(W, x(i) W + AR(W) , .=1 (1) where y(;) are target values corresponding to the inputs x(;) , M is the number of training patterns, and the regularization parameter A controls the importance of the prior constraints relative to the fit to the data. Several approaches can be applied to estimate A (e.g. Eubank (1988) or Wahba (1990». Regularization reduces model variance at the cost of some model bias. An important question arises: What are the right biases? (Geman, Bienenstock & Doursat 1992). A good choice of R(W) will result in lower expected prediction error than will a poor choice. Weight decay is often used effectively, but it is an ad hoc technique that controls weight values without regard to the function 1(' ). It is thus not necessarily optimal and not appropriate for arbitrary function parameterizations. It will give very different results, depending upon whether a function is parameterized, for example, as f (w, x) or as I ( w -I, x). Since many real world problems are intrinsically smooth, we propose that in many cases, an appropriate bias to impose is to favor solutions with low mth-order curvature. Direct penalization of curvature is a parametrization-independent approach. The desired regularizer is the standard D dimensional curvature functional of order m: (2) Here 1111 denotes the ordinary euclidean tensor norm and am /axm denotes the mth order differential operator. The weighting function n (x) ensures that the integral converges and detennines the region over which we require the function to be smooth. n( x) is not required to be equal to the input density p( x), and will most often be different. The use of smoothing functionals like (2) has been extensively studied for smoothing splines (Eubank 1988, Hastie & Tibshirani 1990, Wahba 1990) and for radial basis function (RBF) networks (powell 1987, Poggio & Girosi 1990, Girosi, Jones & Poggio 1995). However, no general class of smoothing regularizers that directly enforce smoothness sew, m) for projective basis junctions (PBFs), such as the widely used sigmoidal PBFs, has been previously proposed. Since explicit enforcement of smoothness using (2) requires costly, impractical MonteCarlo integrations, 1 we derive algebraically-simple regularizers R(W, m) that tightly bound S(W,m). 2 Derivation of Simple Regularizers from Smoothing Functionals We consider single hidden layer networks with D input variables, Nh nonlinear hidden units, and No linear output units. For clarity, we set No = 1, and drop the subscript on Nh 'Note that (2) is not just one integral, but actually O(Dm) integrals, since the norm of the operator 8 m /8xm has O(Dm) terms. This is extremely expensive to compute for large D or large m . Smoothing RegularizersJor Projective Basis Function Networks (the derivation is trivially extended to the case No > 1). Thus, our network function is N j(x) = L ujg[9j, x] + Uo j=1 587 (3) where g[.] are the nonlinear transfer functions of the internal hidden units, x E RD is the input vector , 9 j are the parameters associated with internal unit j, and W denotes all parameters in the network. For regularizers R(W), we wiIl derive strict upper bounds for S(W, m). We desire the regularizers to be as general as possible so that they can easily be applied to different network models. Without making any assumptions about n(x) or g(.), we have the upper bound S(W,m) ~ Nt. u; J dDx{l( .. ) Iliim~~~,"1 II'(4) which follows from the inequality (L~I a;)2 $ NL~I ar. We consider two possible options for the weighting function n(x). One is to require global smoothness, in which case n(x) is a very wide function that covers all relevant parts of the input space (e.g. a very wide gaussian distribution or a constant distribution). The other option is to require local smoothness, in which case n( x) approaches zero outside small regions around some reference points (e.g. the training data). 2.1 Projective Basis Representations Projective basis functions (PBFs) are of the form g[9j , x] = 9 [xT Vj + Vjol , where 9j = {Vj, Vjo}, Vj = (Vj), Vj2, ••• ,VjD) is the vector of weights connecting hidden unit j to the inp.uts, and VjO is the bias, offset, or threshold. For PBFs, expression (4) simplifies to N S(W,m) $ NL u;llvjIl2mlj(W,m), (5) j=) with I;(W,m) " J dDx{l( .. ) (dm~;t)l) 2 (6) where Zj(x) = x T Vj + Vjo. Although the most commonly used g[.]' s are sigmoids, our analysis applies to many other forms, for example flexible fourier units, polynomials, andfational functions.3 The classes of PBF transfer functions g[.] that are applicable (as determined by n(x» are those for which the integral (8) is finite and well-defined. 2.2 Global weighting For the global case, we select a gaussian form for the weighting function na(x) = (.Ji;u)-D exp [ _~:~12] 2Throughout, we use smaliletler boldface to denote vector quantities. 3See for example Moody & Yarvin (1992). (7) 588 1. E. Moody and T S. Rognvaldsson and require (J to be large. Integrating out all dimensions, except the one associated with the projection vector U j, we are left with (8) If (dmg[z]/dzm)2 is integrable and approaches zero outside a region that is small compared to (J, we can bound (8) by setting the exponential equal to unity. This implies I(m). _ 1 1 00 (dmg[z])2 Ij(W,m) =:; IIUjll WIth I(m) = (J.,fi; -00 dz dzm (9) The bound of equation (5) then becomes N S(W,m) =:; NI(m) LU;IIUjWm - 1 = NI(m)Re(W,m) , (10) j;l where the subscript G stands for global. Since A absorbs all constant multiplicative factors, we need only weigh Re(W, m) into the training objective function. 2.2.1 Local weighting For the local case, we consider weighting functions of the general form (11) where x(i) are a set of points, and n( xU) , (J) is a function that decays rapidly for large II x - x(i) II. We require that limu-to n( x(i) , (J) = 6( x - xCi) ). Thus, when the xCi) are the training data points, the limiting distribution of (11) is the empirical distribution. In the limit (J -+ 0, equation (5) becomes (12) For the empirical distribution, we could compute the expression within parenthesis in (12) for each input pattern x( i) during training and use it as our regularization cost. This is done by Bishop (1993) for the special case m = 2. However, this requires explicit design for each transfer function and becomes increasingly complicated as we go to higher m. To construct a simpler and more general form, we instead assume that the m th derivative of g[.] isboundedfromabovebyCL(m) == maxz (d;z9lZlf This gives the bound N S(W,m) =:; NCL(m) L u;lIujWm = NCL(m)RL(W,m) (13) j;l for the maximum local curvature of the function (the subscript L denotes local limit). Smoothing RegularizersJor Projective Basis Function Networks 589 3 Empirical Example We have done extensive simulation studies that demonstrate the efficacy of our new regularizers for PBF networks on a variety of problems. An account is given in Moody & Rognvaldsson (1996). Here, we demonstrate the value of using smoothing regularizers on a simple problem which illustrates a key difference between smoothing and quadratic weight decay, the two dimensional bilinear function (14) This example was used by Friedman & Stuetzle (1981) to demonstrate projection pursuit regression. It is the simplest function with interactions between input variables. We fit this function with one hidden layer networks using the m = {I, 2, 3} smoothing regularizers, comparing the results with using weight decay. In a large set of experiments, we find that both the global and local smoothing regularizers with m = 2 and m = 3 outperform weight decay. An example is shown in figure 1. The local m = 1 case performs poorly, which is unsurprising, given that the target function is quadratic. Weight decay performs poorly because it lacks any form of interaction between the input layer and output layer weights v j and U j. ~ 1,8 I J O.t (a)GIoboI _al_ontOfdor .... N ......... ·.~ ·.·)· , , Otobol_. mot Otobol_.rn.2 - - Otobol_. ""'" - '~~ ... ,"\ ~\ '. \ \:" I g . I I O.t I J 1 -I(b) Weight doc<o)' va. gIobaJ _ 'I , 11 ,l '1 , H '1-I' Figure 1: (8) Generalization errors on the XtX2 problem, with 40 training data points and a signal-to-noise ratio of2l3, for different values of the regularization parameter and different orders of the smoothing regularizer. For each value of A, 10 networks with 8 hidden units' have been trained and averaged (geometric average). The shaded area shows the 95% confidence bands for the average performance of a linear model on the same problem. (b) Similar plot for the m = 3 smoother compared to the standard weight decay method. Error bars mark the estimated standard deviation of the mean generalization error of the 10 networks. The m = 3 regularizer performs significantly better than weight decay. 4 Quality of the Regularizers: Approximations vs Bounds Equations (10) and (13) are strict upper bounds to the smoothness functional S(W, m), eq. (2), in the global and local limits, U -+ 00 and U -+ O. However, if the bounds are not sufficiently tight, then penalizing R(W, m) may not have the effect of penalizing S (W, m )4 . 4For the proposed regularizers R(W, m) to be effective in penalizing S(W, m), we need only have an approximate monotonic relationship between them. 590 1. E. Moody and T. S. Rognvaldsson The bound (4) is tighter the more uncorrelated the mth derivatives of the internal unit activities are. If they are uncorrelated, then the bounds of equations (10) and (13) can be replaced by the approximations: SG(W,m) SL(w,m) ~ IG(m)RG(W, m) ~ CL(m)RL(W,m) , (15) (16) using (~~I a j ) 2 ~ ~~I a;. The right hand sides differ from those in equations (10) and (13) only by a factor of N, so these approximations are actually proportional to the bounds. For our regularizers, the constant factor N doesn't matter, since it can be absorbed into the regularization parameter A (along with the values of the factors IG(m) or CL(m» . In practical terms, there is no difference between using the upper bounds (10) and (13) or the uncorrelated approximations (15) and (16). Our empirical results (see figure 2) indicate that an approximate linear relationship holds between S(W, m) and R(W; m) for both the global and the local cases. This suggests that the uncorrelated hidden unit assumption yields a good approximation. This approximation also improves with the dimensionality of the input space. Extensive results and discussion are presented in (Moody & Rognvaldsson 1996). Pi' + !!!. ~ if Ii -8 Ct ~ • " Ii > ,. FIlS! crde! global SII"IOOIhe!, projection-based tam tI1~S Lineal regression (COIIelation: 0.992) --12 t,, ' t 10 ,-, t .~ • / t 8 6 • 2 , .. / 0 " 0 1 2 3 • 5 6 Monte Carlo eslimated value a S(W,I) [E +3) ~ + !!!. N ~ if Ii -8 Ct ~ • " Ii > 7 ~ orde! global smoother, projection-based tanh unls Unearregression (correlation: 0.998) ._6 5 • tf"· .,' 3 .' O· 0 0.5 1.0 1.5 2.0 2.5 '-bie Carlo eslimaled value a S(W,2) [E+6) Figure 2: Linear correlation between S(W; m) and the global RG(W, m) for neural networks with 10 input units, 10 internal tanh[·] PBF units, and one linear output. The values of S(W; m) are computed through Monte Carlo integration. The left graph shows m = 1 and the right graph shows m = 2. Results are similar for the local form R L (W; m). 5 Summary Our regularizers R(W; m) are the first general class of mth-order smoothing regularizers to be proposed for projective basis function (PBF) networks. They apply to large classes of transfer functions g[.], including sigmoids. They differ fundamentally from quadratic weight decay in that they distinguish the roles of the input and output weights and capture the interactions between them. Our approach is quite different from that developed for smoothing splines and smoothing radial basis functions (RBFs), since we derive smoothing regularizers for given classes of units g[8, x], rather than derive the forms of the units g[.] by requiring them to be Greens functions of the smoothing operator S(·). Our approach thus has the advantage that it can be applied to the types of networks most often used in practice, namely PBFs. Smoothing Regularizersfor Projective Basis Function Networks 591 In Moody & Rognvaldsson (1996), we present further analysis and simulation results for PBFs. We have also extended our work to RBFs (Moody & Rognvaldsson 1997). Acknowledgements Both authors thank Steve Rehfuss and Lizhong Wu for stimulating input. John Moody thanks Volker Tresp for a provocative discussion at a 1991 Neural Networks Workshop sponsored by the Deutsche Informatik Akademie. We gratefully acknowledge support for this work from ARPA and ONR (grant NOOO 14-92-J-4062), NSF (grant CDA-9503968), the Swedish Institute, and the Swedish Research Council for Engineering Sciences (contract TFR-282-95-847). References Bishop, C. (1993), 'Curvature-driven smoothing: A learning algorithm for feed forward networks', IEEE Trans. Neural Networks 4, 882-884. Eubank, R. L. (1988), Spline Smoothing and Nonparametric Regression, Marcel Dekker, Inc. Friedman, J. H. & Stuetzle, W. (1981), 'Projection pursuit regression', J. Amer. Stat. Assoc. 76(376),817-823. Geman, S., Bienenstock, E. & Doursat, R. (1992), 'Neural networks and the bias/variance dilemma', Neural Computation 4(1), 1-58. Girosi, E, Jones, M. & Poggio, T. (1995), 'Regularization theory and neural network architectures', Neural Computation 7,219-269. Hastie, T. 1. & libshirani, R. 1. (1990). Generalized Additive Models, Vol. 43 of Monographs on Statistics and Applied Probability, Chapman and Hall. Moody, 1. E. & Yarvin, N. (1992), Networks with learned unit response functions, in J. E. Moody, S. J. Hanson & R. P. Lippmann, eds. 'Advances in Neural Information Processing Systems 4', Morgan Kaufmann Publishers, San Mateo. CA. pp. 1048-55. Moody, J. & Rognvaldsson. T. (1996), Smoothing regularizers for projective basis function networks. Submitted to Neural Computation. Moody, 1. & Rognvaldsson, T. (1997), Smoothing regularizers for radial basis function networks, Manuscript in preparation. Poggio, T. & Girosi, E (1990). 'Networks for approximation and learning', IEEE Proceedings 78(9). Powell, M. (1987). Radial basis functions for multivariable interpolation: a review .• in 1. Mason & M. Cox, eds, 'Algorithms for Approximation', Clarendon Press, Oxford. Wahba, G. (1990), Spline models for observational data, CBMS-NSF Regional Conference Series in Applied Mathematics.
1996
52
1,250
Bangs. Clicks, Snaps, Thuds and Whacks: an Architecture for Acoustic Transient Processing Fernando J. Pineda(l) fernando. pineda@jhuapl.edu Gert Cauwenberghs(2) gert@jhunix.hcf.jhu.edu R. Timothy Edwards(2) tim@bach.ece.jhu.edu (iThe Applied Physics Laboratory The Johns Hopkins University Laurel, Maryland 20723-6099 (2Dept. of Electrical and Computer Engineering The Johns Hopkins University 34th and Charles Streets Baltimore Maryland 21218 ABSTRACT We propose a neuromorphic architecture for real-time processing of acoustic transients in analog VLSI. We show how judicious normalization of a time-frequency signal allows an elegant and robust implementation of a correlation algorithm. The algorithm uses binary multiplexing instead of analog-analog multiplication. This removes the need for analog storage and analog-multiplication. Simulations show that the resulting algorithm has the same out-of-sample classification performance (-93% correct) as a baseline template-matching algorithm. 1 INTRODUCTION We report progress towards our long-term goal of developing low-cost, low-power, lowcomplexity analog-VLSI processors for real-time applications. We propose a neuromorphic architecture for acoustic processing in analog VLSI. The characteristics of the architecture are explored by using simulations and real-world acoustic transients. We use acoustic transients in our experiments because information in the form of acoustic transients pervades the natural world. Insects, birds, and mammals (especially marine mammals) all employ acoustic signals with rich transient structure. Human speech, is largely composed of transients and speech recognizers based on transients can perform as well as recognizers based on phonemes (Morgan, Bourlard,Greenberg, Hermansky, and Wu, 1995). Machines also generate transients as they change state and as they wear down. Transients can be used to diagnose wear and abnormal conditions in machines. Architecture for Acoustic Transient Processing 735 In this paper, we consider how algorithmic choices that do not influence classification performance, make an initially difficult-to-implement algorithm, practical to implement. In particular, we present a practical architecture for performing real-time recognition of acoustic transients via a correlation-based algorithm. Correlation in analog VLSI poses two fundamental implementation challenges. First, there is the problem of template storage, second, there is the problem of accurate analog multiplication. Both problems can be solved by building sufficiently complex circuits. This solution is generally unsatisfactory because the resulting processors must have less area and consume less power than their digital counterparts in order to be competitive. Another solution to the storage problem is to employ novel floating gate devices. At present such devices can store analog values for years without significant degradation. Moreover, this approach can result in very compact, yet computationally complex devices. On the other hand, programming floating gate devices is not so straight-forward. It is relatively slow, it requires high voltage and it degrades the floating gate each time it is reprogrammed. Our "solution" is to side-step the problem completely and to develop an algorithmic solution that requires neither analog storage nor analog multiplication. Such an approach is attractive because it is both biologically plaUSible and electronically efficient. We demonstrate that a high level of classification performance on a real-world data set is achievable with no measurable loss of performance, compared to a baseline correlation algorithm. The acoustic transients used in our experiments were collected by K. Ryals and O. Steigerwald and are described in (Pineda, Ryals, Steigerwald and Furth, 1995). These transients consist of isolated Bangs, Claps, Clicks, Cracks, Oinks, Pings, Pops, Slaps, Smacks, Snaps, Thuds and Whacks that were recorded on OAT tape in an office environment. The ambient noise level was uncontrolled, but typical of a single-occupant office. Approximately 221 transients comprising 10 classes were collected. Most of the energy in one of our typical transients is dissipated in the first 10 ms. The remaining energy is dissipated over the course of approximately 100 ms. The transients had durations of approximately 20-100 ms. There was considerable in-class and extra-class variability in duration. The duration of a transient was determined automatically by a segmentation algorithm described below. The segmentation algorithm was also used to align the templates in the correlation calculations. 2 THE BASELINE ALGORITHM The baseline classification algorithm and its performance is described in Pineda, et al. (1995). Here we summarize only its most salient features. Like many biologically motivated acoustic processing algorithms, the preprocessing steps include time-frequency analysis, rectification, smoothing and compression via a nonlinearity (e.g. Yang, Wang and Shamma, 1992). Classification is performed by correlation against a template that represents a particular class. In addition , there is a "training" step which is required to create the templates. This step is described in the "correlation" section below. We turn now to a more detailed description of each processing step. A. Time-frequency Analysis: Time-frequency analysis for the baseline algorithm and the simulations performed in this work, was performed by an ultra-low power (5.5 mW) analog VLSI filter bank intended to mimic the processing performed by the mammalian cochlea (Furth, Kumar, Andreou and Goldstein, 1994). This real-time device creates a time-frequency representation that would ordinarily require hours of computation on a 736 F J. Pineda. G. Cauwenberghs and R. T. Edwards high-speed workstation. More complete descriptions can be found in the references. The time-frequency representation produced by the filter bank is qualitatively similar to that produced by a wavelet transformation. The center frequencies and Q-factors of each channel are uniformly spaced in log space. The low frequency channel is tuned to a center frequency of 100 Hz and Q-factor of 1.0, while the high frequency channel is tuned to a center frequency of 6000 Hz and Q-factor 3.5. There are 31 output channels. The 31-channel cochlear output was digitized and stored on disk at a raw rate of 256K samples per second. This raw rate was distributed over 32 channels, at rates appropriate for each channel (six rates were used, 1 kHz for the lowest frequency channels up to 32 kHz for the highest-frequency channels and the unfiltered channel). B. Segmentation: Both the template calculation and the classification algorithm rely on having a reliable segmenter. In our experiments, the transients are isolated and the noise level is low, therefore a simple segmenter is all that is needed. Figure 2. shows a segmenter that we implemented in software and which consists of a three layer neural network. noisy segmentation bit clean segmentation bit Figure 2: Schematic diagram showing the segmenter network The input layer receives mean subtracted and rectified signals from the cochlear filters. The first layer simply thresholds these signals. The second layer consists of a single unit that accumulates and rethresholds the thresholded signals. The second layer outputs a noisy segmentation signal that is nonzero if two or more channels in the input layer exceed the input threshold. Finally, the output neuron cleans up the segmentation signal by low-pass filtering it with a time-scale of 10 ms (to fill in drop outs) and by low-pass filtering it with a time-scale of 1 ms (to catch the onset of a transient). The outputs of the two low-pass filters are OR'ed by the output neuron to produce a clean segmentation bit. The four adjustable thresholds in the network were determined empirically so as to maximize the number of true transients that were properly segmented while minimizing the number of transients that were missed or cut in half. C. Smoothing & Normalization: The raw output of the filter bank is rectified and smoothed with a single pole filter and subsequently normalized. Smoothing was done with a the Architecture for Acoustic Transient Processing 737 same time-scale (l-ms) in all frequency channels. Let X(t) be the instantaneous vector of rectified and smoothed channel data, then the instantaneous output of the normalizer is X(t) = ~(t) II. Where () is a positive constant whose purpose is to prevent the ()+ X(t) normalization stage from amplifying noise in the absence of a transient signal. With this normalization we have IIX(t)lt z 0 if IIX(t)lll «(), and IIX(t)lll z 1 if IIX(t)lll » (). Thus () effectively determines a soft input threshold that transients must exceed if they are to be normalized and passed on to higher level processing. A sequence of normalized vectors over a time-window of length T is used as the feature vector for the correlation and classification stages of the algorithm. Figure 3. shows four normalized feature vectors from one class of transients (concatenated together) . ......I""'~~~ ~-~:..... ... ""--'"'~~ ...I\. .... .A.~~~ J,.;~ 1'0. .-. .,..... ~ ~ -...... ..-.. ...... ....... ~ .~ ~ -. ~ ~I'-.. ~ ~ ~ '-----....., )., ~ ~ ~ ........................ , I I I I I I o 50 100 150 200 250 300 Time (ms) Figure 3.: Normalized representation of the first 4 exemplars from one class of transients. D. Correlation: The feature-vectors are correlated in the time-frequency domain against a set of K time-frequency templates. The k - th feature-vector-template is precalculated by averaging over a corpus of vectors from the k - th class. Thus, if Ck represents the k - th transient class, and if ( ) k represents an average over the elements in a class, e.g. (X(t»)k = E{X(t)IX(t)E Ck}. Then the template is of the form bk(t) = (X(t»)k · The instantaneous output of the correlation stage is a K -dimensional vector c(t)whose t A k -th component is ck(t) = LX(t)· bk(t). The time-frequency window over which the t'=t-T correlations are performed is of length T and is advanced by one time-step between correlation calculations. E. Classification The classification stage is a simple winner-take-all algorithm that assigns a class to the feature vector by picking the component of ck(t) that has the largest value at the appropriate time, i.e. class = argmax{ck(tvalid)}' k 738 F. 1. Pineda, G. Cauwenberghs and R. T. Edwards The segmenter is used to determine the time tva1idwhen the output of the winner-take-all is to be used for classification. This corresponds to properly aligning the feature vector and the template. Leave-one-out cross-validation was used to estimate the out-of-sample classification performance of all the algorithms described in this paper. The rate of correct classification for the baseline algorithm was 92.8%. Out of a total of 221 events that were detected and segmented, 16 were misclassified. 3 A CORRELATION ALGORITHM FOR ANALOG VLSI We now address the question of how to perform classification without performing analoganalog multiplication and without having to store analog templates. To provide a better understanding of the algorithm, we present it as a set of incremental modifications to the baseline algorithm. This will serve to make clear the role played by each modification. Examination of the normalized representation in figure 3 suggests that the information content of anyone time-frequency bin cannot be very high. Accordingly, we seek a highly compressed representation that is both easy to form and with which it is easy to compute. As a preliminary step to forming this compressed representation, consider correlating the time-derivative of the feature vector with the time-derivative of the template, ckU)= I,t~(t).bk(t) where bk(t) = (X(t)}k' This modification has no effect on the out-of-sample performance of the winner-take-all classification algorithm. The above representation, by itself, has very few implementation advantages. It can, in principal, mitigate the effect of any systematic offsets that might emerge from the normalization circuit. Unfortunately, the price for this small advantage would be a very complex multiplier. This is evident since the time-derivative of a positive quantity can have either sign, both the feature vector and the template are now bipolar. Accordingly the correlation hardware would now require 4-quadrant analog-analog multipliers. Moreover the storage circuits must handle bipolar quantities as well. The next step in forming a compressed representation is to replace the time-differentiated template with just a sign that indicates whether the template value in a particular channel is increasing or decreasing with time. This template is b' k (t) = Sign( (XU)} k J. We denote this template as the [-1,+ 1]-representation template. The resulting classification algorithm yields exactly the same out-of-sample performance as the baseline algorithm. The 4-quadrant analog-analog multiply of the differentiated representation is reduced to a "4-quadrant analog-binary" multiply. The storage requirements are reduced to a single bit per timefrequency bin. To simplify the hardware yet further, we exploit the fact that the time derivative of a random unit vector net) (with respect to the I-norm) satisfies E{ ~Sign«(Uv))iv} = 2E{ ~e«(uv))iv} where e is a step function. Accordingly, if we use a template whose elements are in [0,1] instead of [-1, + 1], i.e. b' I k (t) = e( (X(t)} k)' we expect E{ ~ b' vXv } = 2E{ b' I v Xv} = IlxlI" provided the feature vector X(t) is drawn from the Architecture/or Acoustic Transient Processing 739 same class as is used to calculate the template. Furthermore, if the feature vector and the template are statistically independent, then we expect that either representation will produce a zero correlation, E{ ~ h' j(v } = E{ h" v Xv} = 0 . In practice, we find that the difference in correlation values between using the [0,1] and the [-1,+1] representations is simply a scale factor (approximately equal to 2 to several digits of precision). This holds even when the feature vectors and the templates do not correspond to the same class. Thus the difference between the two representations is quantitatively minor and qualitatively nonexistent, as evidenced by our classification experiments, which show that the out-ofsample performance of the [0,1] representation is identical to that of the [-1,+1] representation. Furthermore, changing to the [0,1] representation has no impact on the storage requirements since both representations require the storage of single bit per timefrequency bin. On the other hand, consider that by using the [0,1] representation we now have a "2-quadrant analog-binary" multiply instead of a "4-quadrant analog-binary" multiply. Finally, we observe that differentiation and correlation are commuting operations, A thus rather than differentiating X(t) before correlation, we can differentiate after the correlation without changing the result. This reduces the complexity of the correlation operation still further, since the fact that both X(t) and h" k (t) are positive means that we need only implement a correlator with I-quadrant analog-binary multiplies. The result of the above evolution is a correlation algorithm that empirically performs as well as a baseline correlation algorithm, but only requires binary-multiplexing to perform the correlation. We find that with only 16 frequency channels and 64 time bins (1024bits/templates) , we are able to achieve the desired level of performance. We have undertaken the design and fabrication of a prototype chip. This chip has been fabricated and we will report on it's performance in the near future. Figure 4 illustrates the key architectural features of the correlator/memory implementation. The rectified and 1-norm correlator/memory input Figure 4: Schematic architecture of the k-th correlator-memory. smoothed frequency-analyzed signals are input from the left as currents. The currents are normalized before being fed into the correlator. A binary time-frequency template is stored as a bit pattern in the correlator/memory. A single bit is stored at each time and frequency bin. If this bit is set, current is mirrored from the horizontal (frequency) lines onto vertical (aggregation) lines. Current from the aggregation lines is integrated and shifted in a bucket-brigade analog shift register. The last two stages of the shift register are differenced to estimate a time-derivative. 4 DISCUSSION AND CONCLUSIONS The correlation algorithm described in the previous section is related to the zero-crossing 740 F. J Pineda, G. Cauwenberghs and R. T. Edwards representation analyzed by Yang, Wang. and Shamma (1992). This is because bit flips in the templates correspond to the zero crossings of the expected time-derivative of the normalized "energy-envelope." Note that we do not encode the incoming acoustic signal with a zero-crossing representation. Interestingly enough, if both the analog signal and the template are reduced to a binary representation, then the classification performance drops dramatically. It appears that maintaining some analog information in the processing path is significant. The frequency-domain normalization approach presented above throws away absolute intensity information. Thus, low intensity resonances that remain excited after the initial burst of acoustic energy are as important in the feature vector as the initial burst of energy. These resonances can contain significant information about the nature of the transient but would have less weight in an algorithm with a different normalization scheme. Another consequence of the normalization is that even a transient whose spectrum is highly concentrated in just a few frequency channels will spread its information over the entire spectrum through the normalization denominator. The use of a normalized representation thus distributes the correlation calculation over very many frequency channels and serves to mitigate the effect of device mismatch. We consider the proposed correlator/memory as a potential component in more sophisticated acoustic processing systems. For example, the continuously generated output of the correlators , c(t), is itself a feature vector that could be used in more sophisticated segmentation and/or classification algorithms such as the time-delayed neural network approach ofUnnikrishnan, Hopfield and Tank (1991). The work reported in this report was supported by a Whiting School of Engineering! Applied Physics Laboratory Collaborative Grant. Preliminary work was supported by an APL Internal Research & Development Budget. REFERENCES Furth, P.M. and Kumar, N.G., Andreou, A.G. and Goldstein, M.H. , "Experiments with the Hopkins Electronic EAR", 14th Speech Research Symposium, Baltimore, MD pp.183-189, (1994). Pineda, F.J., Ryals, K, Steigerwald, D. and Furth, P., (1995). "Acoustic Transient Processing using the Hopkins Electronic Ear", World Conference on Neural Networks 1995, Washington DC. Yang, X., Wang K and Shamma, S.A. (1992). "Auditory Representations of Acoustic Signals", IEEE Trans. on Information Processing,.3.8., pp. 824-839. Morgan, N. , Bourlard, H., Greenberg, S., Hermansky, H. and Wu, S. L., (1996). "Stochastic Perceptual Models of Speech", IEEE Proc. IntI. Conference on Acoustics, Speech and Signal Processing, Detroit, MI, pp. 397-400. Unnikrishnan, KP., Hopfield J.J., and Tank, D.W. (1991). "Connected-Digit SpeakerDependent Speech Recognition Using a Neural Network with Time-Delayed Connections", IEEE Transactions on Signal Processing, 3.2, pp. 698-713
1996
53
1,251
Salient Contour Extraction by Temporal Binding in a Cortically-Based Network Shih.Cheng Yen and Leif H. Finkel Department of Bioengineering and Institute of Neurological Sciences University of Pennsylvania Philadelphia, PA 19104, U. S. A. syen @jupiter.seas.upenn.edu leif@jupiter.seas.upenn.edu Abstract It has been suggested that long-range intrinsic connections in striate cortex may play a role in contour extraction (Gilbert et aI., 1996). A number of recent physiological and psychophysical studies have examined the possible role of long range connections in the modulation of contrast detection thresholds (Polat and Sagi, 1993,1994; Kapadia et aI., 1995; Kovacs and Julesz, 1994) and various pre-attentive detection tasks (Kovacs and Julesz, 1993; Field et aI., 1993). We have developed a network architecture based on the anatomical connectivity of striate cortex, as well as the temporal dynamics of neuronal processing, that is able to reproduce the observed experimental results. The network has been tested on real images and has applications in terms of identifying salient contours in automatic image processing systems. 1 INTRODUCTION Vision is an active process, and one of the earliest, preattentive actions in visual processing is the identification of the salient contours in a scene. We propose that this process depends upon two properties of striate cortex: the pattern of horizontal connections between orientation columns, and temporal synchronization of cell responses. In particular, we propose that perceptual salience is directly related to the degree of cell synchronization. We present results of network simulations that account for recent physiological and psychophysical "pop-out" experiments, and which successfully extract salient contours from real images. 916 S. Yen and L. H. Finkel 2 MODEL ARCHITECTURE Linear quadrature steerable filter pyramids (Freeman and Adelson, 1991) are used to model the response characteristics of cells in primary visual cortex. Steerable filters are computationally efficient as they allow the energy at any orientation and spatial frequency to be calculated from the responses of a set of basis filters. The fourth derivative of a Gaussian and its Hilbert transform were used as the filter kernels to approximate the shape of the receptive fields of simple cells. The model cells are interconnected by long-range horizontal connections in a pattern similar to the co-circular connectivity pattern of Parent and Zucker (1989), as well as the "association field" proposed by Field et at. (1993). For each cell with preferred orientation, e, the orientations, ¢, of the pre-synaptic cell at position (i,j) relative to the post-synaptic cell, are specified by: ¢(e,i,j) = 2tan-l(f)-e (see Figure 1a). These excitatory connections are confined to two regions, one flaring out along the axis of orientation of the cell (co-axial), and another confined to a narrow zone extending orthogonally to the axis of orientation (trans-axial). The fan-out of the co-axial connections is limited to low curvature deviations from the orientation axis while the trans-axial connections are limited to a narrow region orthogonal to the cell's orientation axis. These constraints are expressed as: 1, if tan-I (f) -e < 1(, r(e i J. III) 1 zif tan -I (1) _ e = 7! ± E , , ,.." -, i 2' 0, otherwise. where K represents the maximum angular deviation from the orientation axis of the postsynaptic cell and £ represents the maximum angular deviation from the orthogonal axis of the post-synaptic cell. Connection weights decrease for positions with increasing angular deviation from the orientation axis of the cell, as well as positions with increasing distance, in agreement with the physiological and psychophysical findings. Figure 1 b illustrates the connectivity pattern. There is physiological, anatomical and psychophysical evidence consistent with the existence of both sets of connections (Nelson and Frost, 1985; Kapadia et at., 1995; Rockland and Lund, 1983; Lund et at., 1985; Fitzpatrick, 1996; Polat and Sagi, 1993, 1994). Cells that are facilitated by the connections inhibit neighboring cells that lie outside the facilitatory zones. The magnitude of the inhibition is such that only cells receiving strong support are able to remain active. This is consistent with the physiological findings of Nelson and Frost (1985) and Kapadia et al. (1995) as well as the intra-cellular recordings of Weliky et al. (1995) which show EPSPs followed by IPSPs when the longdistance connections were stimulated. This inhibition is thought to occur through disynaptic pathways. In the model, cells are assumed to enter a "bursting mode" in which they synchronize with other bursting cells. In cortex, bursting has been associated with supragranular "chattering cells" (Gray and McCormick (1996). In the model, cells that enter the bursting Salient Contour Extraction by Temporal Binding in a Cortically-based Network 917 mode are modeled as homogeneous coupled neural oscillators with a common fundamental frequency but different phases (Kopell and Ermentrout, 1986; Baldi and Meir, 1990). The p~ase of each oscillator is modulated by the phase of the oscillators to which it is coupled. Oscillators are coupled only to other oscillators with which they have strong, reciprocal, connections. The oscillators synchronize using a simple phase averaging rule: L w .. 8 .(1 -1) 8 j (/)= IJ J , W ii =1 LWij where e represents the phase of the oscillator and W U represents the weight of the connection between oscillator i and j. The oscillators synchronize iteratively with synchronization defined as the following condition: 18;Ct)-8/t)1 < 8, i.j E C, t < t max where C represents all the coupled oscillators on the same contour, 8 represents the maximum phase difference between oscillators, and Imax represents the maximum number of time steps the oscillators are allowed to synchronize. The salience of the chain is then represented by the sum of the activities of all the synchronized elements in the group, C. The chain with the highest salience is chosen as the output of the network. This allows us to compare the output of the model to psychophysical results on contour extraction. It has been postulated that the 40 Hz oscillations observed in the cortex may be responsible for perceptual binding across different cortical regions (Singer and Gray, 1995). Recent studies have questioned the functional significance and even the existence of these oscillations (Ghose and Freeman, 1992; Bair el ai., 1994). We use neural oscillators only as a simple functional means of computing synchronization and make no assumption regarding their possible role in cortex. r-------------------, (x " y,l =-6 ifLll=O. 4,=0 b Figure I: a) Co-circularity constraint. b) Connectivity pattern of a horizontally oriented cell. Length of line indicates connection strength. 3 RESULTS This model was tested by simulating a number of psychophysical experiments. A number of model parameters remain to be defined through further physiological and psychophysical experiments, thus we only attempt a qualitative fit to the data. All simulations were conducted with the same parameter set. 3.1 EXTRACTION OF SALIENT CONTOURS Using the same methods as Field el al. (1993), we tested the model's ability to extract contours embedded in noise (see Figure 2). Pairs of stimulus arrays were presented to the 918 S. Yen and L H. Finkel network, one array contains a contour, the other contains only randomly oriented elements. The network determines the stimulus containing the contour with the highest salience. Network performance was measured by computing the percentage of correct detection. The network was tested on a range of stimulus variables governing the target contour: 1) the angle, ~, between elements on a contour, 2) the angle between elements on a contour but with the elements aligned orthogonal to the contour passing through them, 3) the angle between elements with a random offset angle, ±a, with respect to the contour passing through them, and 4) average separation of the elements. 500 simulations were run at each data point. The results are shown in Figure 2. The model shows good qualitative agreement with the psychophysical data. When the elements are aligned, the performance of the network is mostly modulated by the co-axial connections, whereas when the elements are oriented orthogonal to the contour, the trans-axial connections mediate performance. Both the model and human subjects are adversely affected as the weights between consecutive elements decrease in strength. This reduces the length of the contour and thus the saliency of the stimulus. a) Elements aligned parallell to path b) Elements aligned orthogonal to path c) Elements aligned parallel to path with a = 30' d) Elements at 0.9' separation ~I:~ -U:":-,,~_,:_, ,,~ I :~~~o,. I:~ I :~ ~ ~\~ .. , ......... All ~ 70 ''0 70 :"'~ 70 70 'b ---0--DW ~ 60 "'"0 60 " 60 60 ' ", ~ 50 50 .. _._." • 50 50 _ •• _ •• _ •• _. ..::& -0-Model 4 0 ~~~~~~ 40 ~~-~1~~~4 0 ~~~~~~40~~~r-~~ 15 :\0 45 60 75 90 o 15 30 45 60 75 Angle (deg) Angle (deg) o 15 30 45 60 75 Angle (deg) 15 :\0 45 60 75 Angle (deg) Figure 2: Simulation results are compared to the data from 2 subjects (AH, DJF) in Field ef al. (1993). Stimuli consisted of 256 randomly oriented Gabor patches with 12 elements aligned to form a contour. Each data point represents results for 500 simulations. 3.2 EFFECTS OF CONTOUR CLOSURE In a series of experiments using similar stimuli to Field et al. (1993), Kovacs and Julesz (1993) found that closed contours are much more salient than open contours. They reported that when the inter-element spacing between all elements was gradually increased, the maximum inter-element separation for detecting closed contours, ~ c (defined at 75% performance), is higher than that for open contours, ~ o . In addition, they showed that when elements spaced at ~ o are added to a 'Jagged" (open) contour, the saliency of the contour increases monotonically but when elements spaced at ~ c are added to a circular contour, the saliency does not change until the last element is added and the contour becomes closed. In fact, at ~c , the contour is not salient until it is closed, at which point it suddenly "pops-out" (see Figure 3c). This finding places a strong constraint on the computation of saliency in visual perception. Interestingly, it has been shown that synchronization in a chain of coupled neural oscillators is enhanced when the chain is closed (Kopell and Ermentrout, 1986; Ermentrout, 1985; Somers and Kopell, 1993). This property seems to be related to the differences in boundary effects on synchronization between open and closed chains and appears to hold across different families of coupled oscillators. It has also been shown that synchronization is dependent on the coupling between oscillators -- the stronger the coupling, the better the synchronization, both in terms of speed and coherence (Somers Salient Contour Extraction by Temporal Binding in a Cortically-based Network 919 and Kopell, 1993; Wang, 1995). We believe these findings may apply to the psychophysical results. As in Kovacs and Julesz (1993), the network is presented with two stimuli, one containing a contour and the other made up of all randomly oriented elements. The network picks the stimulus containing the synchronized contour with the higher salience. In separate trials, the threshold for maximum separation between elements was determined for open and closed contours. The ratio of the separation of the background elements to the that of elements on a closed curve, <Pc, was found to be 0.6 (which is similar to the threshold of 0.65 recently reported by Kovacs et al., 1996), whereas the ratio for open contours, <Po, was found to be 0.9. (11 is the threshold separation of contour elements, <P, at a particular background separation). We then examined the changes in salience for open and closed contours. The performance of the network was measured as additional elements were added to an initial short contour of elements. The results are shown in Figure 3b. At <Po, both open and closed contours are synchronized but at <Pc, elements are synchronized only when the chains are closed. If salience can only be computed for synchronized contours, then as additional elements are added to an open chain at <Po' the salience would increase since the whole chain is synchronized. On the other hand, at <Pc, as long as the last element is missing, the chain is really an open chain, and since <Pc is smaller than <Po' the elements on the chain will not be able to synchronize and adding elements has no effect on salience. Once the last element is added, the chain is immediately able to synchronize and the salience of the contour increases dramatically and causes the contour to "pop-out". a) Performance as a function of background to contour separation. 100 _ 90 g 80 u C 70 ~ ~ 60 .. _" 00cP, -.. ~ 0 \ . --------~---- ~--~o ».,b , 50,..-r----r-"T"" ___ rA-.... 1.4 1.2 I 0.8 0.6 0.4 Ip c) Performance as a function of closure. b) Performance as a function of closure. Data from Kovacs and Julesz (1993) 85 85 "T""--------, 80 • ,t"'-~\ 80_ I \ ~ 75 --------1------8~ 75 ------------'4:i~ 70. J. ,0' 8 70,If: --o()-c 65 ~' , ~ 65 .... .,..,.~ / ~ 60 o.~'.o. (S,d d! 60." ~ 55' /l' '0' 55 9.,,-.0-_--0-0' 50 •• 50 • 0123456789 01234567 89 Additional ElemenLS Additional Elements Closed Open Figure 3: Simulation of the experiments of Kovacs and lulesz (1993). Stimuli consisted of 2025 randomly oriented Gabor patches, with 24 elements aligned to form a contour. Each data point represents results from 500 trials. a) Plot of the performance of the model with respect to the ratio of the separation of the background elements to the contour elements. Results show closed contours are salient to a more salient than open contours. b) Changes in salience as additional elements are added to open and closed contours. Results show that the salience of open contours increase monotonically while the salience of closed contours only change with the addition of the last element. Open contours were initially made up of 7 elements while closed contours were made up of 17 elements. c) The data from Kovacs and lulesz (\993) are re-plotted for comparison. 3.3 REAL IMAGES A stringent test of the model's capabilities is the ability to extract perceptually salient contours in real images. Figure 4 and 5 show results for a typical image. The original grayscale image, the output of the steerable filters, and the output of the model are shown in Figure 4a,b,c and Figure 5a,b,c respectively. The network is able to extract some of the more salient contours and ignore other high contrast edges detected by the steerable filters. Both simulations used filters at only one spatial scale and could be improved through interactions across multiple spatial frequencies. Nevertheless, the model shows promise for automated image processing applications 920 S. Yen and L. H. Finkel Figure 4: a) Plane image. b) Steerable filter response. c) Result of model showing the most salient contours. Figure 5: a) Satellite image of Bangkok. b) Steerable filter response. c) Salient contours extracted from the image. The model included filters at only one spatial frequency. 4 CONCLUSION We have presented a cortically-based model that is able to identify perceptually salient contours in images containing high levels of noise. The model is based on the use of long distance intracortical connections that facilitate the responses of cells lying along smooth contours. Salience is defined as the combined activity of the synchronized population of cells responding to a particular contour. The model qualitatively accounts for a range of physiological and psychophysical results and can be used in extracting salient contours in real images. Acknowledgements Supported by the Office of Naval Research (NOO014-93-1-0681), The Whitaker Foundation, and the McDonnell-Pew Program in Cognitive Neuroscience. References Bair, W., Koch, c., Newsome, W. & Britten, K. (1994). Power spectrum analysis of bursting cells in area MT in the behaving monkey. Journal of Neuroscience, 14, 2870-2892. Baldi, P. & Meir, R. (1990). Computing with arrays of coupled oscillators: An application to preattentive texture discrimination. Neural Computation, 2,458-471. Ermentrout, G. B. (1985). The behavior of rings of coupled oscillators. Journal of Mathematical Biology, 23, 55-74. Field, D. J., Hayes, A. & Hess, R. F. (1993). Contour integration by the human visual system: Evidence for a local "Association Field". Vision Research, 33, 173-193. Fitzpatrick, D. (1996). The functional-organization of local circuits in visual-cortex insights from the study of tree shrew striate cortex. Cerebral Cortex, 6, 329-341. Freeman, W. T. & Adelson, E. H. (1991). The design and use of steerable filters. IEEE Transactions on Pattern Analysis and Machine Intelligence, 13,891-906. Salient Contour Extraction by Temporal Binding in a Cortically-based Network 921 Gilbert, C. D., Das, A., Ito, M., Kapadia, M. & Westheimer, G. (1996). Spatial integration and cortical dynamics. Proceedings of the National Academy of Sciences USA, 93, 615-622. Ghose, G. M. & Freeman, R. D. (1992). Oscillatory discharge in the visual system: Does it have a functional role? Journal of Neurophysiology, 68, 1558-1574. Gray, C. M. & McCormick, D. A. (1996). Chattering cells -- superficial pyramidal neurons contributing to the generation of synchronous oscillations in the visualcortex. Science, 274, 109-113. Kapadia, M. K., Ito, M., Gilbert, C. D. & Westheimer. G. (1995). Improvement in visual sensitivity by changes in local context: Parallel studies in human observers and in VI of alert monkeys. Neuron, 15, 843-856. Kopell, N. & Ermentrout, G. B. (1986). Symmetry and phaselocking in chains of weakly coupled oscillators. Communications on Pure and Applied Mathematics, 39, 623660. Kovacs, 1. & Julesz, B. (1993). A closed curve is much more than an incomplete one: Effect of closure in figure-ground segmentation. Proceedings of National Academy of Sciences, USA, 90, 7495-7497. Kovacs, 1. & Julesz, B. (1994). Perceptual sensitivity maps within globally defined visual shapes. Nature, 370, 644-646. Kovacs, 1., Pol at, U. & Norcia, A. M. (1996). Breakdown of binding mechanisms in amblyopia. Investigative Ophthalmology & Visual Science, 37, 3078. Lund, J., Fitzpatrick, D. & Humphrey, A. L. (1985). The striate visual cortex of the tree shrew. In Jones, E. G. & Peters, A. (Eds), Cerebral Cortex (pp. 157-205). New York: Plenum. Nelson, J. 1. & Frost, B. J. (1985). Intracortical facilitation among co-oriented, co-axially aligned simple cells in cat striate cortex. Experimental Brain Research, 61, 54-61. Parent, P. & Zucker, S. W. (1989). Trace inference, curvature consistency, and curve detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 11, 823839. Polat, U. & Sagi, D. (1993). Lateral interactions between spatial channels: Suppression and facilitation revealed by lateral masking experiments. Vision Re5earch, 33, 993999. Polat, U. & Sagi, D. (1994). The architecture of perceptual spatial interactions. Vision Research, 34, 73-78. Rockland, K. S. & Lund, J. S. (1983). Intrinsic laminar lattice connections in primate visual cortex. Journal of Comparative Neurology, 216, 303-318. Singer, W. & Gray, C. M. (1995). Visual feature integration and the temporal correlation hypothesis. Annual Review of Neuroscience, 18, 555-586. Somers, D. & Kopell, N. (1993). Rapid synchronization through fast threshold modulation. Biological Cybernetics, 68, 393-407. Wang, D. (1995). Emergent synchrony in locally coupled neural oscillators. IEEE Transactions on Neural Networks, 6, 941-948. Weliky M., Kandler, K., Fitzpatrick, D. & Katz, L. C. (1995). Patterns of excitation arxl inhibition evoked by horizontal connections in visual cortex share a common relationship to orientation columns. Neuron, 15, 541-552. PART VIII ApPLICATIONS
1996
54
1,252
Learning Decision Theoretic Utilities Through Reinforcement Learning Magnus Stensmo Computer Science Division University of California Berkeley, CA 94720, U.S.A. magnus@cs.berkeley.edu Terrence J. Sejnowski Howard Hughes Medical Institute The Salk Institute 10010 North Torrey Pines Road La Jolla, CA 92037, U.S.A. terry@salk.edu Abstract Probability models can be used to predict outcomes and compensate for missing data, but even a perfect model cannot be used to make decisions unless the utility of the outcomes, or preferences between them, are also provided. This arises in many real-world problems, such as medical diagnosis, where the cost of the test as well as the expected improvement in the outcome must be considered. Relatively little work has been done on learning the utilities of outcomes for optimal decision making. In this paper, we show how temporal-difference reinforcement learning (TO(A» can be used to determine decision theoretic utilities within the context of a mixture model and apply this new approach to a problem in medical diagnosis. TO( A) learning of utilities reduces the number of tests that have to be done to achieve the same level of performance compared with the probability model alone, which results in significant cost savings and increased efficiency. 1 INTRODUCTION Decision theory is normative or prescriptive and can tell us how to be rational and behave optimally in a situation [French, 1988]. Optimal here means to maximize the value of the expected future outcome. This has been formalized as the maximum expected utility principle by [von Neumann and Morgenstern, 1947]. Decision theory can be used to make optimal choices based on probabilities and utilities. Probability theory tells us how probable different future states are, and how to reason with and represent uncertainty information. 1062 M. Stensmo and T. 1. Sejnowski Utility theory provides values for these states so that they can be compared with each other. A simple form of a utility function is a loss function. Decision theory is a combination of probability and utility theory through expectation. There has previously been a lot of work on learning probability models (neural networks, mixture models, probabilistic networks, etc.) but relatively little on representing and reasoning about preference and learning utility models. This paper demonstrates how both linear utility functions (i.e., loss functions) and non-linear ones can be learned as an alternative to specifying them manually. Automated fault or medical diagnosis is an interesting and important application for decision theory. It is a sequential decision problem that includes complex decisions (What is the most optimal test to do in a situation? When is it no longer effective to do more tests?), and other important problems such as missing data (both during diagnosis, i.e., tests not yet done, and in the database which learning is done from). We demonstrate the power of the new approach by applying it to a real-world problem by learning a utility function to improve automated diagnosis of heart disease. '2 PROBABILITY, UTILITY AND DECISION THEORY MODELS The system has separate probability and decision theory models. The probability model is used to predict the probabilities for the different outcomes that can occur. By modeling the joint probabilities these predictions are available no matter how many or few of the input variables are available at any instant. Diagnosis is a missing data problem because of the question-and-answer cycle that results from the sequential decision making process. Our decision theoretic automated diagnosis system is based on hypotheses and deductions according to the following steps: 1. Any number of observations are made. This means that the values of one or several observation variables of the probability model are determined. 2. The system finds probabilities for the different possible outcomes using the joint probability model to calculate the conditional probability for each of the possible outcomes given the current observations. 3. Search for the next observation that is expected to be most useful for improving the diagnosis according to the Maximum Expected Utility principle. Each possible next variable is considered. The expected value of the system prediction with this variable observed minus the current maximum value before making the additional observation and the cost of the observation is computed and defined as the net value of information for this variable [Howard, 1966]. The variable with the maximum of all of these is then the best next observation to make. 4. The steps 1-3 above are repeated until further improvements are not possible. This happens when none of the net value of information values in step 3 is positive. They can be negative since a positive cost has been subtracted. Note that we only look ahead one step (called a myopic approximation [Gorry and Barnett, 1967]). This is in principle suboptimal, however, the reinforcement learning procedure described below can compensate for this. The optimal solution is to consider all possible sequences, but the search tree grows exponentially in the number of unobserved variables. Learning Decision Theoretic Utilities through Reinforcement Learning 1063 Joint probabilities are modeled using mixture models [McLachlan and Basford, 1988]. Such models can be efficiently trained using the Expectation-Maximization (EM) algorithm [Dempster et al., 1977], which has the additional benefit that missing variable values in the training data also can be handled correctly. This is important since most real-world data sets are incomplete. More detail on the probability model can be found in [Stensmo and Sejnowski, 1995; Stensmo, 1995]. This paper is concerned with the utility function part of the decision theoretic model. The utilities are values assigned to different states so that their usefulness can be compared and actions are chosen to maximize the expected future utility. Utilities are represented as preferences when a certain disease has been classified but the patient in reality has another one [Howard, 1980; Heckerman et al., 1992]. For each pair of diseases there is a utility value between 0 and 1, where a 0 means maximally bad and a 1 means maximally good. This is a d x d matrix for d diseases, and the matrix can be interpreted as a kind of a loss function. The notation is natural and helps for acquiring the values, which is a non-trivial problem. Preferences are subjective contrary to probabilities which are objective (for the purposes of this paper). For example, a doctor, a patient and the insurance company may have different preferences, but the probabilities for the outcomes are the same. Methods have been devised to convert perceived risk to monetary values [Howard, 1980]. Subjects were asked to answer questions such as: "How much would you have to be paid to accept a one in a millionth chance of instant painless death r' The answers are recorded for various low levels of risk. It has been empirically found that people are relatively consistent and that perceived risk is linear for low levels of probability. Howard defined the unit micromort (mmt) to mean one in J millionth chance of instant painless death and [Heckerman et al., 1992] found that one subject valued 1 micromort to $20 (in 1988 US dollars) linearly to within a factor of two. We use this to convert utilities in [0,1] units to dollar values and vice versa. Previous systems asked experts to supply the utility values, which can be very complicated, or used some simple approximation. [Heckerman et al., 1992] used a utility value of 1 for misclassification penalty when both diseases are malign or both are benign, and 0 otherwise (see Figure 4, left). They claim that it worked in their system but this approximation should reduce accuracy. We show how to adapt and learn utilities to find better ones. 3 REINFORCEMENT LEARNING OF UTILmES Utilities are adapted using a type of reinforcement learning, specifically the method of temporal differences [Sutton, 1988]. This method is capable of adjusting the utility values correctly even though a reinforcement signal is only received after each full sequence of questions leading to a diagnosis. The temporal difference algorithm (ID(A» learns how to predict future values from past experience. A sequence of observations is used, in our case they are the results of the medical tests that have been done. We used ID(A) to learn how to predict the expected utility of the final diagnosis. Using the notation of Sutton, the function Pt predicts the expected utility at time t. Pt is a vector of expected utilities, one for each outcome. In the linear form described above, Pt = P(Xt, Wt) = WtXt, where Wt is a matrix of utility values and Xt is the vector of probabilities of the outcomes, our state description. The objective is to learn the utility matrix Wt. 1064 M. Stensmo and T. J. Sejnowski We use an intra-sequence version of the ID(,\) algorithm so that learning can occur during normal operation of the system [Sutton, 1988]. The update equation is t Wt+! = Wt + a[P(xt+!, Wt) - P(Xt, Wt)) 2: ,\t-kv wP(Xk, Wt), (1) k=1 where a is the learning rate and ,\ is a discount factor. With Pk = P (x k, Wt) = x k Wt and et = E!=l ,\t-kv wP(Xk, wt} = E~=I ,\t-kxk , (1) becomes the two equations Wt+l Wt + awt[x t+1 - xt)et et+l Xt+l + '\et, starting with el = Xl . These update equations were used after each question was answered. When the diagnosis was done, the reinforcement signal z (considered to be observation Pt+1) was obtained and the weights were updated: Wt+! = Wt + awt[z - Xt)et. A final update of et was not necessary. Note that t~is method allows for the use of any differentiable utility function, specifically a neural network, in the place of P(Xk, wt}. Preference is sUbjective. In this paper we investigated two examples of reinforcement. One was to simply give the highest reinforcement (z = 1) on correct diagnosis and the lowest (z = 0) for errors. This yielded a linear utility function or loss function that was the unity matrix which confirmed that the method works. When applied to a non-linear utility function the result is non-trivial. In the second example the reinforcement signal was modified by a penalty for the use of a high number questions by multiplying each z above with (maXq -q)j(maXq - minq), where q is the number of questions used for the diagnostic sequence, and the minimum and maximum number of questions are minq and maXq, respectively. The results presented in the next section used this reinforcement signal. 4 RESULTS The publicly available Cleveland heart-disease database was used to test the method. It consists of 303 cases where the disorder is one of four types of heart-disease or its absence. There are fourteen variables as shown in Figure 1. Continuous variables were converted into a 1-0/-N binary code based on their distributions among the cases in the database. Nominal and categorical variables were coded with one unit per value. In total 96 binary variables coded the 14 original variables. To find the parameter values for the mixture model that was used for probability estimation, the EM algorithm was run until convergence [Stensmo and Sejnowski, 1995; Stensmo, 1995]. The classification error was 16.2%. To get this result all of the observation variables were set to their correct values for each case. Note that all this information might not be available in a real situation, and that the decision theory model was not needed in this case. To evaluate how well the complete sequential decision process system does, we went through each case in the database and answered the questions that came up according to the correct values for the case. When the system completed the diagnosis sequence, the result was compared to the actual disease that was recorded in the database. The number of questions that were answered for each case was also recorded ( q above). After all of the cases had been processed in this way, the average number of questions needed, its standard Learning Decision Theoretic Utilities through Reinforcement Learning 1065 Observ. Description Values Cost (mmt) Cost ($) 1 age Age in years continuous 0 0 2 sex Sex of subject male/female 0 0 3 cp Chest pain four types 20 400 4 trestbps Resting blood pressure continuous 40 800 5 chol Serum cholesterol continuous 100 2000 6 fbs Fasting blood sugar <,or> 100 2000 120mgldl 7 restecg Resting electrocardiographic five values 100 2000 result 8 thalach Maximum heart rate achieved continuous 100 2000 9 exang Exercise induced angina yes/no 100 2000 lO oldpeak ST depression induced by continuous 100 2000 exercise relative to rest 11 slope Slope of peak exercise up/flat/down 100 2000 STsegment 12 ca Number major vessels colored 0-3 100 2000 by flouroscopy 13 thaI Defect type normaVfixedi 100 2000 reversible Disorder Description Values 14 num Heart disease No disease/ four types Figure 1: The Cleveland Heart Disease database. The database consists of 303 cases described by 14 variables. Observation costs are somewhat arbitrarily assigned and are given in both dollars and converted to micromorts (mmt) in [0,1] units based on $20 per micromort (one in 1 millionth chance of instant painless death). deviation, and the number of errors were calculated. If the system had several best answers, one was selected randomly. Observation costs were assigned to the different variables according to Figure 1. Using the full utility/decision model and the O/I-approximation for the utility function (left part of Figure 4), there were 29.4% errors. The results are summarized in Figure 2. Over the whole data set an average of 4.42 questions were used with a standard deviation of 2.02. Asking about 4-5 questions instead of 13 is much quicker but unfortunately less accurate. This was before the utilities were adapted. With TD(~) learning (Figure 3), the number of errors decreased to 16.2% after 85 repeated presentations of all of the cases in random order. We varied ~ from 0 to 1 in increments of 0.1, and a over several orders of magnitude to find the reported results. The resulting average number of questions were 6.05 with a standard deviation of 2.08. The utility matrix after 85 iterations is shown in Figure 4 with 0'=0.0005 and ~=O.I. The price paid for increased robustness was an increase in the average number of questions from 4.42 to 6.05, but the same accuracy was achieved using only less than half of them on average. Many people intuitively think that half of the questions should be enough. There is, however, no reason for this; furthermore there is no procedure to stop asking questions if observations are chosen randomly. In this paper a simple state description has been used, namely the predicted probabilities of the outcomes. We have also tried other representations by including the test results in the state description. On this data set similar results were obtained. 1066 M. Stensmo and T. 1. Sejnowski Model Errors # Questions St. Dev. Probability model only 16.2% 13 011 approximation 29.4% 4.42 2.02 After 85 iterations of TD("\) learning 16.2% 6.05 2.08 Figure 2: Results on the Cleveland Heart Disease Database. The three methods are described in the text. The first method does not use a utility model. The 011 approximation use the matrix in Figure 4, left. The utility matrix that was learned by TD("\) is shown in Figure 4, right. 7 30 '" 6.5 25 c 6 ,...-.." ~ 0 .;: 5.5 g 20 '" Q) ::I 5 ~ CI 15 4.5 4 10 0 20 40 60 80 0 20 40 60 80 Iteration Iteration Figure 3: Learning graphs with discount-rate parameter ..\=0.1, and learning rate a=O.OOO5 for the TD("\) algorithm. One iteration is a presentation of all of the cases in random order. 5 SUMMARY AND DISCUSSION We have shown how utilities or preferences can be learned for different expected outcomes in a complex system for sequential decision making based on decision theory. Temporaldifferences reinforcement learning was efficient and effective. This method can be extended in several directions. Utilities are usually modeled linearly in decision theory (as with the misclassification utility matrix), since manual specification and interpretation of the utility values then is quite straight-forward. There are advantages with non-linear utility functions and, as indicated above, our method can be used for any utility function that is differentiable. Initial After 8S iterations 1 0 0 0 0 0.8179 0.0698 0.0610 0.0435 0.0505 0 1 1 1 1 0.0579 0.6397 0.2954 0.3331 0.6308 0 1 1 1 1 0.0215 0.1799 0.6305 0.3269 0.6353 0 1 1 1 1 0.0164 0.1430 0.2789 0.7210 0.6090 0 1 1 1 1 0.0058 0.1352 0.2183 0.2742 0.8105 Figure 4: Misclassification utility matrices. The disorder no disease is listed in the first row and column, followed by the four types of heart disease. Left: Initial utility matrix. Right: After TD learning with discount-rate parameter ..\=0.1 and learning rate a=O.OOO5. Element Uij (row i, column j) is the utility when outcome i has been chosen but when it actually is j. Maximally good has value 1, and maximally bad has value O. Learning Decision Theoretic Utilities through Reinforcement Learning 1067 An alternative to learning the utility or value function is to directly learn the optimal actions to take in each state, as in Q-Iearning [Watkins and Dayan, 1992]. This would require one to learn which question to ask in each situation instead of the utility values but would not be directly analyzable in terms of maximum expected utility. Acknowledgements Financial support for M.S. was provided by the Wenner-Gren Foundations and the Foundation Blanceftor Boncompagni-Ludovisi, nee Bildt. The heart-disease database is from the University of California, Irvine Repository of Machine Learning Databases and originates from R. Detrano, Cleveland Clinic Foundation. Stuart Russell is thanked for discussions. References Dempster, A. P., Laird, N. M. and Rubin, D. B. (1977). Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, Series, B. , 39, 1-38. French, S. (1988). Decision Theory: An Introduction to the Mathematics of Rationality. Ellis Horwood, Chichester, UK. Gorry, G. A. and Barnett, G. O. (1967). Experience with a model of sequential diagnosis. Computers and Biomedical Research, 1,490-507. Heckerman, D. E., Horvitz, E. J. and Nathwani, B. N. (1992). Toward normative expert systems: Part I. The Pathfinder project. Methods of Information in Medicine, 31, 90105. Howard, R. A. (1966). Information value theory. IEEE Transactions on Systems Science and Cybernetics, SSC-2, 22-26. Howard, R. A. (1980). On making life and death decisions. In Schwing, R. C. and Albers, Jr., W. A., editors, Societal risk assessment: How safe is safe enough? Plenum Press, New York, NY. McLachlan, G. J. and Basford, K. E. (1988). Mixture Models: Inference and Applications to Clustering. Marcel Dekker, Inc., New York, NY. Stensmo, M. (1995). Adaptive Automated Diagnosis. PhD thesis, Royal Institute of Technology (Kungliga Tekniska Hogskolan), Stockholm, Sweden. Stensmo, M. and Sejnowski, T. J. (1995). A mixture model system for medical and machine diagnosis. In Tesauro, G., Touretzky, D. S. and Leen, T. K., editors, Advances in Neural Information Processing Systems, vol. 7, pp 1077-1084. MIT Press, Cambridge, MA. Sutton, R. S. (1988). Learning to predict by the method of temporal differences. Machine Learning, 3,9-44. von Neumann, J. and Morgenstern, O. (1947). Theory of Games and Economic Behavior. Princeton University Press, Princeton, NJ. Watkins, C. J. and Dayan, P. (1992). Q-Iearning. Machine Learning, 8, 279-292.
1996
55
1,253
Spatial Decorrelation in Orientation Tuned Cortical Cells Alexander Dimitrov Department of Mathematics University of Chicago Chicago, IL 60637 a-dimitrov@uchicago.edu Jack D. Cowan Department of Mathematics University of Chicago Chicago, IL 60637 cowan@math.uchicago.edu Abstract In this paper we propose a model for the lateral connectivity of orientation-selective cells in the visual cortex based on informationtheoretic considerations. We study the properties of the input signal to the visual cortex and find new statistical structures which have not been processed in the retino-geniculate pathway. Applying the idea that the system optimizes the representation of incoming signals, we derive the lateral connectivity that will achieve this for a set of local orientation-selective patches, as well as the complete spatial structure of a layer of such patches. We compare the results with various physiological measurements. 1 Introduction In recent years much work has been done on how the structure of the visual system reflects properties of the visual environment (Atick and Redlich 1992; Attneave 1954; Barlow 1989). Based on the statistics of natural scenes compiled and studied by Field (1987) and Ruderman and Bialek (1993), work was done by Atick and Redlich (1992) on the assumption that one of the tasks of early vision is to reduce the redundancy of input signals, the results of which agree qualitatively with numerous physiological and psychophysical experiments. Their ideas were further strengthened by research suggesting the possibility that such structures develop via simple correlation-based learning mechanisms (Atick and Redlich 1993; Dong 1994). As suggested by Atick and Li (1994), further higher-order redundancy reduction of the luminosity field in the visual processing system is unlikely, since it gives little benefit in information compression. In this paper we apply similar ideas to a different input signal which is readily available to the system and whose statistical properties are lost in the analysis of the luminosity signal. We note that after the Spatial Decorrelation in Orientation Tuned Cortical Cells 853 application of the retinal "mexican hat" filter the most obvious salient features that are left in images are sharp changes in luminosity, for which the filter is not optimal, i.e. local edges. Such edges have correlations which are very different from the luminosity autocorrelation of natural images (Field 1987), and have zero probability measure in visual scenes, so they are lost in the ensemble averages used to compute the autocorrelation function of natural images. We know that this signal is projected to a set of direction-sensitive units in VI for each distinct retinal position, thereby introducing new redundancy in the signal. Thus the necessity for compression and use of factorial codes arises once again. Since local edges are defined by sharp changes in the luminosity field, we can use a derivative operation to pick up the pertinent structure. Indeed, if we look at the gradient of the luminosity as a vector field, its magnitude at a point is proportional to the change of luminosity, so that a large magnitude signals the possible presence of a discontinuity in the luminosity profile. Moreover, in two dimensions, the direction of the gradient vector is perpendicular to the direction of the possible local edge, whose presence is given by the magnitude. These properties define a one-to-one correspondence between large gradients and local edges. The structure of the network we use reflects what is known about the structure of V!. We select as our system a layer of direction sensitive cells which are laterally connected to one another, each receiving input from the previous layer. We assume that each unit receives as input the directional derivative of the luminosity signal along the preferred visuotopic axis of the cell. This implies that locally the input to a cell is proportional to the cosine of the angle between the unit's preferred direction and the local gradient (edge). Thus each unit receives a broadly tuned signal, with HW-HH approximately 60 0 • With this feed-forward structure, the idea that the system is trying to decorrelate its inputs suggests a way to calculate the lateral connections that will perform this task. This calculation, and a further study of the statistical properties of the input is the topic of the paper. 2 Mathematical Model Let G(x) = (Gl(x), G2(x» be the gradient of luminosity at x. Assume that there is a set of N detectors with activity Oi at x, each with a preferred direction ni. Let the input from the previous layer to each detector be the directional derivative along its preferred direction. d Vi(x) = IGrad(L(x».nil = Idni L(x)1 (1) There are long range correlations in the inputs to the network due both to the statistical structure of the natural images and the structure of the input. The simplest of them are captured in the two-point correlation matrix Rij(Xl , X2) =< Vi(XI)Vj(X2) >, where the averaging is done across images. Then R is a block matrix, with an N x N matrix at each spatial position (Xl, X2)' We formulate the problem in terms of a recurrent kernel W, so that (2) The biological interpretation of this is that V is the effective input to VI from the LGN and W specifies the lateral connectivity in V!. This equation describes the steady state of the linear dynamical system 6 = -0 + W * 0 + V. The 854 A. Dimitrov and J. D. Cowan above recurrent system has a solution for 0 not an eigenfunction of W in the form 0= (6 - W)-l * V = K * V. This suggests that there is an equivalent feed-forward system with a transfer function K = (6 - W)-l and we can consider only such systems. The corresponding feed-forward system is a linear system that acts on the input Vex) to produce an output O(x) = (1<. V)(x) == J K(x, y). V(y)dy. If we use Barlow's redundancy reduction hypothesis (Barlow 1989), this filter should decorrelate the output signal. This is achieved by requiring that 6(X1 - X2) '" < O(xd 0 0(X2) >=< (K . V)(X1) 0 (K . V)(X2) >{:} 6(X1 - X2) '" K· R . KT (3) The aim then is to solve (3) for K. Obviously, this is equivalent to KT . K "" R- 1 (assuming K and R are non-singular), which has a solution K '" R-t, unique up to a unitary transformation. The corresponding recurrent filter is then (4) This expression suggests an immediate benefit in the use of lateral kernels by the system. As (4) shows, the filter does not now require inverting the correlation matrix and thus is more stable than a feed-forward filter. This also helps preserve the local structure of the autocorrelator in the optimal filter, while, because of the inversion process, a feed-forward system will in general produce non-local, non-topographic solutions. To obtain a realistic connectivity structure, we need to explicitly include the effects of noise on the system. The system is then described by 0 1 = V + N1 + M * W * (01 + N2), where N1 is the input noise and N2 is the noise, generated by individual units in the recurrently connected layer. Similarly to a feed-forward system (Atick and Redlich 1992), we can modify the decorrelation kernel W derived from (2) to M * W. The form of the correction M, which minimizes the effects of noise on the system, is obtained by minimizing the distance between the states of the two systems. If we define X2(M) =< 10 - 0 1 12 > as the distance function, the solution to ox;1M ) = 0 will give us the appropriate kernel. A solution to this problem is M * W = 6 - (R + Nf + Ni) * (p R 1/ 2 + Ni)-l (5) We see that it has the correct asymptotics as N1, N2 approach zero. The filter behaves well for large N2 , turning mostly into a low-pass filter with large attenuation. It cannot handle well large N1 and reaches -00 proportionally to N'f. 3 Results 3.1 Local Optimal Linear Filter As a first calculation with this model, consider its implications for the connectivity between units in a single hypercolumn. This allows for a very simple application of the theory and does not require any knowledge of the input signal under very general assumptions. We assume that direction selective cells receive as input from the previous layer the projection of the gradient onto their preferred direction. Thus they act as directional Spatial Decorrelation in Orientation Tuned Cortical Cells 855 derivatives, so that their response to a signal with the luminosity profile L( x) and no input from other lateral units is Vi(x) = IGrad(L(x)).nil = Idldni(L(x))1 With this assumption the outputs of the edge detectors are correlated. Define a (local) correlation matrix R;j =< ViVj >. By assumption (1), Vk = la Gas(a ak)1, where a and a are random, independent variables, denoting the magnitude and direction of the local gradient and ak is the preferred angle of the detector. Assuming spatially isotropic local structure for natural scenes, we can calculate the average of R by integrating over a uniform probability measure in a. Then (6) where A =< a2 > can be factored because of the assumption of statistical independence. By the homogeneity assumption, Rij is a function of the relative angle lai - aj I only. This allows us to easily calculate the integral in (6) from its Fourier series. Indeed, in Fourier space R is just the square of the power spectrum of the underlying signal, i.e., cos(a) on [0, 7r]. Thus we obtain the form of R analytically. Knowing the local correlations, we can find a recurrent linear filter which decorrelates the outputs after it is applied. This filter is W = 0 - p R-! (Sec.2), unique up to a unitary transformation. w 0.8 0.6 0.4 0.2 e -0.2 - 0. 4 Figure 1: Local recurrent filter in the presence of noise. The connection strength W depends only on the relative angle (J between units. If we include noise in the calculation according to (5), we obtain a filter which depends on the signal to noise ratio of the input level. We model the noise process here as a set of independent noise processes for each unit, with (Ndi being the input noise and (N2)i the output noise for unit i. All noise processes are assumed statistically independent. The result for SI N2 "" 3 is shown on Fig.I. We observe the broadening of the central connections, caused by the need to average local results in order to overcome the noise. It was calculated at very low Nl level, since, as mentioned in Section 2, the filter is unstable with respect to input noise. With this filter we can directly compare calculations obtained from applying it to a specific input signal, with physiological measurements of the orientation selectivity of cells in the cortex. The results of such comparisons are presented in Fig.2, in which we plot the activity of orientation selective cells in arbitrary units vs stimulus angle in degrees. We see very good matches with experimental results of Celebrini, Thorpe, Trotter, and Imbert (1993), Schiller, Finlay, and Volman (1976) and Orban (1984). We expect some discrepancies, such as in Figures 2.D and 2.F, which can be attributed to the threshold nature of real neural units. We see that we can use the model to classify physiologically distinct cells by the value of the N2 parameter 856 A. Dimitrov and J. D. Cowan that describes them. Indeed, since this parameter models the intrinsic noise of a neural unit, we expect it to differ across populations. A B c -10 -tOO ..... D E F Figure 2: Comparison with experimental data. The activity of orientation selective cells in arbitrary units is plotted against stimulus angle in degrees. Experimental points are denoted with circles, calculated result with a solid line. The variation in the forms of the tuning curves could be accounted for by selecting different noise levels in our noise model. A - data from cell CAJ4 in Celebrini et.al. and fit for Nl = 0.1, N2 = 0.2. Bdata from cell CAK2 in Celebrini et.al. and fit for Nl = 0.35, N2 = 0.1. C - data from a complex cell from Orban and fit for Nl = 0.3, N2 = 0.45. D - data from a simple cell from Orban and fit for Nl = 1.0, N2 = 0.45. E - data from a simple cells in Schiller et.al. and fit for Nl = 0.06, N2 = 0.001. F - data from a simple cells in Schiller et.al. and fit for Nl = 15.0, N2 = 0.01. 3.2 Non-Local Optimal Filter We can perform a similar analysis of the non-local structure of natural images to design a non-local optimal filter. This time we have a set of detectors Vk(X) = la(x) Cos(a(x) - k 7r/N) I and a correlation function Rij(X, y) =< Vi(x) Vj(y) >, averaged over natural scenes. We assume that the function is spatially translation invariant and can be represented as Rij(X, y) = ~j(x - y). The averaging was done over a set of about 100 different pictures, with 10-20 2562 samples taken from each picture. The structure of the correlation matrix depends both on the autocorrelator of the gradient field and the structure of the detectors, which are correlated. Obviously the fact that the output units compute la(x) Cos(a(x) - k7r/N)1 creates many local correlations between neighboring units. Any non-local structure in the detector set is due to a similar structure, present in the gradient field autocorrelator. The structure of the translation-invariant correlation matrix R( x) is shown in Fig.3A. This can be interpreted as the correlation between the input to the center hypercolumn with the input to rest of the hypercolumns. The result of the complete model (5) can be seen in Fig.3B. Since the filter is also assumed to be translation invariant, the pictures can be interpreted as the connectivity of the center hypercolumn with the rest of the network. This is seen to be concentrated near the diagonal, Spatial Decorrelation in Orientation Tuned Cortical Cells 857 A B 2 -1 ... .. I . .i-I ~ , II Figure 3: A. The autocorrelation function of a set with 8 detectors. Dark represents high correlation, light - low correlation. The sets are indexed by the preferred angles Oi, OJ in units of f and each RiJ has spatial structure, which is represented as a 32 x 32 square. B. The lateral connectivity for the central horizontal selective unit with neighboring horizontal (1) and 1r/4 (2) selective units. Note the anisotropic connectivity and the rotation of the connectivity axis on the second picture. and weak in the two adjacent bands, which represent connections to edge detectors with a perpendicular preferred direction. The noise minimizing filter is a low pass filter, as expected, and thus decreases the high frequency component of the power spectrum of the respective decorrelating filter. 4 Conclusions and Discussion We have shown that properties of orientation selective cells in the visual cortex can be partially described by some very simple linear systems analysis. Using this we obtain results which are in very good agreement with physiological and anatomical data of single-cell recordings and imaging. We can use the parameters of the model to classify functionally and structurally differing cells in the visual cortex. We achieved this by using a recurrent network as the underlying model. This was chosen for several reasons. One is that we tried to give the model biological plausibility and recurrency is well established on the cortical level. Another related heuristic argument is that although there exists a feed-forward network with equivalent properties, as shown in Section 2, such a network will require an additional layer of cells, while the recurrent model allows both for feed-forward processing (the input to our model) as well as manipulation of the output of that (the decorrelation procedure in our model). Finally, while a feed-forward network needs large weights to amplify the signal, a recurrent network is able to achieve very high gains on the input signal with relatively small weights by utilizing special architecture. As can be seen from our equivalence model, K = (6 - W)-l, so if W is so constructed as to have an eigenvalues close to 1, it will produce enormous amplification. Our work is based on previous suggestions relating the structure of the visual environment to the structure of the visual pathway. It was thought before (Atick and Li 1994) that this particular relation can describe only early visual pathways, but is insufficient to account for the structure of the striate cortex. We show here that redundancy reduction is still sufficient to describe many of the complexities of the visual cortex, thus strengthening the possibility that this is a basic building princi858 A. Dimitrov and J. D. Cowan pIe for the visual system and one should anticipate its appearance in later regions of the latter. What is even more intriguing is the possibility that this method can account for the structure of other sensory pathways and cortices. We know e.g. that the somatosensory pathway and cortex are similar to the visual ones, because of the similar environments that they encounter (luminosity, edges and textures have analogies in somesthesia). Similar analogies may be expected for the auditory pathway. We expect even better results if we consider a more realistic non-linear model for the neural units. In fact this improves tremendously the information-processing abilities of a bounded system, since it captures higher order correlations in the signal and allows for true minimization of the mutual information in the system, rather than just decorrelating. Very promising results in this direction have been recently described by Bell and Sejnowski (1996) and Lin and Cowan (1997) and we intend to consider the implications for our model. Acknowledgements Supported in part by Grant # 96-24 from the James S. McDonnell Foundation. References Atick, J. J. and Z. Li (1994). Towards a theory of the striate cortex. Neural Computation 6, 127-146. Atick, J. J. and N. N. Redlich (1992). What does the retina know about natural scenes? Neural Computation 4, 196-210. Atick, J . J. and N. N. Redlich (1993). Convergent algorithm for sensory receptive field developement. Neural Computation 5, 45-60. Attneave, F. (1954). Some informational aspects of visual perception. Psychological Review 61, 183-193. Barlow, H. B. (1989). Unsupervised learning. Neural Computation 1,295-311. Bell, A. T. and T. J. Sejnowski (1996). The "independent components" of natural scences are edge filters. Vision Research (submitted). Celebrini, S., S. Thorpe, Y. Trotter, and M. Imbert (1993). Dynamics of orientation coding in area VI of the awake primate. Visual Neuroscience 10, 811-825. Dong, D. (1994). Associative decorrelation dynamics: a theory of selforganization and optimization in feedback networks. Volume 7 of Advances in Neural Information Processing Systems, pp. 925-932. The MIT Press. Field, D. J. (1987). Relations between the statistics of natural images and the response properties of cortical cells. J. Opt. Soc. Am. 4, 2379-2394. Lin, J . K. and J. D. Cowan (1997). Faithful representation of separable input distributions. Neural Computation, (to appear). Orban, G. A. (1984). Neuronal Operations in the Visual Cortex. Springer-Verlag, Berlin. Ruderman, D. L. and W. Bialek (1993). Statistics of natural images: Scaling in the woods. In J. D. COWall, G. Tesauro, and J. Alspector (Eds.), Advances in Neural Information Processing Systems, Volume 6. Morgan Kaufman, San Mateo, CA. Schiller, P., B. Finlay, and S. Volman (1976). Quantitative studies of singlecell properties in monkey striate cortex. II. Orientation specificity and ocular dominance. J. Neuroph. 39(6), 1320-1333.
1996
56
1,254
Sequential Tracking in Pricing Financial Options using Model Based and Neural Network Approaches Mahesan Niranjan Cambridge University Engineering Department Cambridge CB2 IPZ, England niranjan@eng.cam.ac.uk Abstract This paper shows how the prices of option contracts traded in financial markets can be tracked sequentially by means of the Extended Kalman Filter algorithm. I consider call and put option pairs with identical strike price and time of maturity as a two output nonlinear system. The Black-Scholes approach popular in Finance literature and the Radial Basis Functions neural network are used in modelling the nonlinear system generating these observations. I show how both these systems may be identified recursively using the EKF algorithm. I present results of simulations on some FTSE 100 Index options data and discuss the implications of viewing the pricing problem in this sequential manner. 1 INTRODUCTION Data from the financial markets has recently been of much interest to the neural computing community. The complexity of the underlying macro-economic system and how traders react to the flow of information leads to highly nonlinear relationships between observations. Further, the underlying system is essentially time varying, making any analysis both difficult and interesting. A number of problems, including forecasting a univariate time series from past observations, rating credit risk, optimal selection of portfolio components and pricing options have been thrown at neural networks recently. The problem addressed in this paper is that of sequential estimation, applied to pricing of options contracts. In a nonstationary environment, such as financial markets, sequential estimation is the natural approach to modelling. This is because data arrives at the modeller sequentially, and there is the need to build and apply the Sequential Tracking of Financial Option Prices 961 best possible model with available data. At the next point in time, some additional data is available and the task becomes one of optimally updating the model to account for the new data. This can either be done by reestimating the model with a moving window of data or by sequentially propagating the estimates of model parameters and some associated information (such as the error covariance matrices in the Kalman filtering framework discussed in this paper). 2 SEQUENTIAL ESTIMATION Sequential estimation of nonlinear models via the Extended Kalman Filter algorithm is well known (e.g. Candy, 1986; Bar-Shalom & Li, 1993). This approach has also been widely applied to the training of Neural Network architectures (e.g. Kadirkamanathan & Niranjan, 1993; Puskorius & Feldkamp, 1994). In this section, I give the necessary equations for a second order EKF, i.e. Taylor series expansion of the nonlinear output equations, truncated at order two, for the state space model simplified to the system identification framework considered here. The parameter vector or state vector, 0, is assumed to have the following simple random walk dynamics. (t(n + 1) = (t(n) + .Y(n) , where .Y( n) is a noise term, known as process noise . .Y( n) is of the same dimensionality as the number of states used to represent the system. The process noise gives a random walk freedom to the state dynamics facilitating the tracking behaviour desired in non stationary environments. In using the Kalman filtering framework, we assume the covariance matrix of this noise process, denoted Q, is known. In practice, we set Q to some small diagonal matrix. The observations from the system are given by the equation ~(n) = [({t, U) + w(n), where, the vector ~(n) is the output of the system consisting of the call and put option prices at time n. U denotes the input information. In the problem considered here, U consists of the price of the underlying asset and the time to maturity if the option. w is known as the measurement noise, covariance matrix of which, denoted R, is also assumed to be known. Setting the parameters Rand Q is done by trial and error and knowledge about the noise processes. In the estimation framework considered here, Q and R determine the tracking behaviour of the system. For the experiments reported in this paper, I have set these by trial and error, but more systematic approaches involving multiple models is possible (Niranjan et al, 1994). The prior estimates at time (n + 1), using all the data upto time (n) and the model of the dynamical system, or the prediction phase of the Kalman algorithm is given by the equations: ~(n + lin) P(n + lin) ~(n + lin) = ~(nln) P(nln) + Q(n) 1 n8 10 ({t(n + lin)) + 2" L~i tr (mo(n + 1) P(n + lin)) i=l where 10 and H~o are the Jacobian and Hessians of the output e; also no = 2. ~i are unit vectors in direction i. tr(.) denotes trace of a matrix. The posterior esti962 M. Niranjan mates or the correction phase of the Kalman algorithm are given by the equations: 8(n+1) = l..o(n + l)P(n + 1In)~(n + 1) 1 nil nil + 2 L L~i~j (H~o(n + 1) P(n + 1In)illo(n + 1) P(n + lin») K(n + 1) i=1 j=1 +R P(n + 1In)l..o(n + 1)8-I (n + 1) ~(n + 1) l..o(n + l)~(n + lin) ~(n + lin) + K(n + l)!L(n + 1) !L(n + 1) ~(n + lin + 1) P(n + lin + 1) (I K(n + l)l..o(n + 1» P(n + lin) (I K(n + l)l..o(n + I»' + K(n + 1) R K(n + I)' Here, K(n + 1) is the Kalman Gain matrix and !L(n + 1) is the innovation signal. 3 BLACK-SCHOLES MODEL The Black-Scholes equation for calculating the price of an European style call option (Hull, 1993) is where, 2 In(8j X) + (r + -%-)0m (70m d1 (70m Here, C is the price of the call option, 8 the underlying asset price, X the strike price of the option at maturity, tm the time to maturity and r is the risk free interest rate. (7 is a term known as volatility and may be seen as an instantaneous variance of the time variation of the asset price. N(.) is the cumulative normal function. For a derivation of the formula and the assumptions upon which it is based see Hull, 1993. Readers unfamiliar with financial terms only need to know that all the quantities in the above equation, except (7, can be directly observed. (7 is usually estimated from a small moving window of data of about 50 trading days. The equivalent formula for the price of a put option is given by P = -8 N(-dl ) + X e- rt.,. N(-d2 ), For recursive estimation of the option prices with this model, I assume that the instantaneous picture given by the Black Scholes model is correct. The state vector is two dimensional and consists of the volatility (7 and the interest rate r . The Jacobian and Hessian required for applying EKF algorithm are ( aC Be ) ( a2c B'e ) ~. = ( a2p a2p ) au or au2 auar ; au2 auar 10 IDo aP aP a2c a2c a2p a2p au Or auar 8r2 auar ar2 Expressions for the terms in these matrices are given in table 1. Sequential Tracking of Financial Option Prices 963 Table 1: First and Second Derivatives of the Black Scholes Model aC aP S 0m N'(d1) au au aC Xtmexp( -rtm)N(d2 ) Or aP -Xtmexp( -rtm)N( -d2 ) Or a2C a2p S~dld2 N'(dd au2 au2 a2C -Xtmexp( -rtm) (tmN(d2 ) d2 'fFN' (d2 ) ) ar2 a2p Xtmexp( -rtm) (tmN( -d2 ) d2 'fFN' ( -d2 )) ar2 a2c a2p -S!lt", N'(dd auar aUaT 4 NEURAL NETWORK MODELS The data driven neural network model considered here is the Radial Basis FUnctions Network (RBF). Following Hutchinson et al, I use the following architecture: where U is the two dimensional input data vector consisting of the asset price and time to maturity. The asset price S is normalised by the strike price of the option X. The time to maturity, trn , is also normalised such that the full lifetime of the option gets a value 1.0. These normalisations is the reason for considering options in pairs with the same strike price and time of maturity in this study. The nonlinear function 1>(.) is set to 1>( Q) = va and m = 4. With the nonlinear part of the network fixed, Kalman filter training of the RBF model is straightforward (see Kadirkamanathan & Niranjan, 1993). In the simulations studied in this paper, I used two approaches to fix the nonlinear functions. The first was to use the /-l.S -J and the ~ published in Hutchinson et al. The second was to select the /-l . terms as -J random subsets of the training data and set ~ to I. The estimation problem is now linear and hence the Kalman filter equations become much simpler than the EKF equations used in the training of the Black-Scholes model. In addition to training by EKF, I also implemented a batch training of the RBF model in which a moving window of data was used, training on data from (n 50) to n days and testing on day (n + 1). Since it is natural to assume that data closer to the test day is more appropriate than data far back in time, I incorporated a weighting function to weight the errors linearly, in the minimisation process. The least squares solution, with a weighting function, is' given by the modified pseudo 964 M. Niranjan Table 2: Comparison of the Approximation Errors for Different Methods Strike Price Trivial RBF Batch RBF Kalman BS Historic BS Kalman 2925 0.0790 0.0632 0.0173 0.0845 0.0180 3025 0.0999 0.1109 0.0519 0.1628 0.0440 3125 0.0764 0.0455 0.0193 0.0343 0.0112 3225 0.1116 0.0819 0.0595 0.0885 0.0349 inverse l = (Y' W y)-l y' W t Matrix W is a diagonal matrix, consisting of the weighting function in its diagonal elements, t is the target values of options prices, and I is the vector containing the unknown coefficients AI, ... ,Am . The elements of Yare given by Yij = </>j(Ui ), with j = 1, ... ,m and i = n - 50, ... ,n. 5 SIMULATIONS The data set for teh experiments consisted of call and put option contracts on the FTSE-100 Index, during the period February 1994 to December 1994. The date of maturity of all contracts was December 1994. Five pairs (Call and Put) of contracts at strike prices of 2925, 3025, 3125, 3225, and 3325. The tracking behaviour of the EKF for one of the pairs is shown in Fig. 1 for a call/put pair with strike price 3125. Fig. 2 shows the trajectories of the underlying state vector for four different call/put option pairs. Table 2 shows the squared errors in the approximation errors computed over the last 100 days of data (allowing for an initial period of convergence of the recursive algorithms). 6 DISCUSSION This paper presents a sequential approach to tracking the price of options contracts. The sequential approach is based on the Extended Kalman Filter algorithm, and I show how it may be used to identify a parametric model of the underlying nonlinear system. The model based approach of the finance community and the data driven approach of neural computing community lead to good estimates of the observed price of the options contracts when estimated in this manner. In the state space formulation of the Black-Scholes model, the volatility and interest rate are estimated from the data. I trust the instantaneous picture presented by the model based approach, but reestimate the underlying parameters. This is different from conventional wisdom, where the risk free interest rate is set to some figure observed in the bond markets. The value of volatility that gives the correct options price through Black Scholes equation is called option implied volatility, and is usually different for different options. Option traders often use the differences in implied volatility to take trading positions. In the formulation presented here, there is an extra freedom coming in the form what one might call implied interest rates. It's difference from the interest rates observed in the markets might explain trader speculation about risk associated with a particular currency. The derivatives of the RBF model output with respect to its inputs is easy to compute. Hutchinson et a1 use this to define a highly relevant performance measure Sequential Tracking of Financial Option Prices 965 0.25 \ 0.2 ~ 0.15 0.1 0.05 (a) Estimated Call Option Price ---True ~~Estimaie O~--~----L---~----~--~----~----L----L----~--~----~ 20 40 60 80 100 120 140 160 180 200 220 (b) Estimated Put Option Price 0.1.---.----.----.----.----.----.----.----.----.----,----.. 0.08 0.06 ~ 0.04 0.02 20 40 60 80 ---True . :Estimate . . . 100 120 140 160 180 200 220 time Figure 1: Tracking Black-Scholes Model with EKF; Estimates of Call and Put Prices suitable to this particular application, namely the tracking error of a delta neutral portfolio. This is an evaluation that is somewhat unfair to the RBF model since at the time of training, the network is not shown the derivatives. An interesting combination of the work presented in this paper and Hutchinson et al's performance measure is to train the neural network to approximate the observed option prices and simultaneously force the derivative network to approximate the delta observed in the markets. References Bar-Shalom, Y. & Li, X-R. (1993), 'Estimation and Tracking: Principles, Techniques and Software', Artech House, London. Candy, J. V. (1986), 'Signal Processing: The Model Based Aproach', McGraw-Hill, New York. Hull, J. (1993), 'Options, Futures and Other Derivative Securities', Prentice Hall, NJ. Hutchinson, J. M., Lo, A. W. & Poggio, T. (1994), 'A Nonparametric Approach to Pricing and Hedging Derivative Securities Via Learning Networks', The Journal of Finance, Vol XLIX, No.3., 851-889. 966 M. Niranjan Trajectory of State Vector 0.055,....----.-----.----.-----,-----.----.-----,----. 0.05 0.045 0.04 Q) 0.035 iii c:: ~ 0.03 $ c: - 0.025 0.02 0.015 0.01 ..... .... ", : ........ ... . " o ·2925 x 3025 lIE 3125 .. + . ~225. 0.005L.----...J-----'----L...----"-----'----L...----"------' 0.14 0.16 0.18 0.2 0.22 0.24 0.26 0.28 0.3 Volatility Figure 2: Tracking Black-Scholes Model with EKF; Estimates of Call and Put Prices and the Trajectory of the State Vector Kadirkamanathan, V. & Niranjan, M {1993}, 'A Function Estimation Approach to Sequential Learning with Neural Networks', Neural Computation 5, pp. 954-975. Lowe, D. {1995}, 'On the use of Nonlocal and Non Positive Definite Basis Functions in Radial Basis Function Networks', Proceedings of the lEE Conference on Artificial Neural Networks, lEE Conference Publication No. 409, pp 206-211. Niranjan, M., Cox, I. J., Hingorani, S. {1994}, 'Recursive Estimation of Formants in Speech', Proceedings of the International Conference on Acoustics, Speech and Signal Processing, ICASSP '94, Adelaide. Puskorius, G.V. & Feldkamp, L.A. {1994}, 'Neurocontrol of Nonlinear Dynamical Systems with Kalman Filter-Trained Recurrent Networks', IEEE Transactions on Neural Networks, 5 {2}, pp 279-297.
1996
57
1,255
Are Hopfield Networks Faster Than Conventional Computers? Ian Parberry* and Hung-Li Tsengt Department of Computer Sciences University of North Texas P.O. Box 13886 Denton, TX 76203-6886 Abstract It is shown that conventional computers can be exponentiallx faster than planar Hopfield networks: although there are planar Hopfield networks that take exponential time to converge, a stable state of an arbitrary planar Hopfield network can be found by a conventional computer in polynomial time. The theory of 'P.cS-completeness gives strong evidence that such a separation is unlikely for nonplanar Hopfield networks, and it is demonstrated that this is also the case for several restricted classes of nonplanar Hopfield networks, including those who interconnection graphs are the class of bipartite graphs, graphs of degree 3, the dual of the knight's graph, the 8-neighbor mesh, the hypercube, the butterfly, the cube-connected cycles, and the shuffle-exchange graph. 1 Introduction Are Hopfield networks faster than conventional computers? This apparently straightforward question is complicated by the fact that conventional computers are universal computational devices, that is, they are capable of simulating any discrete computational device including Hopfield networks. Thus, a conventional computer could in a sense cheat by imitating the fastest Hopfield network possible. * Email: ianGcs. unt .edu. URL: http://hercule .csci. unt. edu/ian. t Email: ht sengGponder. csci. unt . edu. 240 I. Parberry and H. Tseng But the question remains, is it faster for a computer to imitate a Hopfield network, or to use other computational methods? Although the answer is likely to be different for different benchmark problems, and even for different computer architectures, we can make our results meaningful in the long term by measuring scalability, that is, how the running time of Hopfield networks and conventional computers increases with the size of any benchmark problem to be solved. Stated more technically, we are interested in the computational complexity of the stable state problem for Hopfield networks, which is defined succinctly as follows: given a Hopfield network, determine a stable configuration. As previously stated, this stable configuration can be determined by imitation, or by other means. The following results are known about the scalability of Hopfield network imitation. Any imitative algorithm for the stable state problem must take exponential time on some Hopfield networks, since there exist Hopfield networks that require exponential time to converge (Haken and Luby [4], Goles and Martinez [2]). It is unlikely that even non-imitative algorithms can solve the stable state problem in polynomial time, since the latter is PeS-complete (Papadimitriou, Schaffer, and Yannakakis [9]). However, the stable state problem is more difficult for some classes of Hopfield networks than others. Hopfield networks will converge in polynomial time if their weights are bounded in magnitude by a polynomial of the number of nodes (for an expository proof see Parberry [11, Corollary 8.3.4]). In contrast, the stable state problem for Hopfield networks whose interconnection graph is bipartite is peS-complete (this can be proved easily by adapting techniques from Bruck and Goodman [1]) which is strong evidence that it too requires superpolynomial time to solve even with a nonimitative algorithm. We show in this paper that although there exist planar Hopfield networks that take exponential time to converge in the worst case, the stable state problem for planar Hopfield networks can be solved in polynomial time by a non-imitative algorithm. This demonstrates that imitating planar Hopfield networks is exponentially slower than using non-imitative algorithmic techniques. In contrast, we discover that the stable state problem remains peS-complete for many simple classes of nonplanar Hopfield network, including bipartite networks, networks of degree 3, and some networks that are popular in neurocomputing and parallel computing. The main part of this manuscript is divided into four sections. Section 2 contains some background definitions and references. Section 3 contains our results about planar Hopfield networks. Section 4 describes our peS-completeness results, based on a pivotal lemma about a nonstandard type of graph embedding. 2 Background This section contains some background which are included for completeness but may be skipped on a first reading. It is divided into two subsections, the first on Hopfield networks, and the second on PeS-completeness. 2.1 Hopfield Networks 4- Hopfield network [6] is a discrete neural network model with symmetric connections. Each processor in the network computes a hard binary weighted threshold Are Hopfield Networks Faster than Conventional Computers? 241 function. Only one processor is permitted to change state at any given time. That processor becomes active if its excitation level exceeds its threshold, and inactive otherwise. A Hopfield network is said to be in a stable state if the states of all of its processors are consistent with their respective excitation levels. It is well-known that all Hopfield networks converge to a stable state. The proof defines a measure called energy, and demonstrates that energy is positive but decreases with every computation step. Essentially then, a Hopfield network finds a local minimum in some energy landscape. 2.2 P .cS-completeness While the theory of NP-completeness measures the complexity of global optimization, the theory of p.cS-completeness developed by Johnson, Papadimitriou, and Yannakakis [7] measures the complexity of local optimization. It is similar to the theory of NP-completeness in that it identifies a set of difficult problems known collectively as p.cS-complete problems. These are difficult in the sense that if a fast algorithm can be developed for any P .cS-complete problem, then it can be used to give fast algorithms for a substantial number of other local optimization problems including many important problems for which no fast algorithms are currently known. Recently, Papadimitriou, Schaffer, and Yannakakis [9] proved that the problem of finding stable states in Hopfield networks is P .cS-complete. 3 Planar Hopfield Networks A planar Hopfield network is one whose interconnection graph is planar, that is, can be drawn on the Euclidean plane without crossing edges. Haken and Luby [4] describe a planar Hopfield network that provably takes exponential time to converge, and hence any imitative algorithm for the stable state problem must take exponential time on some Hopfield network. Yet there exists a nonimitative algorithm for the stable state problem that runs in polynomial time on all Hopfield networks: Theorem 3.1 The stable state problem for Hopfield networks with planar interconnection pattern can be solved in polynomial time. PROOF: (Sketch.) The prooffollows from the fact that the maximal cut in a planar graph can be found in polynomial time (see, for example, Hadlock [3]), combined with results of Papadimitriou, Schaffer, and Yannakakis [9]. 0 4 P .cS-completeness Results Our P .cS-completeness results are a straightforward consequence of a new result that characterizes the difficulty of the stable state problem of an arbitrary class of Hopfield networks based on a graph-theoretic property of their interconnection patterns. Let G = (V, E) and H = (V', E') be graphs. An embedding of G into H is a function f: V -+ 2 Vi such that the following properties hold. (1) For all v E V, the subgraph of H induced by f(v) is connected. (2) For all (u, v) E E, there exists a path (which we will denote f(u , v)) in H from a member of f(u) to a member of f(v). (3) Each vertex w E H is used at most once, either as a member of f(v) 242 I. Parberry and H. Tseng for some v E V, or as an internal vertex in a path feu, v) for some u, v E V. The graph G is called the guest graph, and H is called the host graph. Our definition of embedding is different from the standard notion of embedding (see, for example, Hong, Mehlhorn, and Rosenberg [5]) in that we allow the image of a single guest vertex to be a set of host vertices, and we insist in properties (2) and (3) that the images of guest edges be distinct paths. The latter property is crucial to our results, and forms the major difficulty in the proofs. Let 5, T be sets of graphs. 5 is said to be polynomial-time embeddable into T, written 5 ::;e T, if there exists polynomials Pl(n),P2(n) and a function f with the following properties: (1) f can be computed in time PI(n), and (2) for every G E 5 with n vertices, there exists H E T with at most p2(n) vertices such that G can be embedded into H by f. A set 5 of graphs is said to be pliable if the set of all graphs is polynomial-time embeddable into 5. Lemma 4.1 If 5 is pliable, then the problem of finding a stable state in Hopfield networks with interconnection graphs in 5 is 'P £S-complete. PROOF: (Sketch.) Let 5 be a set of graphs with the property that the set of all graphs is polynomial-time embeddable into 5. By the results of Papadimitriou, Schaffer, and Yannakakis [9], it is enough to show that the max-cut problem for graphs in 5 is 'P £S-complete. Let G be an arbitrary labeled graph. Suppose G is embedded into H E 5 under the polynomial-time embedding. For each edge e in G of cost c, select one edge from the path connecting the vertices in f( e) and assign it cost c. We call this special edge f' ( e ). Assign all other edges in the path cost -00. For all v E V, assign the edges linking the vertices in f(v) a cost of -00. Assign all other edges of H a cost of zero. It can be shown that every cut in G induces a cut of the same cost in H, as follows. Suppose G ~ E is a cut in G, that is, a set of edges that if removed from G, disconnects it into two components containing vertices VI and V2 respectively. Then, removing vertices f'(G) and all zero-cost edges from H will disconnect it into two components containing vertices f(VI ) and f(V2 ) respectively. Furthermore, each cut of positive cost in H induces a cut of the same cost in G, since a positive cost cut in H cannot contain any edges of cost -00, and hence must consist only of f'(e) for some edges e E E. Therefore, every max-cost cut in H induces in polynomial time a max-cost cut in G. 0 We can now present our 'P £S-completeness results. A graph has degree 3 if all vertices are connected to at most 3 other vertices each. Theorem 4.2 The problem of finding stable states in Hopfield networks of degree 3 is 'P £S-complete. PROOF: (Sketch.) By Lemma 4.1, it suffices to prove that the set of degree-3 graphs is pliable. Suppose G = (V, E) is an arbitrary graph. Replace each degree-k vertex x E V by a path consisting of k vertices, and attach each edge incident with v by a new edge incident with one of the vertices in the path. Figure 1 shows an example of this embedding. 0 Are Hopfield Networks Faster than Conventional Computers? 243 Figure 1: A guest graph of degree 5 (left), and the corresponding host of degree 3 (right). Shading indicates the high-degree nodes that were embedded into paths. All other nodes were embedded into single nodes. Figure 2: An 8-neighbor mesh with 25 vertices (left), and the 8 X 8 knight's graph superimposed on an 8 x 8 board (right). The 8-neighbor mesh is the degree-8 graph G = (V, E) defined as follows: V = {1,2, ... ,m} x {1,2, ... ,n}, and vertex (u,v) is connected to vertices (u,v± 1), (u ± 1, v), (u ± 1, v ± 1). Figure 2 shows an 8-neighbor mesh with 25 vertices. Theorem 4.3 The problem of finding stable states in H opfield networks on the 8-neighbor mesh is peS-complete. PROOF: (Sketch.) By Lemma 4.1, it suffices to prove that the 8-neighbor mesh is pliable. An arbitrary graph can be embedded on an 8-neighbor mesh by mapping each node to a set of consecutive nodes in the bottom row of the grid, and mapping edges to disjoint rectilinear paths which use the diagonal edges of the grid for crossovers. 0 The knight's graph for an n X n chessboard is the graph G = (V, E) where V = {(i, j) 11 ~ i, j ~ n}, and E = {((i, j), (k, i» I {Ii - kl, Ij - il} = {I, 2}}. That is, there is a vertex for every square of the board and an edge between two vertices exactly when there is a knight's move from one to the other. For example, Figure 2 shows the knight's graph for the 8 x 8 chessboard. Takefuji and Lee [15] (see also Parberry [12]) use the dual of the knight's graph for a Hopfield-style network to solve the knight's tour problem. That is, they have a vertex Ve for each edge e of the knight's graph, and an edge between two vertices Vd and Ve when d and e share a common vertex in the knight's graph. 244 I. Parberry and H. Tseng Theorem 4.4 The problem of finding stable states in H opfield networks on the dual of the knight's graph is pes -complete. PROOF: (Sketch.) By Lemma4.1, it suffices to prove that the dual of the knight's graph is pliable. It can be shown that the knight 's graph is pliable using the technique of Theorem 4.3. It can also he proved that if a set S of graphs is pliable, then the set consisting of the duals of graphs in S is also pliable. 0 The hypercube is the graph with 2d nodes for some d, labelled with the binary representations of the d-bit natural numbers, in which two nodes are connected by an edge iff their labels differ in exactly one bit. The hypercube is an important graph for parallel computation (see, for example, Leighton [8], and Parberry [lOD. Theorem 4.5 The problem of finding stable states in Hopfield networks on the hypercube is peS-complete. PROOF : (Sketch.) By Lemma 4.1, it suffices to prove that the hypercube is pliable. Since the " ~e" relation is transitive, it further suffices by Theorem 4.2 to show that the set of degree-3 graphs is polynomial-time embeddable into the hypercube. To embed a degree-3 graph G into the hypercube, first break it into a degree-1 graph G1 and a degree-2 graph G2 . Since G2 consists of cycles, paths, and disconnected vertices, it can easily be embedded into a hypercube (since a hypercube is rich in cycles). G1 can be viewed as a permutation of vertices in G and can hence be realized using a hypercube implementation of Waksman 's permutation network [16]. o We conclude by stating PeS-completeness results for three more graphs that are important in the parallel computing literature the butterfly (see, for example, Leighton [8]), the cube-connected cycles (Preparata and Vuillemin [13D, and the shuffle-exchange (Stone [14]). The proofs use Lemma 4.1 and Theorem 4.5, and are omitted for conciseness. Theorem 4.6 The problem of finding stable states in Hopfield networks on the butterfly, the cube-connected cycles, and the shuffle-exchange is peS-complete. Conclusion Are Hopfield networks faster than conventional computers? The answer seems to be that it depends on the interconnection graph of the Hopfield network. Conventional nonimitative algorithms can be exponentially faster than planar Hopfield networks. The theory of peS-completeness shows us that such an exponential separation result is unlikely not only for nonplanar graphs, but even for simple nonplanar graphs such as bipartite graphs, graphs of degree 3, the dual of the knight's graph, the 8-neighbor mesh, the hypercube, the butterfly, the cube-connected cycles, and the shuffle-exchange graph. Acknowledgements The research described in this paper was supported by the National Science Foundation under grant number CCR- 9302917, and by the Air Force Office of Scientific Are Hopfield Networks Faster than Conventional Computers? 245 Research, Air Force Systems Command, USAF, under grant number F49620-93-10100. References [1] J. Bruck and J. W. Goodman. A generalized convergence theorem for neural networks. IEEE Transactions on Information Theory, 34(5):1089-1092, 1988. [2] E. Goles and S. Martinez. Exponential transient classes of symmetric neural networks for synchronous and sequential updating. Complex Systems, 3:589597, 1989. [3] F. Hadlock. Finding a maximum cut of a planar graph in polynomial time. SIAM Journal on Computing, 4(3):221-225, 1975. [4] A. Haken and M. Luby. Steepest descent can take exponential time for symmetric conenction networks. Complex Systems, 2:191-196,1988. [5] J.-W. Hong, K. Mehlhorn, and A.L. Rosenberg. Cost tradeoffs in graph embeddings. Journal of the ACM, 30:709-728,1983. [6] J. J . Hopfield. Neural networks and physical systems with emergent collective computational abilities. Proc. National Academy of Sciences, 79:2554-2558 , April 1982. [7] D. S. Johnson, C. H. Papadimitriou, and M. Yannakakis. How easy is local search? In 26th Annual Symposium on Foundations of Computer Science, pages 39-42. IEEE Computer Society Press, 1985. [8] F. T. Leighton. Introduction to Parallel Algorithms and Architectures: Arrays . Trees· Hypercubes. Morgan Kaufmann, 1992. [9] C. H. Papadimitrioll, A. A. Schaffer, and M. Yannakakis. On the complexity of local search. In Proceedings of the Twenty Second Annual ACM Symposium on Theory of Computing, pages 439-445. ACM Press, 1990. [10] I. Parberry. Parallel Complexity Theory. Research Notes in Theoretical Computer Science. Pitman Publishing, London, 1987. [11] I. Parberry. Circuit Complexity and Neural Networks. MIT Press, 1994. [12] I. Patberry. Scalability of a neural network for the knight's tour problem. Neurocomputing, 12:19-34, 1996. [13] F. P. Preparata and J. Vuillemin. The cube-connected cycles: A versatile network for parallel computation. Communications of the ACM, 24(5):300309, 1981. [14] H. S. Stone. Parallel processing with the perfect shuffle. IEEE Transactions on Computers, C-20(2):153-161, 1971. [15] Y. Takefuji and K. C. Lee. Neural network computing for knight's tour problems. Neurocomputing, 4(5):249-254, 1992. [16] A. Waksman. A permutation network. Journal of the ACM, 15(1):159-163, January 1968.
1996
58
1,256
Learning From Demonstration Stefan Schaal sschaal @cc .gatech.edu; http://www.cc.gatech.edulfac/Stefan.Schaal College of Computing, Georgia Tech, 801 Atlantic Drive, Atlanta, GA 30332-0280 ATR Human Information Processing, 2-2 Hikaridai, Seiko-cho, Soraku-gun, 619-02 Kyoto Abstract By now it is widely accepted that learning a task from scratch, i.e., without any prior knowledge, is a daunting undertaking. Humans, however, rarely attempt to learn from scratch. They extract initial biases as well as strategies how to approach a learning problem from instructions and/or demonstrations of other humans. For learning control, this paper investigates how learning from demonstration can be applied in the context of reinforcement learning. We consider priming the Q-function, the value function, the policy, and the model of the task dynamics as possible areas where demonstrations can speed up learning. In general nonlinear learning problems, only model-based reinforcement learning shows significant speed-up after a demonstration, while in the special case of linear quadratic regulator (LQR) problems, all methods profit from the demonstration. In an implementation of pole balancing on a complex anthropomorphic robot arm, we demonstrate that, when facing the complexities of real signal processing, model-based reinforcement learning offers the most robustness for LQR problems. Using the suggested methods, the robot learns pole balancing in just a single trial after a 30 second long demonstration of the human instructor. 1. INTRODUCTION Inductive supervised learning methods have reached a high level of sophistication. Given a data set and some prior information about its nature, a host of algorithms exist that can extract structure from this data by minimizing an error criterion. In learning control, however, the learning task is often less well defined. Here, the goal is to learn a policy, i.e., the appropriate actions in response to a perceived state, in order to steer a dynamical system to accomplish a task. As the task is usually described in terms of optimizing an arbitrary performance index, no direct training data exist which could be used to learn a controller in a supervised way. Even worse, the performance index may be defined over the long term behavior of the task, and a problem of temporal credit assignment arises in how to credit or blame actions in the past for the current performance. In such a setting, typical for reinforcement learning, learning a task from scratch can require a prohibitively timeconsuming amount of exploration of the state-action space in order to find a good policy. On the other hand, learning without prior knowledge seems to be an approach that is rarely taken in human and animal learning. Knowledge how to approach a new task can be transferred from previously learned tasks, and/or it can be extracted from the performance of a teacher. This opens the questions of how learning control can profit from these kinds of information in order to accomplish a new task more quickly. In this paper we will focus on learning from demonstration. Learning from demonstration, also known as "programming by demonstration", "imitation learning", and "teaching by showing" received significant attention in automatic robot assembly over the last 20 years. The goal was to replace the time-consuming manual proLearningfrom Demonstration a,s (a) mI2jj = -/.tiH mg/sin9 + r . 9 E [-tr.tr] r(9.r)= (;)' _{;)210gCO{~ r:) m =1 = I. g = 9.81, J.I = 0.05. r..., =5Nm define : x = (9.8)'. U = r (b) mlXcos9 + ml2e - mglsin9 = 0 (m + me)x + m/ecos9 - mU'J2 sin9 = F define: x = (x,i,9.e) T, u = F r(x.u) = XTQX + uTRu 1= 0.75m. m = 0.15kg, me = LOkg Q = diag(1.25, I. 12. 0.25). R = om Figure 1 : a) pendulum swing up, b) cart pole balancing 1041 gramming of a robot by an automatic programming process, solely driven by showing the robot the assembly task by an expert. In concert with the main stream of Artificial Intelligence at the time, research was driven by symbolic approaches: the expert's demonstration was segmented into primitive assembly actions and spatial relationships between manipulator and environment, and subsequently submitted to symbolic reasoning processes (e.g., LozanoPerez, 1982; Dufay & Latombe, 1983; Segre & Dejong, 1985). More recent approaches to programming by demonstration started to include more inductive learning components (e.g., Ikeuchi, 1993; Dillmann, Kaiser, & Ude, 1995). In the context of human skill learning, teaching by showing was investigated by Kawato, Gandolfo, Gomi, & Wada (1994) and Miyamoto et al. (1996) for a complex manipulation task to be learned by an anthropomorphic robot arm. An overview of several other projects can be found in Bakker & Kuniyoshi (1996). In this paper, the focus lies on reinforcement learning and how learning from demonstration can be beneficial in this context. We divide reinforcement learning into two categories: reinforcement learning for nonlinear tasks (Section 2) and for (approximately) linear tasks (Section 3), and investigate how methods like Q-Iearning, valuefunction learning, and model-based reinforcement learning can profit from data from a demonstration. In Section 2.3, one example task, pole balancing, is placed in the context of using an actual, anthropomorphic robot to learn it, and we reconsider the applicability of learning from demonstration in this more complex situation. 2. REINFORCEMENT LEARNING FROM DEMONSTRATION Two example tasks will be the basis of our investigation of learning from demonstration. The nonlinear task is the "pendulum swing-up with limited torque" (Atkeson, 1994; Doya, 19%), as shown in Figure 1a. The goal is to balance the pendulum in an upright position starting from hanging downward. As the maximal torque available is restricted such that the pendulum cannot be supported against gravity in all states, a "pumping" trajectory is necessary, similar as in the mountain car example of Moore (1991), but more delicately in its timing since building up too much momentum during pumping will overshoot the upright position. The (approximately) linear example, Figure 1b, is the well-known cart-pole balancing problem (Widrow & Smith, 1964; Barto, Sutton, & Anderson, 1983). For both tasks, the learner is given information about the one-step reward r (Figure 1), and both tasks are formulated as continuous state and continuous action problems. The goal of each task is to find a policy which minimizes the infinite horizon discounted reward: co (S-I) 00 v(x(t)) = J e --~ r(x(s), u(s))ds or V(x(t)) = L ri-1r(x(i), u(i)) (1) i=t where the left hand equation is the continuous time formulation, while the right hand equation is the corresponding discrete time version, and where x and u denote a ndimensional state vector and a m-dimensional command vector, respectively. For the Swing-Up, we assume that a teacher provided us with 5 successful trials starting from dif1042 S. Schaal ferent initial conditions. Each trial consists of a time series of data vectors (0, e, -r) sampled at 60Hz. For the Cart-Pole, we have a 30 second demonstration of successful balancing, represented as a 60Hz time series of data vectors (x, X, 0, e, F). How can these demonstrations be used to speed up reinforcement learning? 2.1 THE NONLINEAR TASK: SWING-UP We applied reinforcement learning based on learning a value function (V-function) (Dyer & McReynolds, 1970) for the Swing-Up task, as the alternative method, Q-learning (Watkins, 1989), has yet received very limited research for continuous state-action spaces. The V-function assigns a scalar reward value V(x(t») to each state x such that the entire Vfunction fulfills the consistency equation: V(x(t)) = arg min(r(x(t), u(t)) + r V(x(t + 1))) (2) u(t) For clarity, this equation is given for a discrete state-action system; the continuous formulation can be found, e.g., in Doya (1996). The optimal policy, u =Jt(x), chooses the action u in state x such that (2) is fulfilled. Note that this computation involves an optimization step that includes knowledge of the subsequent state x(t+ 1). Hence, it requires a model of the dynamics of the controlled system, x(t+ 1)=f(x(t),u(t». From the viewpoint of learning from demonstration, V-function learning offers three candidates which can be primed from a demonstration: the value function V(x), the policy 1t(x), and the modelf(x,u). 60 2.1.1 V-Learning 50-+-----------------+------~~'§.JJl~ r~30+----------r~-+------------~ 20+-----~~~---+------------~ In order to assess the benefits of a demonstration for the Swing-Up, we implemented V-learning as suggested in Doya's (1996) continuous TD (CTD) learning algorithm. The V-function and the dynamics model were incrementally learned by a 10 ...,-" -- a)scratch c) primed model nonlinear function approximator, Recepo - b) primed actor d) primed actor&model tive Field Weighted Regression (RFWR) 10 100 (Schaal & Atkeson (1996»). Differing Trial from Doya's (1996) implementation, we Figure 2: Smoothed learning curves of the average used the optimal action suggested by CTD of 10 learning trials for the learning conditions a) to learn a model of the policy 1t (an to d) (see text). Good performance is characterized "actor" as in Barto et al. (1983», again reby T up >45s; below this value the system is usupresented by RFWR. The following learnally able to swing up properly but it does not know ing conditions were tested empirically: b) c) d) how to stop in the upright position. a) Scratch: Trial by trial learning of value function V, model f, and actor 1t from scratch. Primed Actor: Initial training of 1t from the demonstration, then trial by trial learning. Primed Model: Initial training of f from the demonstration, then trial by trial learning. Primed Actor&Model: Priming of 1t and f as in b) and c), then trial by trial learning. Figure 2 shows the results of learning the Swing-Up. Each trial lasted 60 seconds. The time Tup the pole spent in the interval ° E [-7r / 2, 7r /2] during each trial was taken as the performance measure (Doya, 1996). Comparing conditions a) and c), the results demonstrate that learning the pole model from the demonstration did not speed up learning. This is not surprising since learning the V-function is significantly more complicated than learning the model, such that the learning process is dominated by V-function learning. Interestingly, priming the actor from the demonstration had a significant effect on the initial performance (condition a) vs. b»). The system knew right away how to pump up the pendUlum, but, in order to learn how to balance the pendulum in the upright position, it finally took the same amount of time as learning from scratch. This behavior is due to the Learning from Demonstration 1043 fact that, theoretically, the V-function can only be approximated correctly if the entire state-action space is explored densely. Only if the demonstration covered a large fraction of the entire state space one would expect that V-learning can profit from it. We also investigated using the demonstration to prime the V-function by itself or in combination with the other functions. The results were qualitatively the same as in shown in Figure 2: if the policy was included in the priming, the learning traces were like b) and d), otherwise like a) and c). Again, this is not totally surprising. Approximating a V-function is not just supervised learning as for ;t and f, it requires an iterative procedure to ensure the validity of (2) and amounts to a complicated nonstationary function approximation process. Given the limited amount of data from the demonstration, it is generally very unlikely to approximate a good value function. 60 50 ~ --/' / V 40 1--~30 20 I a)scratch 10 /' .....,.. -. b) primed model 2.1.2 Model-Based V-Learning If learning a model f is required, one can make more powerful use of it. According to the certainty equivalence principle, f can substitute the real world, and planning can be run in "mental simulations" instead of interaction with the real world. In reinforcement learning, this idea was originally pursued by Sutton's (1990) DYNA o 10 Trial 100) algorithms for discrete state-action spaces. Figure 3: Smoothed learning curves of the average of 10 learning trials for the learning conditions a) and b) (see text) of the Swing-Up problem using "mental simulations". See Figure 2 for explanations how to interpret the graph. Here we will explore in how far a continuous version of DYNA, DYNA-CTD, can help in learning from demonstration. The only difference compared to CTD in Section 2.1.1 is that after every real trial, DYNA-CTD performs five "mental trials" in which the model of the dynamics acquired so far replaces the actual pole dynamics. Two learning conditions we be explored: a) Scratch: Trial by trial learning of V, model f, and policy ;t from scratch. b) Primed Model: Initial training of f from the demonstration, then trial by trial learning. Figure 3 demonstrates that in contrast to V-learning in the previous section, learning from demonstration can make a significant difference now: after the demonstration, it only takes about 2-3 trials to accomplish a good swing-up with stable balancing, indicated by T up >45s. Note that also learning from scratch is significantly faster than in Figure 2. 2.2 THE LINEAR TASK: CART -POLE BALANCING One might argue that applying reinforcement learning from demonstration to the SwingUp task is premature, since reinforcement learning with nonlinear function approximators has yet to obtain appropriate scientific understanding. Thus, in this section we turn to an easier task: the cart-pole balancer. The task is approximately linear if the pole is started in a close to upright position, and the problem has been well studied in the dynamic programming literature in the context of linear quadratic regulation (LQR) (Dyer & McReynolds, 1970). 2.2.1 Q-Learning In contrast to V-learning, Q-Iearning (Watkins, 1989; Singh & Sutton, 1996) learns a more complicated value function, Q(x,u), which depends both on the state and the command. The analogue of the consistency equation (2) for Q-Iearning is: Q(x(t), u(t») = r(x(t), u(t») + r arg min(Q(x(t + 1), u(t + 1»)) (3) u(I+1) 1044 S. Schaal At every state x, picking the action u which minimizes Q is the optimal action under the reward function (l). As an advantage, evaluating the Q-function to find the optimal policy does not require a model the dynamical system f that is to be controlled; only the value of the one-step reward r is needed. For learning from demonstration, priming the Qfunction and/or the policy are the two candidates to speed up learning. For LQR problems, Bradtke (1993) suggested a Q-Iearning method that is ideally suited for learning from demonstration, based on extracting a policy. He observed that for LQR the Q-function is quadratic in the states and commands: Q(x,u) = [xT,uTl[HHIl H I2][XT,UTY, HIl =nxn, H22 =mxm, HI2 =H;I =nxm (4) 21 H22 0.045 0.04 0.035 '0 ~ 0.03 £ 0.025 a. ~ 0.02 ~ 0.Q15 o 0.01 KdOmo = [·0.59. -1.81. -18.71. ·6.67) ~nal = [·5.76, -11.37, -83.05, -21.92) and that the (linear) policy, represented as a gain matrix K, can be extracted from (4) as: uopt = -K x = -H;;H2Ix (5) Conversely, given a stabilizing initial policy K demo' the current Q-function can be approximated by a recursive least squares procedure, and it can be optimized by a policy iteration 0.005 process with guaranteed convergence (Bradkte, o-tJcr~Dl"!NI(fj~~~~II!~ 1993). As a demonstration allows one to extract o 20 40 60 80 100 120 an initial policy K demo by linearly regressing Time[s) Figure 4: Typical learning curve of a noisy simulation of the cart-pole balancer using Qlearning. The graph shows the value of the one-step reward over time for the first learning trial. The pole is never dropped. the observed command u against the corresponding observed states x, one-shot learning of pole balancing is achievable. As shown in Figure 4, after about 120 seconds (12 policy iteration steps), the policy is basically indistinguishable from the optimal policy. A caveat of this Q-Iearning, however, is that it cannot not learn without a stabilizing initial policy. 2.2.2 Model-based V -Learning Learning an LQR task by learning the V-function is one of the classic forms of dynamic programming (Dyer & McReynolds, 1970). Using a stabilizing initial policy K demo' the current V-function can be approximated by recursive least squares in analogy with Bradtke (1993). Similarly as for K demo' a (linear) model f demo of the cart-pole dynamics can be extracted from a demonstration by linear regression of the cart-pole state x(t) vs. the previous state and command vector (x(t-1), u(t-1», and the model can be refined with every new data point experienced during learning. The policy update becomes: K= y(R + yBTHBtBTHA, where Vex) = xTHx, idemo = [AB], A = n x n,B = n X m (6) Thus, a similar process as in Bradtke (1993) can be used to find the optimal policy K, and the system accomplishes one shot learning, qualitatively indistinguishable from Figure 4. Again, as pointed out in Section 2.1.2, one can make more efficient use of the learned model by performing mental simulations. Given the model f demo' the policy K can be calculated by off-line policy iteration from an initial estimate ofH, e.g., taken to be the identity matrix (Dyer & McReynolds, 1970). Thus, no initial (stabilizing) policy is required, but rather an estimate of the task dynamics. Also this method achieves one shot learning. 2.3 POLE BALANCING WITH AN ACTUAL ROBOT As a result of the previous section, it seems that there are no real performance differences between V-learning, Q-Iearning, and model-based V-learning for LQR problems. To explore the usefulness of these methods in a more realistic framework, we implemented Leamingfrom Demonstration 1045 learning from demonstration of pole balancing on an anthropomorphic robot arm. The robot is equipped with a 60 Hz video-based stereo vision. The pole is marked by two color blobs which can be tracked in real-time. A 30 second long demonstration of pole balaocing was is provided by a human standing in front of the two robot cameras. There are a few crucial differences in comparison with the simulations. First, as the demonstration is vision-based, only kinematic variables can be extracted from the demonstration. Second, visual signal processing has about 120ms time delay. Third, a command given to the robot is not executed with very high accuracy due to unknown nonlinearities of the robot. And lastly, humans use internal state for pole balancing, i.e., their policy is partially based on non-observable variables. These issues have the following impact: Kinematic Variables: In this implementation, the robot arm replaces the cart of the Cart-Pole problem. Since we have an estimate of the inverse dynamics and inverse kinematics of the arm, we can use the acceleration of the finger in Cartesian space as command input to the task. The arm is also much heavier than the pole which allows us to neglect the interaction forces the pole exerts on the arm. Thus, the pole balancing dynamics of Figure Ib can be reformulated as: , .. i uml cosO + Oml 2 - mgl sin 0= 0, x = u (7) Figure 5: Sketch of SARCOS An variables in this equation can be extracted from a demanthropomorphic robot arm onstration. We omit the 3D extension of these equations. Delayed Visual Information: There are two possibilities of dealing with delayed variables. Either the state of the system is augmented by delayed commands corresponding to 7* 1/60s:::::120s delay time, x T = (x, x,O, (}, Ut_1' ut- 2 ' ... , ut- 7 ) , or a state predictive controller is employed. The former method increases the complexity of a policy significantly, while the latter method requires a model f. Inaccuracies of Command Execution: Given an acceleration command u, the robot will execute something close to u, but not u exactly. Thus, learning a function which includes u, e.g., the dynamics model (7), can be dangerous since the mapping (x,i,O,(},u) ~ (x,ii) is contaminated by the nonlinear dynamics of the robot arm. Indeed, it turned out that we could not learn such a model reliably. This could be remedied by "observing" the command u, i.e., by extracting u = x from visual feedback. Internal State in Demonstrated Policy: Investigations with human subjects have shown that humans use internal state in pole balancing. Thus, a policy cannot be observed that easily anymore as claimed in Section 2.2: a regression analysis for extracting the policy of a teacher must find the a~propriate time-alignment of observed current state and command(s) in the past. This can become a numerically involved process as regressing a policy based on delayed commands is endangered by singUlar regression matrices. Consequently, it easily happens that one extracts a nonstabilizing policy from the demonstration, which prevents the application of Q-Iearning and V-learning as described in Section 2.2. As a result of these considerations, the most trustworthy item to extract from a demonstration is the model of the pole dynamics. In our implementation it was used in two ways, for calculating the policy as in (6), and in state-predictive control with a Kalman filter to overcome the delays in visual information processing. The model was learned incrementally in real-time by an implementation of RFWR (Schaal & Atkeson 1996). Figure 6 shows the results of learning from scratch and learning from demonstration of the actual robot. Without a demonstration, it took about 10-20 trials before learning succeeded in reliable performance longer than one minute. With a 30 second long demonstration, learning was reliably accomplished in one single trial, using a large variety of physically different poles and using demonstrations from arbitrary people in the laboratory. 1046 ~~------------------, 70 60+--- ------.,..------1 50 ,.'l' 40 30 20 S. Schaal 3. CONCLUSION We discussed learning from demonstration in the context of reinforcement learning, focusing on Qlearning, value function learning, and model based reinforcement learning. Q-Iearning and value function learning can theoretically profit from a demonstration by extracting a policy, by a)scratcn _ b) primed model using the demonstration data to prime the Q/value O-fl=:::::;:==;:::;:::;='~~lrO -L:::;::'.::!.;:.~~~I00 function, or, in the case of value function learn10 /!Trial ing, by extracting a predictive model of the Figure 6: Smoothed average of 10 learning curves of the robot for pole balancing. The trials were aborted after successful balancing of 60 seconds. We also tested long term performance of the learning system by running pole balancing for over an hour-the pole was never dropped. world. Only in the special case of LQR problems, however, could we find a significant benefit of priming the learner from the demonstration. In contrast, model-based reinforcement learning was able to greatly profit from the demonstration by using the predictive model of the world for "mental simulations". In an implementation with an anthropomorphic robot arm, we illustrated that even in LQR problems, model-based reinforcement learning offers larger robustness towards the complexity in real learning systems than Q-Iearning and value function learning. Using a model-based strategy, our robot learned pole-balancing from a demonstration in a single trial with great reliability. The important message of this work is that not every learning approach is equally suited to allow knowledge transfer and/or the incorporation of biases. This issue may serve as a critical additional constraint to evaluate artificial and biological models of learning. Acknowledgments Ikeuchi, K. (1993b). "Assembly plan from observation.", School of Computer Science, Carnegie Mellon Support was provided by the A TR Human Infor- University, Pittsburgh, PA. mation Processing Labs the German Research Kawato, M., Gandolfo, F., Gomi, H., & Wada, Y. . . 'b (1994b). "Teaching by showing in kendama based on ~ssocIatlOn, the Alexander v. ~um oldt ~ounda- optimization principle." In: Proceedings of the InternatlOn, and the German ScholarshIp FoundatIOn. tional Conference on Artificial Neural Networks (ICANN'94), 1, pp.601-606. References Lozano-Perez, T. (1982). "Task-Planning." In: Brady, Atkeson, C. G. (1994). "Using local trajectory optimiz- M., Hollerbach, 1. M., Johnson, T. L., Lozano-P _rez, T., ers to speed up global optimization in dynamic pro- & Mason, M. T. (Eds.), , pp,473-498. MIT Press. gramming." In: Moody, Hanson, & Lippmann (Ed.), Miyamoto, H., Schaal, S., Gandolfo, F., Koike, Y., Osu, Adv. in Neural In! Proc. Sys. 6. Morgan Kaufmann. R., Nakano, E., Wada, Y., & Kawato, M. (in press). "A Bakker, P., & Kuniyoshi, Y. (1996). "Robot see, robot Kendama learning robot based on bi-directional theory." do: An overview of robot imitation.", Autonomous Neural Networks. Systems Section, Electrotechnical Laboratory, Tsukuba Moore, A. (1991a). "Fast, robust adaptive control by Science City, Japan. learning only forward models." In: Moody, 1. E., HanBarto, A. G., Sutton, R. S., & Anderson, C. W. (1983). son, S. J., & and Lippmann, R. P. (Eds.), Advances in "Neuronlike adaptive elements that can solve difficult NeuralInt Proc. Systems 4. Morgan Kaufmann. learning control problems." IEEE Transactions on Sys- Schaal, S., & Atkeson, C. G. (1996). "From isolation to tems, Man, and Cybernetics, SMC-13, 5. cooperation: An alternative of a system of experts." In: Bradtke, S. 1. (1993). "Reinforcement learning applied Touretzky, D. S., Mozer, M. c., & Hasselmo, M. E. to linear quadratic regulation." In: Hanson, J. S., Cowan, (Eds.), Advances in Neural Information Processing 1. D., & Giles, C. L. (Eds.), Advances in Neural In! Systems 8. Cambridge, MA: MIT Press. Processing Systems 5, pp.295-302. Morgan Kaufmann. Segre, A. B., & Dejong, G. (1985). "Explanation-based Dillmann, R., Kaiser, M., & Ude, A. (1995). manipulator learning: Acquisition of planning ability "Acquisition of elementary robot skills from human through observation." In: Conference on Robotics and demonstration." In: International Symposium on Intelli- Automation, pp.555-560. gent Robotic Systems (SIRS'95), Pisa, Italy. Singh, S. P., & Sutton, R. S. (1996). "Reinforcement Doya, K. (1996). "Temporal difference learning in con- learning with eligibility traces." Machine Learning. tinuous time and space." In: Touretzky, D. S., Mozer, Sutton, R. S. (1990). "Integrated architectures for M. c., & Hasselmo, M. E. (Eds.), Advances in Neural learning, planning, and reacting based on approximating Information Processing Systems 8. MIT Press. dynamic programming." In: Proceedings of the InternaDufay, B., & Latombe, J.-c. (1984). "An approach to tional Machine Learning Conference. automatic robot programming based on inductive Watkins, C. 1. C. H. (1989). "Learning with delayed relearning." In: Brady, M., & Paul, R. (Eds.), Robotics Re- wards." Ph.D. thesis, Cambridge University (UK), . search, pp.97-115. Cambridge, MA: MIT Press. Widrow, B., & Smith, F. W. (1964). "Pattern recognizDyer, P., & Mc~eynolds, S. R. (1970). The ~omputation ing control systems." In: 1963 Compo and In! Sciences and theory of opumal control. NY: AcademIC Press. (COINS) Symp. Proc., 288-317, Washington: Spartan.
1996
59
1,257
Adaptive On-line Learning in Changing Environments Noboru Murata, Klaus-Robert Miiller, Andreas Ziehe GMD-First, Rudower Chaussee 5, 12489 Berlin, Germany {mura.klaus.ziehe}~first.gmd.de Shun-ichi Amari Laboratory for Information Representation, RIKEN Hirosawa 2-1, Wako-shi, Saitama 351-01, Japan amari~zoo.riken.go.jp Abstract An adaptive on-line algorithm extending the learning of learning idea is proposed and theoretically motivated. Relying only on gradient flow information it can be applied to learning continuous functions or distributions, even when no explicit loss function is given and the Hessian is not available. Its efficiency is demonstrated for a non-stationary blind separation task of acoustic signals. 1 Introduction Neural networks provide powerful tools to capture the structure in data by learning. Often the batch learning paradigm is assumed, where the learner is given all training examples simultaneously and allowed to use them as often as desired. In large practical applications batch learning is often experienced to be rather infeasible and instead on-line learning is employed. In the on-line learning scenario only one example is given at a time and then discarded after learning. So it is less memory consuming and at the same time it fits well into more natural learning, where the learner receives new information and should adapt to it, without having a large memory for storing old data. On-line learning has been analyzed extensively within the framework of statistics (Robbins & Monro [1951]' Amari [1967] and others) and statistical mechanics (see ego Saad & Solla [1995]). It was shown that on-line learning is asymptotically as effective as batch 600 N. Murata, K. MUller, A. Ziehe and S. Amari learning (cf. Robbins & Monro [1951]). However this only holds, if the appropriate learning rate 1] is chosen. A too large 1] spoils the convergence of learning. In earlier work on dichotomies Sompolinsky et al. [1995] showed the effect on the rate of convergence of the generalization error of a constant, annealed and adaptive learning rate. In particular, the annealed learning rate provides an optimal convergence rate, however it cannot follow changes in the environment. Since on-line learning aims to follow the change of the rule which generated the data, Sompolinsky et al. [1995], Darken & Moody [1991] and Sutton [1992] proposed adaptive learning rates, which learn how to learn. Recently Cichoki et al. [1996] proposed an adaptive online learning algorithm for blind separation based on low pass filtering to stabilize learning. We will extend the reasoning of Sompolinsky et al. in several points: (1) we give an adaptive learning rule for learning continuous functions (section 3) and (2) we consider the case, where no explicit loss function is given and the Hessian cannot be accessed (section 4). This will help us to apply our idea to the problem of on-line blind separation in a changing environment (section 5). 2 On-line Learning Let us consider an infinite sequence of independent examples (Zl, Yl), (Z2' Y2), .... The purpose of learning is to obtain a network with parameter W which can simulate the rule inherent to this data. To this end, the neural network modifies its parameter Wt at time t into Wt+1 by using only the next example (Zt+l' Yt+l) given by the rule. We introduce a loss function l(z, Y; w) to evaluate the performance of the network with parameter w. Let R(w) = (l(z, Y; w)) be the expected loss or the generalization error of the network having parameter w, where ( ) denotes the average over the distribution of examples (z, y). The parameter w* of the best machine is given by w* = argminR(w). We use the following stochastic gradient descent algorithm (see Amari [1967] and Rumelhart et al. [1986]): Wt+l = Wt -1]tC(Wt) 8~ I(Zt+l' Yt+l; Wt), (1) where 1]t is the learning rate which may depend on t and C(Wt) is a positive-definite matrix which rr,ay depend on Wt. The matrix C plays the role of the Riemannian metric tensor of the underlying parameter space {w}. When 1]t is fixed to be equal to a small constant 1], E[wt] converges to w* and Var[wt] converges to a non-zero matrix which is order 0(1]). It means that Wt fluctuates around w* (see Amari [1967], Heskes & Kappen [1991]). If 1]t = cit (annealed learning rate) Wt converges to w* locally (Sompolinsky et al. [1995]). However when the rule changes over time, an annealed learning rate cannot follow the changes fast enough since 1]t = cit is too small. 3 Adapti ve Learning Rate The idea of an adaptively changing 1]t was called learning of the learning rule (Sompolinsky et al. [1995]). In this section we investigate an extension of this idea to differentiable loss functions. Following their algorithm, we consider Wt - 1]tK - 1(wt) 8~ l(zt+1' Yt+1; Wt), (2) Adaptive On-line Learning in Changing Environments 601 (3) where c¥ and f3 are constants, K( Wt) is a Hessian matrix of the expected loss function 82 R(Wt)/8w8w and R is an estimator of R(woO). Intuitively speaking, the coefficient 'I in Eq.(3) is controlled by the remaining error. When the error is large, 'I takes a relatively large value. When the error is small, it means that the estimated parameter is close to the optimal parameter; 'I approaches to ° automatically. However, for the above algorithm all quantities (K, I, R) have to be accessible which they are certainly not in general. Furthermore I(Zt+l' Yt+1; Wt) - R could take negative values. Nevertheless in order to still get an intuition of the learning behaviour, we use the continuous versions of (2) and (3), averaged with respect to the current input-output pair (Zt, Yt) and we omit correlations and variances between the quantities ('It, Wt, I) for the sake of simplicity Noting that (81(z, Y; woO)/8w) = 0, we have the asymptotic evaluations (8~ I(z, Y; W t») KoO(Wt - woO), (/(z, Y; Wt) - R) R(woO) - R+ ~(Wt - woO? KoO(wt - woO), with KoO = 82R(w*)/8w8w. AssumingR(woO)-Ris small and K(wt) ':::: KoO yields :t Wt = -TJt(Wt - woO), !TJt = C¥TJt (~(Wt - woO? K*(wt - woO) - 'It) . (4) Introducing the squared error et = HWt - woOf KoO(Wt - woO), gives rise to (5) The behavior of the above equation system is interesting: The origin (0,0) is its attractor and the basin of attraction has a fractal boundary. Starting from an adequate initial value, it has the solution of the form 1 1 'It = - . -. 2 t (6) It is important to note that this l/t-convergence rate of the generalization error et is the optimal order of any estimator Wt converging to woO. So we find that Eq.( 4) gives us an on-line learning algorithm which converges with a fast rate. This holds also if the target rule is slowly fluctuating or suddenly changing. The technique to prove convergence was to use the scalar distance in weight space et. Note also that Eq.(6) holds only within an appropriate parameter range; for small 'I and Wt - woO correlations and variances between ('It, Wt, I) can no longer be neglected. 4 Modification From the practical point of view (1) the Hessian KoO of the expected loss or (2) the minimum value of the expected loss R are in general not known or (3) in some 602 N Murata, K. Muller, A. Ziehe and S. Amari applications we cannot access the explicit loss function (e.g. blind separation). Let us therefore consider a generalized learning algorithm: (7) where I is a flow which determines the modification when an example (Zt+1' Yt+l) is given. Here we do not assume the existence of a loss function and we only assume that the averaged flow vanishes at the optimal parameter, i.e. (f(z, Y; w*») = o. With a loss function, the flow corresponds to the gradient of the loss. We consider the averaged continuous equation and expand it around the optimal parameter: (8) where K* = (81(z, Y; w*)/8w). Suppose that we have an eigenvector of the Hessian K* vector v satisfying v T K* = ).vT and let us define (9) then the dynamics of e can be approximately represented as (10) By using e, we define a discrete and continuous modification of the rule for 1]: 1]t+l = 1]t + a1]t (Plet I - 1]d and (11) Intuitively e corresponds to a I-dimensional pseudo distance, where the average flow I is projected down to a single direction v. The idea is to choose a clever direction such that it is sufficient to observe all dynamics of the flow only along this projection. In this sense the scalar e is the simplest obtainable value to observe learning. Noting that e is always positive or negative depending on its initial value and 1] can be positive, these two equations (10) and (11) are equivalent to the equation system (5). Therefore their asymptotic solutions are and 1 1 1]t = - . -. ). t (12) Again similar to the last section we have shown that the algorithm converges properly, however this time without using loss or Hessian. In this algorithm, an important problem is how to get a good projection v. Here we assume the following facts and approximate the previous algorithm: (1) the minimum eigenvalue of matrix f{* is sufficiently smaller than the second minimum eigenvalue and (2) therefore after a large number of iterations, the parameter vector Wt will approach from the direction of the minimum eigenvector of K*. Since under these conditions the evolution of the estimated parameter can be thought of as a one-dimensional process, any vector can be used as v except for the vectors which are orthogonal to the minimum eigenvector. The most efficient vector will be the minimum eigenvector itself which can be approximated (for a large number of iterations) by v = (/)/11(/)11, Adaptive On-line Learning in Changing Environments 603 where II " denotes the L2 norm. Hence we can adopt e = II(f)II. Substituting the instantaneous average of the flow by a leaky average, we arrive at Wt 1Jd(~t+l' Yt+l; Wt), (1- 8)rt + 8f(~t+l,Yt+1;Wt), (0 < 8 < 1) 1Jt + a1Jt (,811rt+l" - 1Jt) , (13) (14) (15) where 8 controls the leakiness of the average and r is used as auxiliary variable to calculate the leaky average of the flow f. This set of rules is easy to compute. However 1J will now approach a small value because of fluctuations in the estimation of r which depend on the choice of a,,8, 'Y. In practice, to assure the stability of the algorithm, the learning rate in Eq.(13) should be limited to a maximum value 1Jrnax and a cut-off 1Jrnin should be imposed. 5 Numerical Experiment: an application to blind separation In the following we will describe the blind separation experiment that we conducted (see ego Bell & Sejnowski [1995], Jutten & Herault [1991]' Molgedey & Schuster [1994] for more details on blind separation). As an example we use the two sun audio files (sampling rate 8kHz): "rooster" (sD and "space music" (s;) (see Fig. 1). Both sources are mixed on the computer via J; = (lL. + A)s~ where Os < t < 1.25s and 3.75s ~ t ~ 5s and it = (lL. + B)s~ for 1.25s ~ t < 3.75s, using A = (0 0.9; 0.7 0) and B = (0 0.8; 0.6 0) as mixing matrices. So the rule switches twice in the given data. The goal is to obtain the sources s-; by estimating A and B, given only the measured mixed signals it. A change of the mixing is a scenario often encountered in real blind separation tasks, e.g. a speaker turns his head or moves during his utterances. Our on-line algorithm is especially suited to this nonstationary separation task, since adaptation is not limited by the above-discussed generic drawbacks of a constant learning rate as in Bell & Sejnowski [1995], Jutten & Herault [1991]' Molgedey & Schuster [1994]. Let u~ be the unmixed signals (16) where T is the estimated mixing matrix. Along the lines of Molgedey & Schuster [1994] we use as modification rule for Tt ~T:j ex 1Jtt ((I!u{), (u~u{), (I!UL1), (U~UL1)) ex 1Jt(I!u{)(u~u{)+(I! UL1)(U~uLl)' (i,j = 1, 2,i # j), where we substitute instantaneous averages with leaky averages (I! U{)leaky = (1 - t)(ILl UL1)leaky + d! u{. Note that the necessary ingredients for the flow t in Eq.(13)-(14) are in this case simply the correlations at equal or different times; 1Jt is computed according to Eq.(15). In Fig.2 we observe the results of the simulation (for parameter details, see figure caption). After a short time (t=O.4s) of large 1J and strong fluctuations in 1J the mixing matrix is estimated correctly. Until t=1.25s the learning rate adapts cooling down approximately similar to lit (cf. Fig. 2c), which was predicted in Eq.(12) in the previous section, i.e. it finds the optimal rate for annealing. At the 604 N. Murata, K. Muller; A. Ziehe and S. Amari point of the switch where simple annealed learning would have failed to adapt to the sudden change, our adaptive rule increases TJ drastically and is able to follow the switch within another O.4s rsp. O.ls. Then again, the learning rate is cooled down automatically as intended. Comparing the mixed, original and unmixed signals in Fig.1 confirms the accurate and fast estimate that we already observed in the mixing matrix elements. The same also holds for an acoustic cross check: for a small part of a second both signals are audible, then as time proceeds only one signal, and again after the switches both signals are audible but only for a very short moment. The fading away of the signal is so fast to the listener that it seems that one signal is simply "switched off" by the separation algorithm. Altogether we found an excellent adaptation behavior of the proposed on-line algorithm, which was also reproduced in other simulation examples omitted here. 6 Conclusion We gave a theoretically motivated adaptive on-line algorithm extending the work of Sompolinskyet al. [1995]. Our algorithm applies to general feed-forward networks and can be used to accelerate learning by the learning about learning strategy in the difficult setting where (a) continuous functions or distributions are to be learned, (b) the Hessian K is not available and (c) no explicit loss function is given. Note, that if an explicit loss function or K is given, this additional information can be incorporated easily, e.g. we can make use of the real gradient otherwise we only rely on the flow. Non-stationary blind separation is a typical implementation of the setting (a)-(c) and we use it as an application of the adaptive on-line algorithm in a changing environment. Note that we can apply the learning rate adaptation to most existing blind separation algorithms and thus make them feasible for a nonstationary environment. However, we would like to emphasize that blind separation is just an example for the general adaptive on-line strategy proposed and applications of our algorithm are by no means limited to this scenario. Future work will also consider applications where the rules change more gradually (e.g. drift) . References Amari, S. (1967) IEEE Trans. EC 16(3):299-307. Bell, T ., Sejnowski, T . (1995) Neural Compo 7:1129-1159. Cichocki A., Amari S., Adachi M., Kasprzak W. (1996) Self-Adaptive Neural Networks for Blind Separation of Sources, ISCAS'96 (IEEE), Vol. 2, 157-160. Darken, C., Moody, J. (1991) in NIPS 3, Morgan Kaufmann, Palo Alto. Heskes, T.M., Kappen, B. (1991) Phys. Rev. A 440:2718-2726. Jutten, C., Herault, J . (1991) Signal Processing 24:1-10. Molgedey, L., Schuster, H.G. (1994) Phys. Rev. Lett. 72(23):3634-3637. Robbins, H., Monro, S. (1951) Ann. Math. Statist., 22:400-407. Rumelhart, D., McClelland, J.L and the PDP Research Group (eds.) (1986) , PDP Vol. 1, pp. 318-362, Cambridge, MA: MIT Press. Saad D., and SoIl a S. (1995), Workshop at NIPS '95, see World-Wide-Web page: http://neural-server.aston.ac.uk/nips95/workshop.html and references therein. Sompolinsky, H., Barkai, N., Seung, H.S. (1995) in Neural Networks: The Statistical Mechanics Perspective, pp. 105-130. Singapore: World Scientific. Sutton, R.S. (1992) in Proc. 10th nat. conf. on AI, 171-176, MIT Press. Adaptive On-line Learning in Changing Environments t:~I.~ I ~.~ •• ~ ~.I~.~I ~ I ~~ ... ~ ~ ~~.I~ o 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 t~~ o 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 m 1~--~--~----~--~--~----~--~--~--------~ c: Ol 'iii i 0 'E §_1L-__ ~ __ ~ ____ L-__ -L __ ~ ____ ~ __ -L __ ~ ____ ~ __ ~ o 0.5 1 1.5 5 r~·,,: ' .. :, , til 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 time sec 605 Figure 1: s¥ "space music", the mixture signal Ii, the unmixed signal u¥ and the separation error u¥ - s¥ as functions of time in seconds. ¥l 1 c: III ~0.8 III 8 0l0.6 c: .~ 'E 0.4 0 0.5 1.5 2 2.5 3 3.5 4 4.5 5 15 10 .l!l III 5 0.5 1.5 2 2.5 3 3.5 4 4.5 5 10 ~ 5 ~ 00 0.5 1.5 2 2.5 3 3.5 4 4.5 5 time sec Figure 2: Estimated mixing matrix Tt , evolution ofthe learning rate TJt and inverse learning rate l/TJt over time. Rule switches (t=1.25s, 3.75s) are clearly observed as drastic changes in TJt . Asymptotic 1ft scaling in TJ amounts to a straight line in l/TJt. Simulation parameters are a = 0.002,,B = 20/ maxll(r)II, f = 8 = 0.01. maxll(r)11 denotes the maximal value of the past observations.
1996
6
1,258
Clustering Sequences with Hidden Markov Models Padhraic Smyth Information and Computer Science University of California, Irvine CA 92697-3425 smyth~ics.uci.edu Abstract This paper discusses a probabilistic model-based approach to clustering sequences, using hidden Markov models (HMMs). The problem can be framed as a generalization of the standard mixture model approach to clustering in feature space. Two primary issues are addressed. First, a novel parameter initialization procedure is proposed, and second, the more difficult problem of determining the number of clusters K, from the data, is investigated. Experimental results indicate that the proposed techniques are useful for revealing hidden cluster structure in data sets of sequences. 1 Introduction Consider a data set D consisting of N sequences, D = {SI,"" SN}' Si = (.f.L ... .f.~J is a sequence of length Li composed of potentially multivariate feature vectors.f.. The problem addressed in this paper is the discovery from data of a natural grouping of the sequences into K clusters. This is analagous to clustering in multivariate feature space which is normally handled by methods such as k-means and Gaussian mixtures. Here, however, one is trying to cluster the sequences S rather than the feature vectors.f.. As an example Figure 1 shows four sequences which were generated by two different models (hidden Markov models in this case). The first and third came from a model with "slower" dynamics than the second and fourth (details will be provided later). The sequence clustering problem consists of being given sample sequences such as those in Figure 1 and inferring from the data what the underlying clusters are. This is non-trivial since the sequences can be of different lengths and it is not clear what a meaningful distance metric is for sequence comparIson. The use of hidden Markov models for clustering sequences appears to have first Clustering Sequences with Hidden Markov Models 649 Figure 1: Which sequences came from which hidden Markov model? been mentioned in Juang and Rabiner (1985) and subsequently used in the context of discovering subfamilies of protein sequences in Krogh et al. (1994). This present paper contains two new contributions in this context: a cluster-based method for initializing the model parameters and a novel method based on cross-validated likelihood for determining automatically how many clusters to fit to the data. A natural probabilistic model for this problem is that of a finite mixture model: K fK(S) = L/j(SIOj)pj (1) j=l where S denotes a sequence, Pj is the weight of the jth model, and /j (SIOj) is the density function for the sequence data S given the component model /j with parameters OJ . Here we will assume that the /j's are HMMs: thus, the OJ'S are the transition matrices, observation density parameters, and initial state probabilities, all for the jth component. /j (SIOj) can be computed via the forward part of the forward backward procedure. More generally, the component models could be any probabilistic model for S such as linear autoregressive models, graphical models, non-linear networks with probabilistic semantics, and so forth. It is important to note that the motivation for this problem comes from the goal of building a descriptive model for the data, rather than prediction per se. For the prediction problem there is a clearly defined metric for performance, namely average prediction error on out-of-sample data (cf. Rabiner et al. (1989) in a speech context with clusters of HMMs and Zeevi, Meir, and Adler (1997) in a general time-series context). In contrast, for descriptive modeling it is not always clear what the appropriate metric for evaluation is, particularly when K, the number of clusters, is unknown. In this paper a density estimation viewpoint is taken and the likelihood of out-of-sample data is used as the measure of the quality of a particular model. 2 An Algorithm for Clustering Sequences into ]{ Clusters Assume first that K, the number of clusters, is known. Our model is that of a mixture of HMMs as in Equation 1. We can immediately observe that this mixture can itself be viewed as a single "composite" HMM where the transition matrix A of the model is block-diagonal, e.g., if the mixture model consists of two components with transition matrices Al and A2 we can represent the overall mixture model as 650 P Smyth a single HMM (in effect, a hierarchical mixture) with transition matrix (2) where the initial state probabilities are chosen appropriately to reflect the relative weights ofthe mixture components (the Pic in Equation 1). Intuitively, a sequence is generated from this model by initially randomly choosing either the "upper" matrix Al (with probability PI) or the "lower" matrix with probability A2 (with probability 1 - PI) and then generating data according to the appropriate Ai . There is no "crossover" in this mixture model: data are assumed to come from one component or the other. Given this composite HMM a natural approach is to try to learn the parameters of the model using standard HMM estimation techniques, i.e., some form of initialization followed by Baum-Welch to maximize the likelihood. Note that unlike predictive modelling (where likelihood is not necessarily an appropriate metric to evaluate model quality), likelihood maximization is exactly what we want to do here since we seek a generative (descriptive) model for the data. We will assume throughout that the number of states per component is known a priori, i.e., that we are looking for K HMM components each of which has m states and m is known. An obvious extension is to address the problem of learning K and m simultaneously but this is not dealt with here. 2.1 Initialization using Clustering in "Log-Likelihood Space" Since the EM algorithm is effectively hill-climbing the likelihood surface, the quality of the final solution can depend critically on the initial conditions. Thus, using as much prior information as possible about the problem to seed the initialization is potentially worthwhile. This motivates the following scheme for initializing the A matrix of the composite HMM: 1. Fit N m-state HMMs, one to each individual sequence Si, 1 ~ i ~ N. These HMMs can be initialized in a "default" manner: set the transition matrices uniformly and set the means and covariances using the k-means algorithm, where here k = m, not to be confused with K, the number of HMM components. For discrete observation alphabets modify accordingly. 2. For each fitted model Mi, evaluate the log-likelihood of each of the N sequences given model Mi, i.e., calculate Lij = log L(Sj IMi), 1 ~ i, j ~ N. 3. Use the log-likelihood distance matrix to cluster the sequences into K groups (details of the clustering are discussed below). 4. Having pooled the sequences into K groups, fit K HMMs, one to each group, using the default initialization described above. From the K HMMs we get K sets of parameters: initialize the composite HMM in the obvious way, i.e., the m x m "block-diagonal" component Aj of A (where A is mK x mK) is set to the estimated transition matrix from the jth group and the means and covariances of the jth set of states are set accordingly. Initialize the Pj in Equation 1 to Nj / N where Nj is the number of sequences which belong to cluster j. After this initialization step is complete, learning proceeds directly on the composite HMM (with matrix A) in the usual Baum-Welch fashion using all of the sequences. The intuition behind this initialization procedure is as follows. The hypothesis is that the data are being generated by K models. Thus, if we fit models to each individual sequence, we will get noisier estimates of the model parameters (than if we used all of the sequences from that cluster) but the parameters should be Clustering Sequences with Hidden Markov Models 651 clustered in some manner into K groups about their true values (assuming the model is correct). Clustering directly in parameter space would be inappropriate (how does one define distance?): however, the log-likelihoods are a natural way to define pairwise distances. Note that step 1 above requires the training of N sequences individually and step 2 requires the evaluation of N 2 distances. For large N this may be impractical. Suitable modifications which train only on a small random sample of the N sequences and randomly sample the distance matrix could help reduce the computational burden, but this is not pursued here. A variety of possible clustering methods can be used in step 3 above. The "symmetrized distance" Lij = Ij2(L(Si IMj)+ L(Sj IMi» can be shown to be an appropriate measure of dissimilarity between models Mi and Mj (Juang and Rabiner 1985). For the results described in this paper, hierarchical clustering was used to generate K clusters from the symmetrized distance matrix. The "furthest-neighbor" merging heuristic was used to encourage compact clusters and worked well empirically, although there is no particular reason to use only this method. We will refer to the above clustering-based initialization followed by Baum-Welch training on the composite model as the "HMM-Clustering" algorithm in the rest of the paper. 2.2 Experimental Results Consider a deceptively simple "toy" problem. I-dimensional feature data are generated from a 2-component HMM mixture (K = 2), each with 2 states. We have A - (0.6 0.4) A _ (0.4 0.6) 1 0.4 0.6 2 0.6 0.4 and the observable feature data obey a Gaussian density in each state with 0'1 = 0'2 = 1 for each state in each component, and PI = 0, P2 = 3 for the respective mean of each state of each component. 4 sample sequences are shown in Figure 1. The top, and third from top, sequences are from the "slower" component Al (is more likely to stay in any state than switch). In total the training data contain 20 sample sequences from each component of length 200. The problem is non-trivial both because the data have exactly the same marginal statistics if the temporal sequence information is removed and because the Markov dynamics (as governed by Al and A2 ) are relatively similar for each component making identification somewhat difficult. The HMM clustering algorithm was applied to these sequences. The symmetrized likelihood distance matrix is shown as a grey-scale image in Figure 2. The axes have been ordered so that the sequences from the same clusters are adjacent. The difference in distances between the two clusters is apparent and the hierarchical clustering algorithm (with K = 2) easily separates the two groups. This initial clustering, followed by training separately the two clusters on the set of sequences assigned to each cluster, yielded: A - (0.580 0.402) 1 0.420 0.598 A (0.392 0.611) A2 = 0.608 0.389 A (2.892 PI = 0.040 A (2.911 P2 = 0.138 ) ) A (1.353 0'1 = 1.219 A (1.239 0'2 = 1.339 ) ) Subsequent training of the composite model on all of the sequences produced more refined parameter estimates, although the basic cluster structure of the model remained the same (i.e., the initial clustering was robust) . 652 P. Smyth modeLo." to 15 20 26 30 3S 40 mod,LO." modelO.e Figure 2: Symmetrized log-likelihood distance matrix. For comparative purposes two alternative initialization procedures were used to initialize the training ofthe composite HMM. The "unstructured" method uniformly initializes the A matrix without any knowledge of the fact that the off-block-diagonal terms are zero (this is the "standard" way of fitting a HMM). The "block-uniform" method uniformly initializes the K block-diagonal matrices within A and sets the off-block-diagonal terms to zero. Random initialization gave poorer results overall compared to uniform. Table 1: Differences in log-likelihood for different initialization methods. Initialization Maximum Mean Standard Method Log-Likelihood Log-Likelihood Deviation Value Value of Log-Likelihoods Unstructured 7.6 0.0 1.3 Block-Uniform 44.8 8.1 28.7 HMM-Clustering 55.1 50.4 0.9 The three alternatives were run 20 times on the data above, where for each run the seeds for the k-means component of the initialization were changed. The maximum, mean and standard deviations of log-likelihoods on test data are reported in Table 1 (the log-likelihoods were offset so that the mean unstructured log-likelihood is zero). The unstructured approach is substantially inferior to the others on this problem: this is not surprising since it is not given the block-diagonal structure of the true model. The Block-Uniform method is closer in performance to HMMClustering but is still inferior. In particular, its log-likelihood is consistently lower than that of the HMM-Clustering solution and has much greater variability across different initial seeds. The same qualitative behavior was observed across a variety of simulated data sets (results are not presented here due to lack of space). 3 Learning K, the Number of Sequence Components 3.1 Background Above we have assumed that K, the number of clusters, is known. The problem of learning the "best" value for K in a mixture model is a difficult one in practice Clustering Sequences with Hidden Markov Models 653 even for the simpler (non-dynamic) case of Gaussian mixtures. There has been considerable prior work on this problem. Penalized likelihood approaches are popular, where the log-likelihood on the training data is penalized by the subtraction of a complexity term. A more general approach is the full Bayesian solution where the posterior probability of each value of K is calculated given the data, priors on the mixture parameters, and priors on K itself. A potential difficulty here is the the computational complexity of integrating over the parameter space to get the posterior probabilities on K. Various analytic and sampling approximations are used in practice. In theory, the full Bayesian approach is fully optimal and probably the most useful. However, in practice the ideal Bayesian solution must be approximated and it is not always obvious how the approximation affects the quality of the final answer. Thus, there is room to explore alternative methods for determining K. 3.2 A Monte-Carlo Cross-Validation Approach Imagine that we had a large test data set Dtest which is not used in fitting any of the models. Let LK(Dtest) be the log-likelihood where the model with K components is fit to the training data D but the likelihood is evaluated on Dtest . We can view this likelihood as a function of the "parameter" K, keeping all other parameters and D fixed. Intuitively, this "test likelihood" should be a much more useful estimator than the training data likelihood for comparing mixture models with different numbers of components. In fact, the test likelihood can be shown to be an unbiased estimator of the Kullback-Leibler distance between the true (but unknown) density and the model. Thus, maximizing out-of-sample likelihood over K is a reasonable model selection strategy. In practice, one does not usually want to reserve a large fraction of one's data for test purposes: thus, a cross-validated estimate of log-likelihood can be used instead. In Smyth (1996) it was found that for standard multivariate Gaussian mixture modeling, the standard v-fold cross-validation techniques (with say v = 10) performed poorly in terms of selecting the correct model on simulated data. Instead MonteCarlo cross-validation (Shao, 1993) was found to be much more stable: the data are partitioned into a fraction f3 for testing and 1 - f3 for training, and this procedure is repeated M times where the partitions are randomly chosen on each run (i.e., need not be disjoint). In choosing f3 one must tradeoff the variability of the performance estimate on the test set with the variability in model fitting on the training set. In general, as the total amount of data increases relative to the model complexity, the optimal f3 becomes larger. For the mixture clustering problem f3 = 0.5 was found empirically to work well (Smyth, 1996) and is used in the results reported here. 3.3 Experimental Results The same data set as described earlier was used where now K is not known a priori. The 40 sequences were randomly partitioned 20 times into training and test crossvalidation sets. For each train/test partition the value of K was varied between 1 and 6, and for each value of K the HMM-Clustering algorithm was fit to the training data, and the likelihood was evaluated on the test data. The mean cross-validated likelihood was evaluated as the average over the 20 runs. Assuming the models are equally likely a priori, one can generate an approximate posterior distribution p(KID) by Bayes rule: these posterior probabilities are shown in Figure 3. The cross-validation procedure produces a clear peak at K = 2 which is the true model size. In general, the cross-validation method has been tested on a variety of other simulated sequence clustering data sets and typically converges as a function of the number oftraining samples to the true value of K (from below). As the number of 654 P. Smyth Eallmeted POIIertor PrcbabIlIlu on K lor 2-Clus1er Data 0.9 0.8 0.7 ~0.6 1 0.5 ·~0.4 ~ 0.3 0.2 0.1 °0 2 5 6 7 K Figure 3: Posterior probability distribution on K as estimated by cross-validation. data points grow, the posterior distribution on K narrows about the true value of K. If the data were not generated by the assumed form of the model, the posterior distribution on K will tend to be peaked about the model size which is closest (in K-L distance) to the true model. Results in the context of Gaussian mixture clustering(Smyth 1996) have shown that the Monte Carlo cross-validation technique performs as well as the better Bayesian approximation methods and is more robust then penalized likelihood methods such as BIC. In conclusion, we have shown that model-based probabilistic clustering can be generalized from feature-space clustering to sequence clustering. Log-likelihood between sequence models and sequences was found useful for detecting cluster structure and cross-validated likelihood was shown to be able to detect the true number of clusters. References Baldi, P. and Y. Chauvin, 'Hierarchical hybrid modeling, HMM/NN architectures, and protein applications,' Neural Computation, 8(6), 1541-1565,1996. Krogh, A. et al., 'Hidden Markov models in computational biology: applications to protein modeling,' it J. Mol. Bio., 235:1501-1531, 1994. Juang, B. H., and L. R. Rabiner, 'A probabilistic distance measure for hidden Markov models,' AT&T Technical Journal, vo1.64, no.2, February 1985. Rabiner, L. R., C. H. Lee, B. H. Juang, and J. G. Wilpon, 'HMM clustering for connected word recognition,' Pmc. Int. Conf. Ac. Speech. Sig. Proc, IEEE Press, 405-408, 1989. Shao, J ., 'Linear model selection by cross-validation,' J. Am. Stat. Assoc., 88(422), 486-494, 1993. Smyth, P., 'Clustering using Monte-Carlo cross validation,' in Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, Menlo Park, CA: AAAI Press, pp.126-133, 1996. Zeevi, A. J ., Meir, R., Adler, R., 'Time series prediction using mixtures of experts,' in this volume, 1997.
1996
60
1,259
Promoting Poor Features to Supervisors: Some Inputs Work Better as Outputs Rich Caruana JPRC and Carnegie Mellon University Pittsburgh, PA 15213 caruana@cs.cmu.edu Virginia R. de Sa Sloan Center for Theoretical Neurobiology and W. M. Keck Center for Integrative Neuroscience University of California, San Francisco CA 94143 desa@phy.ucsf.edu Abstract In supervised learning there is usually a clear distinction between inputs and outputs inputs are what you will measure, outputs are what you will predict from those measurements. This paper shows that the distinction between inputs and outputs is not this simple. Some features are more useful as extra outputs than as inputs. By using a feature as an output we get more than just the case values but can. learn a mapping from the other inputs to that feature. For many features this mapping may be more useful than the feature value itself. We present two regression problems and one classification problem where performance improves if features that could have been used as inputs are used as extra outputs instead. This result is surprising since a feature used as an output is not used during testing. 1 Introduction The goal in supervised learning is to learn functions that map inputs to outputs with high predictive accuracy. The standard practice in neural nets is to use all features that will be available for the test cases as inputs, and use as outputs only the features to be predicted. Extra features available for training cases that won't be available during testing can be used as extra outputs that often benefit the original output[2][5]. Other ways of adding information to supervised learning through outputs include hints[l]' tangent-prop[7], and EBNN[8]. In unsupervised learning it has been shown that inputs arising from different modalities can provide supervisory signals (outputs for the other modality) to each other and thus aid learning [3][6]. If outputs are so useful, and since any input could be used as an output, would some inputs be more useful as outputs? Yes. In this paper we show that in supervised backpropagation learning, some features are more useful as outputs than as inputs. This is surprising since using a feature as an output only extracts information from it during training; during testing it is not used. 390 R. Caruana and V. R. de Sa This paper uses the following terms: The Main Task is the output to be learned. The goal is to improve performance on the Main Task. Regular Inputs are the features provided as inputs in all experiments. Extra Inputs (Extra Outputs) are the extra features when used as inputs (outputs). STD is standard backpropagation using the Regular Inputs as inputs and the Main Task as outputs. STD+IN uses the Extra Features as Extra Inputs to learn the Main Task. STD+OUT uses the Extra Features, but as Extra Outputs learned in parallel with the Main Task, using just the Regular Inputs as inputs. 2 Poorly Correlated Features This section presents a simple synthetic problem where it is easy to see why using a feature as an extra output is better than using that same feature as an extra input. Consider the following function: F1(A,B) = SIGMOID(A+B), SIGMOID(x) = 1/(1 + e(-x) The STD net in Figure 1a has 20 inputs, 16 hidden units, and one output. We use backpropagation on this net to learn FlO. A and B are uniformly sampled from the interval [-5,5]. The network's input is binary codes for A and B. The range [-5,5] is discretized into 210 bins and the binary code of the resulting bin number is used as the input coding. The first 10 input units receive the code for A and the second 10 that for B. The target output is the unary real (unencoded) value F1(A,B). A:B'I'D fully connected hidden layer !lain OUtput , ~ .:81'0+11 IIIlin OUtput , folly cOMocted ~ hidden layer ~~oo~ "'''''''' """"I' binary inputs coding for A binary inputs coding for B a • g u I a r I n put • "11'"'" ""'''"' I binary inputs coding for A binary inputs codinq for B aegular Input. Jlxtra IDput C I BTDtOUT !laiD OUtput Bxtra OUtput fully connected hidden layer , , ""'''''' '''''''''' binary inputs codino for A binary inputs coding for B Ragular Input. Figure 1: Three Neural Net Architectures for Learning F1 Backpropagation is done with per-epoch updating and early stopping. Each trial uses new random training, halt, and test sets. Training sets contain 50 patterns. This is enough data to get good performance, but not so much that there is not room for improvement. We use large halt and test sets 1000 cases each to minimize the effect of sampling error in the measured performances. Larger halt and test sets yield similar results. We use this methodology for all the experiments in this paper. Table 1 shows the mean performance of 50 trials of STD Net 1a with backpropagation and early stopping. N ow consider a similar function: J:2(A,B) = SIGMOID(A-B). Suppose, in addition to the 10-bit codings for A and B, you are given the unencoded unary value F2(A,B) as an extra input feature. Will this extra input help you learn F1(A,B) better? Probably not. A+B and A-B do not correlate for random A and B. The correlation coefficient for our training sets is typically less than ±0.0l. Because Promoting Poor Features to Supervisors Table 1: Mean Test Set Root-Mean-Squared-Error on F1 Network I Trials I Mean RMSE I Significance I STD 50 0.0648 STD+IN 50 0.0647 ns STD+OUT 50 0.0631 0.013* 391 of this, knowing the value of F2(A,B) does not tell you much about the target value F1(A,B) (and vice-versa). F1(A,B)'s poor correlation with F2(A,B) hurts backprop's ability to learn to use F2(A,B) to predict F1(A,B). The STD+IN net in Figure 1b has 21 inputs 20 for the binary codes for A and B, and an extra input for F2(A,B). The 2nd line in Table 1 shows the performance of STD+IN for the same training, halting, and test sets used by STD; the only difference is that there is an extra input feature in the data sets for STD+ IN. Note that the performance of STD+ IN is not significantly different from that ofSTD the extra information contained in the feature F2(A,B) does not help backpropagation learn F1(A,B) when used as an extra input. If F2(A,B) does not help backpropagation learn Fl(A,B) when used as an input, should we ignore it altogether? No. F1(A,B) and F2(A,B) are strongly related. They both benefit from decoding the binary input encoding to compute the subfeatures A and B. If, instead of using F2(A,B) as an extra input, it is used as an extra output trained with backpropagation, it will bias the shared hidden layer to learn A and B better, and this will help the net better learn to predict Fl(A,B). Figure 1c shows a net with 20 inputs for A and B, and 2 outputs, one for Fl(A,B) and one for F2(A,B). Error is back-propagated from both outputs, but the performance of this net is evaluated only on the output F1(A,B) and early stopping is done using only the performance of this output. The 3rd line in Table 1 shows the mean performance of 50 trials of this multitask net on F1(A,B). Using F2(A,B) as an extra output significantly improves performance on F1(A,B). Using the extra feature as an extra output is better than using it as an extra input. By using F2( A,B) as an output we make use of more than just the individual output values F2( A , B) but learn to extract information about the function mapping the inputs to F2(A,B). This is a key difference between using features as inputs and outputs. The increased performance of STD+OUT over STD and STD+IN is not due to STD+OUT reducing the capacity available for the main task FlO. All three nets STD, STD+IN, STD+OUT perform better with more hidden units. (Because larger capacity favors STD+OUT over STD and STD+IN, we report results for the moderate sized 16 hidden unit nets to be fair to STD and STD+IN.) 3 Noisy Features This section presents two problems where extra features are more useful as inputs if they have low noise, but which become more useful as outputs as their noise increases. Because the extra features are ideal features for these problems, this demonstrates that what we observed in the previous section does not depend on the extra features being contrived so that their correlation with the main task is low - features with high correlation can still be more useful as outputs. Once again, consider the main task from the previous section: F1(A,B) = SIGMOID(A+B) 392 Now consider these extra features: EF(A) = A + NOISE...sCALE * Noisel EF(B) = B + NOISE...sCALE * Noise2 R. Caruana and V. R. de Sa Noisel and Noise2 are uniformly sampled on [-1,1]. If NOISE...sCALE is not too large, EF(A) and EF(B) are excellent input features for learning FI(A,B) because the net can avoid learning to decode the binary input representations. However, as NOISE_SCALE increases, EF(A) and EF(B) become less useful and it is better for the net to learn FI(A,B) from the binary inputs for A and B. As before, we try using the extra features as either extra inputs or as extra outputs. Again, the training sets have 50 patterns, and the halt and test sets have 1000 patterns. Unlike before, however, we ran preliminary tests to find the best net size. The results showed 256 hidden units to be about optimal for the STD nets with early stopping on this problem. 0.06 0.18 0.17 w 0.055 (/) w ~ 0.16 :::i: ex: 1ii (/) ! 0.05 0.045 'STD+IN" "STD+OUr -+-"STD" · B·· 0.0 1.0 2.0 3.0 4.0 5.0 6.0 7.0 8.0 9.0 10.0 Feature Noise Scale ex: 0.15 0.14 0.13 "STD+IN" ' STD+OUT' -+-'STD" ·B-0.12 "---'---'-_'----'----'-_L..--'-----'---JL..-~ 0.0 1.0 2.0 3.0 4.0 5.0 6.0 7.0 8.0 9.0 10.0 Feature Noise Scale Figure 2: STD, STD+IN, and STD+OUT on FI (left) and F3 (right) Figure 2a plots the average performance of 50 trials of STD+IN and STD+OUT as NOISE...sCALE varies from 0.0 to 10.0. The performance of STD, which does not use EF(A) and EF(B), is shown as a horizontal line; it is independent of NOISE_SCALE. Let's first examine the results of STD+IN which uses EF(A) and EF(B) as extra inputs. As expected, when the noise is small, using EF(A) and EF(B) as extra inputs improves performance considerably. As the noise increases, however, this improvement decreases. Eventually there is so much noise in EF(A) and EF(B) that they no longer help the net if used as inputs. And, if the noise increases further, using EF(A) and EF(B) as extra inputs actually hurts. Finally, as the noise gets very large, performance asymptotes back towards the baseline. Using EF(A) and EF(B) as extra outputs yields quite different results. When the noise is low, they do not help as much as they did as extra inputs. As the noise increases, however, at some point they help more as extra outputs than as extra inputs, and never hurt performance the way the noisy extra inputs did. Why does noise cause STD+IN to perform worse than STD? With a finite training sample, correlations between noisy inputs and the main task cause the network to use the noisy inputs. To the extent that the main task is a function of the noisy inputs, it must pass the noise to the output, causing the output to be noisy. Also, as the net comes to depend on the noisy inputs, it depends less on the noise-free binary inputs. The noisy inputs explain away some of the training signal, so less is available to encourage learning to decode the binary inputs. Promoting Poor Features to Supervisors 393 Why does noise not hurt STD+OUT as much as it hurts STD+IN? As outputs, the net is learning the mapping from the regular inputs to EF(A) and EF(B). Early in training, the net learns to interpolate through the noise and thus learns smooth functions for EF(A) and EF(B) that have reasonable fidelity to the true mapping. This makes learning less sensitive to the noise added to these features. 3.1 Another Problem F1(A,B) is only mildly nonlinear because A and B do not go far into the tails of the SIGMOID. Do the results depend on this smoothness? To check, we modified F1(A,B) to make it more nonlinear. Consider this function: F3(A,B) = SIGMOID(EXPAND(SIGMOID(A)-SIGMOID(B))) where EXPAND scales the inputs from (SIGMOID(A)-SIGMOID(B)) to the range [-12.5,12.5]' and A and B are drawn from [-12.5,12.5]. F3(A,B) is significantly more nonlinear than F1(A,B) because the expanded scales of A and B, and expanding the difference to [-12.5,12.5] before passing it through another sigmoid, cause much of the data to fall in the tails of either the inner or outer sigmoids. Consider these extra features: EF(A) = SIGMOID(A) + NOISE...sCALE * Noise1 EF(B) = SIGMOID(B) + NOISE...sCALE * Noise2 where Noises are sampled as before. Figure 2B shows the results of using extra features EF(A) and EF(B) as extra inputs or as extra outputs. The trend is similar to that in Figure 2A but the benefit of STD+OUT is even larger at low noise. The data for 2a and 2b are generated using different seeds, 2a used steepest descent and Mitre's Aspirin simulator, 2b used conjugate gradient and Toronto's Xerion simulator, and F1 and F3 do not behave as similarly as their definitions might suggest. The similarity between the two graphs is due to the ubiquity of the phenomena, not to some small detail of the test functions or how the experiments were run. 4 A Classification Problem This section presents a problem that combines feature correlation (Section 1) and feature noise (Section 2) into one problem. Consider the 1-D classification problem, shown in Figure 3, of separating two Gaussian distributions with means 0 and 1, and standard deviations of 1. This problem is simple to learn if the 1-D input is coded as a single, continuous input but can be made harder by embedding it non-linearly in a higher dimensional space. Consider encoding input values defined on [0.0,15.0] using an interpolated 4-D Gray code(GC); integer values are mapped to a 4-D binary Gray code and intervening non-integers are mapped linearly to intervening 4-D vectors between the binary Gray codes for the bounding integers. As the Gray code flips only one bit between neighboring integers this involves simply interpolating along the 1 dimension in the 4-D unit cube that changes. Thus 3.4 is encoded as .4(GC(4) - GC(3)) + GC(3). The extra feature is a 1-D value correlated (with correlation p) with the original unencoded regular input, X. The extra feature is drawn from a Gaussian distribution with mean p x (X - .5) + .5 and standard deviation )(1 - p2). Examples of the distributions of the unencoded original dimension and the extra feature for various correlations are shown in Figure 3. This problem has been carefully constructed so that the optimal classification boundary does not change as p varies. Consider the extreme cases. At p = 1, the extra feature is exactly an unencoded 394 R. Caruana and V. R. de Sa p=O p=O.5 p=1 y y y :\, ; \, ./, I I x , X X X o 1 Figure 3: Two Overlapped Gaussian Classes (left), and An Extra Feature (y-axis) Correlated Different Amounts (p = 0: no correlation, p = 1: perfect correlation) With the unencoded version of the Regular Input (x-axis) version of the regular input. A STD+IN net using this feature as an extra input could ignore the encoded inputs and solve the problem using this feature alone. An STD+OUT net using this extra feature as an extra output would have its hidden layer biased towards representations that decode the Gray code, which is useful to the main classification task. At the other extreme (p = 0), we expect nets using the extra feature to learn no better than one using just the regular inputs because there is no useful information provided by the uncorrelated extra feature. The interesting case is between the two extremes. We can imagine a situation where as an output, the extra feature is still able to help STD+OUT by guiding it to decode the Gray code but does not help STD+IN because of the high level of noise. 0,65 ... --------.-. ------ ----------- -- -----.-1:::1 0,64 0 ,63 +-- - ------------ ... -+- - ... ... -- ............ --------+--0,62 ---. 0,61 "STD+IN" _____ 0,6 "STD.OUT" -+--"STD· -0 -0,00 0 ,25 0,50 0 ,75 Correlation of Extra Feature 1,00 Figure 4: STD, STD+IN, and STD+OUT vs. p on the Classification Problem The class output unit uses a sigmoid transfer function and cross-entropy error measure. The output unit for the correlated extra feature uses a linear transfer function and squared error measure. Figure 4 shows the average performance of 50 trials of STD, STD+IN, and STD+OUT as a function of p using networks with 20 hidden units, 70 training patterns, and halt and test sets of 1000 patterns each. As in the previous section, STD+IN is much more sensitive to changes in the extra feature than STD+OUT, so that by p = 0.75 the curves cross and for p less than 0.75, the dimension is actually more useful as an output dimension than an extra input. 5 Discussion Are the benefits of using some features as extra outputs instead of as inputs large enough to be interesting? Yes. Using only 1 or 2 features as extra outputs instead of as inputs reduced error 2.5% on the problem in Section 1, more than 5% in regions of the graphs in Section 2, and more than 2.5% in regions of the graph in Section Promoting Poor Features to Supervisors 395 3. In domains where many features might be moved, the net effect may be larger. Are some features more useful as outputs than as inputs only on contrived problems? No. In this paper we used the simplest problems we could devise where a few features worked better as outputs than as inputs. But our findings explain a result we noted previously, but did not understand, when applying multitask learning to pneumonia risk prediction[4]. There, we had the choice of using lab tests that would be unavailable on future patients as extra outputs, or using poor i.e., noisy predictions of them as extra inputs. Using the lab tests as extra outputs worked better. If one compares the zero noise points for STD+OUT (there's no noise in a feature when used as an output because we use the values in the training set, not predicted values) with the high noise points for STD+IN in the graphs in Section 2, it is easy to see why STD+OUT could perform much better. This paper shows that the benefit of using a feature as an extra output is different from the benefit of using that feature as an input. As an input, the net has access to the values on the training and test cases to use for prediction. As an output, however, the net is instead biased to learn a mapping from the other inputs in the training set to that output. From the graphs it is clear that some features help when used either as an input, or as an output. Given that the benefit of using a feature as an extra output is different from that of using it as an input, can we get both benefits? Our early results with techniques that reap both benefits by allowing some features to be used simultaneously as both inputs and outputs while preventing learning direct feed through identity mappings are promising. Acknowledgements R. Caruana was supported in part by ARPA grant F33615-93-1-1330, NSF grant BES-9315428, and Agency for Health Care Policy and Research grant HS06468. v. de Sa was supported by postdoctoral fellowships from NSERC (Canada) and the Sloan Foundation. We thank Mitre Group for the Aspirin/Migraines Simulator and The University of Toronto for the Xerion Simulator. References [1] Y.S. Abu-Mostafa, "Learning From Hints in Neural Networks," Journal of Complexity 6:2, pp. 192-198, 1989. [2] S. B aluj a, and D.A. Pomerleau, "Using the Representation in a Neural Network's Hidden Layer for Task-Specific Focus of Attention". In C. Mellish (ed.) The International Joint Conference on Artificial Intelligence 1995 (IJCAI-95): Montreal, Canada. IJCAII & Morgan Kaufmann. San Mateo, CA. pp 133-139, 1995. [3] S. Becker and G. E. Hinton, "A self-organizing neural network that discovers surfaces in random-dot stereograms," Nature 355 pp. 161-163, 1992. [4] R. Caruana, S. Baluja, and T. Mitchell, "Using the Future to Sort Out the Present: Rankprop and Multitask Learning for Pneumnia Risk Prediction," Advances in Neural Information Processing Systems 8, 1996. [5] R. Caruana, "Learning Many Related Tasks at the Same Time With Backpropagation," Advances in Neural Information Processing Systems 7, 1995. [6] V. R. de Sa, "Learning classification with unlabeled data," Advances in Neural Information Processing Systems 6, pp. 112-119, Morgan Kaufmann, 1994. [7] P. Simard, B. Victoni, Y. 1. Cun, and J. Denker, "Tangent prop a formalism for specifying selected invariances in an adaptive network," Advances in Neural Information Processing Systems 4, pp. 895-903, Morgan Kaufmann, 1992. [8] S. Thrun and T. Mitchell, "Learning One More Thing," CMU TR: CS-94-184, 1994.
1996
61
1,260
Hidden Markov decision trees Michael I. Jordan*, Zoubin Ghahramanit, and Lawrence K. Saul* {jordan.zoubin.lksaul}~psyche.mit.edu *Center for Biological and Computational Learning Massachusetts Institute of Technology Cambridge, MA USA 02139 t Department of Computer Science University of Toronto Toronto, ON Canada M5S 1A4 Abstract We study a time series model that can be viewed as a decision tree with Markov temporal structure. The model is intractable for exact calculations, thus we utilize variational approximations. We consider three different distributions for the approximation: one in which the Markov calculations are performed exactly and the layers of the decision tree are decoupled, one in which the decision tree calculations are performed exactly and the time steps of the Markov chain are decoupled, and one in which a Viterbi-like assumption is made to pick out a single most likely state sequence. We present simulation results for artificial data and the Bach chorales. 1 Introduction Decision trees are regression or classification models that are based on a nested decomposition of the input space. An input vector x is classified recursively by a set of "decisions" at the nonterminal nodes of a tree, resulting in the choice of a terminal node at which an output y is generated. A statistical approach to decision tree modeling was presented by Jordan and Jacobs (1994), where the decisions were treated as hidden multinomial random variables and a likelihood was computed by summing over these hidden variables. This approach, as well as earlier statistical analyses of decision trees, was restricted to independently, identically distributed data. The goal of the current paper is to remove this restriction; we describe a generalization of the decision tree statistical model which is appropriate for time senes. The basic idea is straightforward-we assume that each decision in the decision tree is dependent on the decision taken at that node at the previous time step. Thus we augment the decision tree model to include Markovian dynamics for the decisions. 502 M. I. Jordan, Z Ghahramani and L. K. Saul For simplicity we restrict ourselves to the case in which the decision variable at a given nonterminal is dependent only on the same decision variable at the same nonterminal at the previous time step. It is of interest, however, to consider more complex models in which inter-nonterminal pathways allow for the possibility of various kinds of synchronization. Why should the decision tree model provide a useful starting point for time series modeling? The key feature of decision trees is the nested decomposition. If we view each nonterminal node as a basis function, with support given by the subset of possible input vectors x that arrive at the node, then the support of each node is the union of the support associated with its children. This is reminiscent of wavelets, although without the strict condition of multiplicative scaling. Moreover, the regions associated with the decision tree are polygons, which would seem to provide a useful generalization of wavelet-like decompositions in the case of a highdimensional input space. The architecture that we describe in the current paper is fully probabilistic. We view the decisions in the decision tree as multinomial random variables, and we are concerned with calculating the posterior probabilities of the time sequence of hidden decisions given a time sequence of input and output vectors. Although such calculations are tractable for decision trees and for hidden Markov models separately, the calculation is intractable for our model. Thus we must make use of approximations. We utilize the partially factorized variational approximations described by Saul and Jordan (1996), which allow tractable substructures (e.g., the decision tree and Markov chain substructures) to be handled via exact methods, within an overall approximation that guarantees a lower bound on the log likelihood. 2 Architectures 2.1 Probabilistic decision trees The "hierarchical mixture of experts" (HME) model (Jordan & Jacobs, 1994) is a decision tree in which the decisions are modeled probabilistically, as are the outputs. The total probability of an output given an input is the sum over all paths in the tree from the input to the output. The HME model is shown in the graphical model formalism in Figure 2.1. Here a node represents a random variable, and the links represent probabilistic dependencies. A conditional probability distribution is associated with each node in the graph, where the conditioning variables are the node's parents. Let Zl, Z2, and z3 denote the (multinomial) random variables corresponding to the first, second and third levels of the decision tree. l We associate multinomial probabilities P(zl lx,1]l), P(z2Ix,zl,1]2), and P(Z3jx,zl,Z2,1]3) with the decision nodes, where 1]1,1]2, and 1]3 are parameters (e.g., Jordan and Jacobs utilized softmax transformations of linear functions of x for these probabilities). The leaf probabilities P(y lx, zl, Z2 , Z3, 0) are arbitrary conditional probability models; e.g., linear/Gaussian models for regression problems. The key calculation in the fitting of the HME model to data is the calculation of the posterior probabilities of the hidden decisions given the clamped values of x and y. This calculation is a recursion extending upward and downward in the tree, in which the posterior probability at a given nonterminal is the sum of posterior probabilities associated with its children. The recursion can be viewed as a special IThroughout the paper we restrict ourselves to three levels for simplicity of presentation. Hidden Markov Decision Trees Figure 1: The hierarchical mixture of experts as a graphical model. The E step of the learning algorithm for HME's involves calculating the posterior probabilities of the hidden (unshaded) variables given the observed (shaded) variables. 503 1t Figure 2: An HMM as a graphical model. The transition matrix appears on the horizontal links and the output probability distribution on the vertical links. The E step of the learning algorithm for HMM's involves calculating the posterior probabilities of the hidden (unshaded) variables given the observed (shaded) variables. case of generic algorithms for calculating posterior probabilities on directed graphs (see, e.g., Shachter, 1990). 2.2 Hidden Markov models In the graphical model formalism a hidden Markov model (HMM; Rabiner, 1989) is represented as a chain structure as shown in Figure 2.1. Each state node is a multinomial random variable Zt. The links between the state nodes are parameterized by the transition matrix a(ZtIZt-l), assumed homogeneous in time. The links between the state nodes Zt and output nodes Yt are parameterized by the output probability distribution b(YtIZt), which in the current paper we assume to be Gaussian with (tied) covariance matrix ~. As in the HME model, the key calculation in the fitting of the HMM to observed data is the calculation of the posterior probabilities of the hidden state nodes given the sequence of output vectors. This calculation-the E step of the Baum-Welch algorithm-is a recursion which proceeds forward or backward in the chain. 2.3 Hidden Markov decision trees We now marry the HME and the HMM to produce the hidden Markov decision tree (HMDT) shown in Figure 3. This architecture can be viewed in one of two ways: (a) as a time sequence of decision trees in which the decisions in a given decision tree depend probabilistically on the decisions in the decision tree at the preceding moment in time; (b) as an HMM in which the state variable at each moment in time is factorized (cf. Ghahramani & Jordan, 1996) and the factors are coupled vertically to form a decision tree structure. Let the state of the Markov process defining the HMDT be given by the values of hidden multinomial decisions z~, z¥, and zr, where the superscripts denote the level ofthe decision tree (the vertical dimension) and the sUbscripts denote the time (the horizontal dimension). Given our assumption that the state transition matrix has only intra-level Markovian dependencies, we obtain the following expression for the 504 M. I. Jordan, Z Ghahramani and L. K. Saul • • • Figure 3: The HMDT model is an HME decision tree (in the vertical dimension) with Markov time dependencies (in the horizontal dimension). HMDT probability model: P( {z}, z;, zn, {Yt}l{xt}) = 7r1 (ZUX1)7r2(zilx1' zt)7r3(zflx1 ' zL zi) T T II a1 (z: IXt, zL1)a2(z; IXt, Z;_1 ' z:)a3(z~lxt , ZL1' z:, z;) II b(Yt IXt, z}, z;, z~) t=2 t=l Summing this probability over the hidden values zI, z;, and z~ yields the HMDT likelihood. The HMDT is a 2-D lattice with inhomogeneous field terms (the observed data). It is well-known that such lattice structures are intractable for exact probabilistic calculations. Thus, although it is straightforward to write down the EM algorithm for the HMDT and to write recursions for the calculations of posterior probabilities in the E step, these calculations are likely to be too time-consuming for practical use (for T time steps, J{ values per node and M levels, the algorithm scales as O(J{M+1T)). Thus we turn to methods that allow us to approximate the posterior probabilities of interest. 3 Algori thms 3.1 Partially factorized variational approximations Completely factorized approximations to probability distributions on graphs can often be obtained variationally as mean field theories in physics (Parisi, 1988). For the HMDT in Figure 3, the completely factorized mean field approximation would delink all of the nodes, replacing the interactions with constant fields acting at each of the nodes. This approximation, although useful, neglects to take into account the existence of efficient algorithms for tractable substructures in the graph. Saul and Jordan (1996) proposed a refined mean field approximation to allow interactions associated with tractable substructures to be taken into account. The basic idea is to associate with the intractable distribution P a simplified distribution Q that retains certain of the terms in P and neglects others, replacing them with parameters Pi that we will refer to as "variational parameters." Graphically the method can be viewed as deleting arcs from the original graph until a forest of tractable substructures is obtained. Arcs that remain in the simplified graph Hidden Markov Decision Trees 505 correspond to terms that are retained in Q; arcs that are deleted correspond to variational parameters. To obtain the best possible approximation of P we minimize the Kullback-Liebler divergence K L( Q liP) with respect to the parameters J-li. The result is a coupled set of equations that are solved iteratively. These equations make reference to the values of expectations of nodes in the tractable substructures; thus the (efficient) algorithms that provide such expectations are run as subroutines. Based on the posterior expectations computed under Q, the parameters defining P are adjusted. The algorithm as a whole is guaranteed to increase a lower bound on the log likelihood. 3.2 A forest of chains The HMDT can be viewed as a coupled set of chains, with couplings induced directly via the decision tree structure and indirectly via the common coupling to the output vector. If these couplings are removed in the variational approximation, we obtain a Q distribution whose graph is a forest of chains. There are several ways to parameterize this graph; in the current paper we investigate a parameterization with time-varying transition matrices and time-varying fields. Thus the Q distribution is given by T 1 II-1( 111 )-2( 212 )-3( 31 3 ) ~ at Zt Zt-1 at Zt Zt_l at Zt Zt_1 Q t=2 T II iji(zDij;(z;)ij~(zn t=l where a~(z~ IzL1) and ij;(zD are potentials that provide the variational parameterization. 3.3 A forest of decision trees Alternatively we can drop the horizontal couplings in the HMDT and obtain a variational approximation in which the decision tree structure is handled exactly and the Markov structure is approximated. The Q distribution in this case is T II fi(zDf;(zi IzDf~(zrlzt, zi) t=l Note that a decision tree is a fully coupled graphical model; thus we can view the partially factorized approximation in this case as a completely factorized mean field approximation on "super-nodes" whose configurations include all possible configurations of the decision tree. 3.4 A Viterbi-like approximation In hidden Markov modeling it is often found that a particular sequence of states has significantly higher probability than any other sequence. In such cases the Viterbi algorithm, which calculates only the most probable path, provides a useful computational alternative. We can develop a Viterbi-like algorithm by utilizing an approximation Q that assigns probability one to a single path {zi, zF, zn: { I if z~ = zL "It, i o otherwise (1) 506 a) 10r-------------------, ~ 0 -5 -10~-------------' o 50 100 150 200 t M. I. Jordan, Z Ghahramani and L. K. SauL b) 600r-------------------, 500 "'0 8 400 ,£; ~300 ~200 100 2 J)) ff//JJjjJ 10 20 30 40 iterations of EM Figure 4: a) Artificial time series data. b) Learning curves for the HMDT. Note that the entropy Q In Q is zero, moreover the evaluation of the energy Q In P reduces to substituting z~ for z~ in P. Thus the variational approximation is particularly simple in this case. The resulting algorithm involves a subroutine in which a standard Viterbi algorithm is run on a single chain, with the other (fixed) chains providing field terms. 4 Results We illustrate the HMDT on (1) an artificial time series generated to exhibit spatial and temporal structure at multiple scales, and (2) a domain which is likely to exhibit such structure naturally-the melody lines from J .S. Bach's chorales. The artificial data was generated from a three level binary HMDT with no inputs, in which the root node determined coarse-scale shifts (±5) in the time series, the middle node determined medium-scale shifts (±2), and the bottom node determined fine-scale shifts (±0.5) (Figure 4a). The temporal scales at these three nodes-as measured by the rate of convergence (second eigenvalue) of the transition matrices, with 0 (1) signifying immediate (no) convergence-were 0.85,0.5, and 0.3, respectively. We implemented forest-of-chains, forest-of-trees and Viterbi-like approximations. The learning curves for ten runs of the forest-of-chains approximation are shown in Figure 4b. Three plateau regions are apparent, corresponding to having extracted the coarse, medium, and fine scale structures of the time series. Five runs captured all three spatia-temporal scales at their correct level in the hierarchy; three runs captured the scales but placed them at incorrect nodes in the decision tree; and two captured only the coarse-scale structure.2 Similar results were obtained with the Viterbi-like approximation. We found that the forest-of-trees approximation was not sufficiently accurate for these data. The Bach chorales dataset consists of 30 melody lines with 40 events each.3 Each discrete event encoded 6 attributes-start time of the event (st), pitch (pitch), duration (dur), key signature (key), time signature (time), and whether the event was under a fermata (ferm). The chorales dataset was modeled with 3-level HMDTs with branching factors (K) 2Note that it is possible to bias the ordering of the time scales by ordering the initial random values for the nodes of the tree; we did not utilize such a bias in this simulation. 3This dataset was obtained from the UCI Repository of Machine Learning Datasets. Hidden Markov Decision Trees 507 Percent variance explained Temporal scale K st pitch dur key time ferm level 1 level 2 level 3 2 3 6 6 84 95 a 1.00 1.00 0.51 3 22 38 7 93 99 0 1.00 0.96 0.85 4 55 48 36 96 99 5 1.00 1.00 0.69 5 57 41 41 97 99 61 1.00 0.95 0.75 6 70 40 58 94 99 10 1.00 0.93 0.76 Table 1: Hidden Markov decision tree models of the Bach chorales dataset: mean percentage of variance explained for each attribute and mean temporal scales at the different nodes. 2, 3, 4, 5, and 6 (3 runs at each size, summarized in Table 1). Thirteen out of 15 runs resulted in a coarse-to-fine progression of temporal scales from root to leaves of the tree. A typical run at branching factor 4, for example, dedicated the top level node to modeling the time and key signatures-attributes that are constant throughout any single chorale-the middle node to modeling start times, and the bottom node to modeling pitch or duration. 5 Conclusions Viewed in the context of the burgeoning literature on adaptive graphical probabilistic models-which includes HMM's, HME's, CVQ's, IOHMM's (Bengio & Frasconi, 1995), and factorial HMM's-the HMDT would appear to be a natural next step. The HMDT includes as special cases all of these architectures, moreover it arguably combines their best features: factorized state spaces, conditional densities, representation at multiple levels ofresolution and recursive estimation algorithms. Our work on the HMDT is in its early stages, but the earlier literature provides a reasonably secure foundation for its development. References Bengio, Y., & Frasconi, P. (1995). An input output HMM architecture. In G. Tesauro, D. S. Touretzky & T. K. Leen, (Eds.), Advances in Neural Information Processing Systems 7, MIT Press, Cambridge MA. Ghahramani, Z., & Jordan, M. r. (1996). Factorial hidden Markov models. In D. S. Touretzky, M. C. Mozer, & M. E. Hasselmo (Eds.), Advances in Neural Information Processing Systems 8, MIT Press, Cambridge MA. Jordan, M. r., & Jacobs, R. A. (1994). Hierarchical mixtures of experts and the EM algorithm. Neural Computation, 6, 181-214. Parisi, G. (1988). Statistical Field Theory. Redwood City, CA: Addison-Wesley. Rabiner, L. (1989). A tutorial on hidden Markov models and selected application s in speech recognition. Proceedings of the IEEE, 77, 257-285. Saul, L. K., & Jordan, M. I. (1996). Exploiting tractable substructures in intractable networks. In D. S. Touretzky, M. C. Mozer, & M. E. Hasselmo (Eds.), Advances in Neural Information Processing Systems 8, MIT Press, Cambridge MA. Shachter, R. (1990). An ordered examination of influence diagrams. Networks, 20, 535-563.
1996
62
1,261
Representing Face Images for Emotion Classification Curtis Padgett Department of Computer Science University of California, San Diego La Jolla, CA 92034 Garrison Cottrell Department of Computer Science University of California, San Diego La Jolla, CA 92034 Abstract We compare the generalization performance of three distinct representation schemes for facial emotions using a single classification strategy (neural network). The face images presented to the classifiers are represented as: full face projections of the dataset onto their eigenvectors (eigenfaces); a similar projection constrained to eye and mouth areas (eigenfeatures); and finally a projection of the eye and mouth areas onto the eigenvectors obtained from 32x32 random image patches from the dataset. The latter system achieves 86% generalization on novel face images (individuals the networks were not trained on) drawn from a database in which human subjects consistently identify a single emotion for the face. 1 Introduction Some of the most successful research in machine perception of complex natural image objects (like faces), has relied heavily on reduction strategies that encode an object as a set of values that span the principal component sub-space of the object's images [Cottrell and Metcalfe, 1991, Pentland et al., 1994]. This approach has gained wide acceptance for its success in classification, for the efficiency in which the eigenvectors can be calculated, and because the technique permits an implementation that is biologically plausible. The procedure followed in generating these face representations requires normalizing a large set of face views (" mug-shots") and from these, identifying a statistically relevant sub-space. Typically the sub-space is located by finding either the eigenvectors of the faces [Pentland et al., 1994] or the weights of the connections in a neural network [Cottrell and Metcalfe, 1991]. In this work, we classify face images based on their emotional content and examine how various representational strategies impact the generalization results of a classifier. Previous work using whole face representations for emotion classification by Representing Face Images for Emotion Classification 895 Cottrell and Metcalfe [Cottrell and Metcalfe, 1991] was less encouraging than results obtained for face recognition. We seek to determine if the problem in Cottrell and Metcalfe's work stems from bad data (i.e., the inability of the undergraduates to demonstrate emotion), or an inadequate representation (i.e. eigenfaces). Three distinct representations of faces are considered in this work- a whole face representation similar to that used in previous work on recognition, sex, and emotion [Cottrell and Metcalfe, 1991]; a more localized representation based on the eyes (eigeneyes and eigenmouths) and mouth [Pentland et aI., 1994]; and a representation of the eyes and mouth that makes use of basis vectors obtained by principal components of random image blocks. By examining the generalization rate of the classifiers for these different face representations, we attempt to ascertain the sensitivity of the representation and its potential for broader use in other vision classification problems. 2 Face Data The dataset used in Cottrell and Metcalfe's work on emotions consisted of the faces of undergraduates who were asked to pose for particular expressions. However, feigned emotions by untrained individuals exhibit significant differences from the prototypical face expression [Ekman and Friesen, 1977]. These differences often result in disagreement between the observed emotion and the expression the actor is attempting to feign. A feigned smile for instance, differs around the eyes when compared with a " natural" smile. The quality of the displayed emotion is one of the reasons cited by Cottrell and Metcalfe for the poor recognition rates achieved by their classifier. To reduce this possibility, we made use of a validated facial emotion database (Pictures of Facial Affect) assembled by Ekman and Friesen [Ekman and Friesen, 1976]. Each of the face images in this set exhibits a substantial agreement between the labeled emotion and the observed response of human subjects. The actors used in this database were trained to reliably produce emotions using Facial Action Coding System [Ekman and Friesen, 1977] and their images were presented to undergraduates for testing. The agreement between the emotion the actor was required to express and the students' observations was at least 70% on all the images incorporated in the database. We digitized a total of 97 images from 12 individuals (6 male, 6 female). Each portrays one of 7 emotions- happy, sad, fear, anger, surprise, disgust or neutral. With the exception of the neutral faces, each image in the set is labeled with a response vector of the remaining six emotions indicating the fraction of total respondents classifying the image with a particular emotion. Each of the images was linearly stretched over the 8 bit greyscale range to reduce lighting variations. Although care was taken in collecting the original images, natural variations in head size and the mouth's expression resulted in significant variation in the distance between the eyes (2.7 pixels) and in the vertical distance from the eyes to the mouth (5.0 pixels). To achieve scale invariance, each image was scaled so that prominent facial features were located in the same image region. Eye and mouth templates were constructed from a number of images, and the most correlated template was used to localize the respective feature. Similar techniques have been employed in previous work on faces [Brunelli and Poggio, 1993]. Examples of the normalized images and typical facial expressions can be found in Figure 1. 896 C. Padgen and G. W. Conrell A B C Figure 1: The image regions from which the representations are derived. Image A is a typical normalized and cropped image used to generate the full face eigenvectors. Image B depicts the feature regions from which the feature eigenvectors are calculated. Image C indicates each of the block areas projected onto the random block eigenvectors. 3 Representation From the normalized database, we develop three distinct representations that form independent pattern sets for a single classification scheme. The selected representations differ in their scope (features or whole face) and in the nature ofthe sub-space (eigen- faces/features or eigenvectors of random image patches). The more familiar representational schemes (eigenfaces, eigenfeatures) are based on PCA of aligned features or faces. They have been shown to provide a reasonably compact representation for recognition purposes but little is known about their suitability for other classification tasks. Random image patches are used to identify an alternative sub-space from which a set of localized face feature patterns are generated. This space is different in that the sub-space is more general, the variance captured by the leading eigenvectors is derived from patches drawn randomly over the set of face images. As we seek to develop generalizations across the rather small portion of image space containing faces or features , perturbations in this space will hopefully reflect more about class characteristics than individual distinctions. For each of the pattern sets, we normalized the resultant set of values obtained from their projections on the eigenvectors by their standard deviation to produce Z scores. The Z score obtained from each image constitutes a single input to the neural network classifier. The highest valued eigenvectors typically contain more average features so that presumably they would be more suitable for object classification. All the representations will make use of the top k principal components. The full-faced pattern has proved to be quite useful in identification and the same techniques using face features have also been valuable [Pentland et al., 1994, Cottrell and Metcalfe, 1991]. However representations useful for identification of individuals may not be suitable for emotion recognition. In determining the appropriate emotion, structural differences in faces need to be suppressed. One way to accomplish this is to eliminate portions of the face image where variation provides little information with respect to emotion. Local changes in facial muscles around the eyes and mouth are generally associated with our perception of emotions [Ekman and Friesen, 1977]. The full face images presumably contain much information that is simply irrelevant to the task at hand which could impact the ability of the classifier to uncover the signal. Representing Face Images for Emotion Classification 897 The feature based representations are derived from local windows around the eyes and mouth of the normalized whole face images (see Fig. IB). The eigenvectors of the feature sub-space are determined independently for each feature (left/right eye and mouth). A face pattern is generated by projecting the particular facial features on their respective eigenvectors. The random block pattern set is formed from image blocks extracted around the feature locations (see Fig. lC). The areas around each eye are divided into two vertically overlapping blocks of size 32x32 and the mouth is sectioned into three. However, instead of performing PCA on each individual block or all of them together, a more general PCA of random 32x32 blocks taken over the entire image was used to generate the eigenvectors. We used random blocks to reduce the uniqueness of a projection for a single individual and provide a more reasonable model of the early visual system. The final input pattern consists of the normalized projection of the seven extracted blocks for the image on the top n principal components. 4 Classifier design and training The principal goal of classification for this study is to examine how the different representational spaces facilitate a classifiers ability to generalize to novel individuals. Comparing expected recognition rate error using the same classification technique with different representations should provide an indication of how well the signal of interest is preserved by the respective representation. A neural network with a hidden layer employing a non-linear activation function (sigmoid) is trained to learn the input-output mapping between the representation of the face image and the associated response vector given by human subjects. A simple, fully connected, feed-forward neural network containing a single hidden layer with 10 nodes, when trained using back propagation, is capable of correctly classifying the input of training sets from each of the three representations (tested for pattern sizes up to 140 dimensions). The architecture of the network is fixed for a particular input size (based on the number of projections on the respective sub-space) and the generalization of the network is found on a set of images from a novel individual. An overview of the network design is shown in Fig. 2. To minimize the impact of choosing a poor hold out set from the training set, each of the 11 individuals in the training set was in turn used as a hold out. The results of the 11 networks were then combined to evaluate the classification error on the test set. A number of different techniques are possible: winner take all, weighted average output, voting, etc. The method that we found to consistently give the highest generalization rate involved combining Z scores from the 11 networks. The average output for each possible emotion across all the networks was calculated along with its deviation over the entire training set. These values were used to normalize each output of the 11 networks and the highest weighted sum for a particular input was the associated emotion. Due to the limited amount of data available for testing and training, a crossvalidation technique using each set of an individual's images for testing was employed to increase the confidence ofthe generalization measurement. Thus, for each individual, 11 networks were combined to evaluate the generalization on the single test individual, and this procedure is repeated for all 12 individuals to give an average generalization error. This results in a total of 132 networks to evaluate the entire database. A single trial consisted of the generalization rate obtained over the whole database for a particular size of input pattern. By varying the initial weights of the network, we can determine the expected generalization performance of this type 898 C. Padgett and G. W. Cottrell Human ReSPOllge. Ensemble Networks Figure 2: The processing path used to evaluate the pattern set of each representation scheme. The original image data (after normalization) is used to generate the eigenvectors and construct the pattern sets. Human responses are used in training the classifiers and determining generalization percentages on the test data. classifier on each representation. The number of projections on the relevant space is also varied to determine a generalization curve for each representation. Constructing, training, and evaluating the 132 networks takes approximately 2 minutes on a SparcStation 10 for input pattern size of 15 and 4 minutes for an input pattern size of 80. 5 Results Fig. 3 provides the expected generalization achIeved by the neural network architecture initially seeded with small random weights for an increasing number of projections in the respective representational spaces. Each data point represents the average of 20 trials, 1 (T error bars show the amount of error with respect to the mean. The curve (generalization rate vs. input pattern size) was evaluated at 6 points for the whole face and at 8 points for each feature based approach. The eigenfeature representation made use of up to 40 eigenvectors for the three regions while the random block representation made use of up to 17 eigenvectors for each of its seven regions. For the most part, all the representations show improvement as the number of projections increase. Variations as input size increases are most likely due to a combination of two factors: lower signal to noise ratios (SNR) for higher order projections; and the increasing number of para~eters with a fixed number of patterns, making generalization difficult. The highest average recognition rate achieved by the neural network ensembles is 86%, found using the random block representation with 15 projections per block. The results indicate that the generalization rate for emotion classification varies significantly depending on the representational strategy. Both local feature-based approaches (eigenfeatures and random block) did significantly better over their shared range than the eigenface representation. Over most of the range, the random block representation is clearly superior to the eigenfeature representation even though both are derived from the same image area. Representing Face Images for Emotion Classification o.9r----,---..,..----,.----,.------,,.-----, 0.85 ~ 0.8 c: .Q ~ ·ill 0.75 ~ ~ ~ .. ~ 0.7 random block eigenface O·~':-O --r--4:':-O---6"':0---BO-.l-:----l~OO-:------'12-0--~140 Number of inputs (projections on eignevectors) 899 Figure 3: Generalization curves for feature-based representation and full-face representation. 6 Discussion Fig. 3 clearly demonstrates that reasonable recognition rates can be obtained for novel individuals using representational techniques that were found useful for identity. The 86% generalization rate achieved by the neural network ensemble using random block patterns with 105 projections compares favorably with the results obtained from techniques that use an expression sequence (neutral to expression) [Mase, 1991, Yacoob and Davis, 1996, Bartlett et al., 1996]. Such schemes make use of a neutral mask which enhances the sequence's expression by simple subtraction, a technique that is not possible on novel, static face images. That our technique works as well or better indicates the possibility that the human visual system need not rely on difference image strategies over sequences of images in classifying emotions. As many psychological studies are performed on static images of individuals, models that can accommodate this aspect of emotion recognition can make predictions that directly guide research [Padgett et al., 1996]. As for the suitability of the various representations for fine grained discrimination over different individual objects (as required by emotion classification), Fig. 3 clearly demonstrates the benefits accrued by concentrating on facial features important to emotion. The generalization of the trained networks making use of the two local feature-based representations averages 6-15% higher than do the networks trained using projections on the eigenfaces. The increased performance can be attributed to a better signal to noise ratio for the feature regions. As much of the face is rigid (e.g. the chin and forehead), these regions provide little in the way of information useful in classifying emotions. However, there are substantial differences in these areas between individuals, which will be expressed by the principal component analysis of the images and thus reflected in the projected values. These variations are essentially noise with respect to emotion recognition making it more difficult for the classifier to extract useful generalizations during learning. The final point is the superiority of the random block representation over the range 900 C. Padgett and G. W. Cottrell examined. One possible explanation for its significant performance edge is that major feature variations (e.g. open mouth, open eyes, etc.) are more effectively preserved by this representation than the eigenfeature approach, which covers the same image area. Due to individual differences in mouth/eye structure, one would expect that many of the eigenvectors of the feature space would be devoted to this variance. Facial expressions could be substantially orthogonal to this variance, so that information pertinent to emotion discrimination is effectively hidden. This of course would imply that the eigenfeature representation should be better than the random block representation for face recognition purposes. However, this is not the case. Nearest neighbor classification of individuals using the same pattern sets shows that the random block representation does better for this task as well (results not shown). We are currently developing a noise model that looks promising as an explanation for this phenomenon. 7 Conclusion We have demonstrated that average generalization rates of 86% can be obtained for emotion recognition on novel individuals using techniques similar to work done in face recognition. Previous work on emotion recognition has relied on image sequences and obtained recognition rates of nearly the same generalization. The model we developed here is potentially of more interest to researchers in emotion that make use of static images of novel individuals in conducting their tests. Future work will compare aspects of the network model with human performance. References [Bartlett et al., 1996] Bartlett, M., Viola, P., Sejnowski, T ., Larsen, J., Hager, J., and Ekman, P. (1996). Classifying facial action. In Touretzky, D., Mozer, M. , and Hasselmo, M., editors, Advances in Neural Information Processing Systems 8, Cambridge, MA. MIT Press. [Brunelli and Poggio, 1993] Brunelli, R. and Poggio, T. (1993). Face recognition: Feature versus templates. IEEE Trans. Patt. Anal. Machine Intell., 15(10). [Cottrell and Metcalfe, 1991] Cottrell, G. W. and Metcalfe, J. (1991). Empath: Face, gender and emotion recognition using holons. In Lippman, R., Moody, J ., and Touretzky, D., editors, Advances in Neural Information Processing Systems 3, pages 564-571, San Mateo. Morgan Kaufmann. [Ekman and Friesen, 1976] Ekman, P. and Friesen, W. (1976). Pictures of facial affect. [Ekman and Friesen, 1977] Ekman, P. and Friesen, W. (1977). Facial Action Coding System. Consulting Psychologists, Palo Alto, CA. [Mase, 1991] Mase, K. (1991). Recognition of facial expression from optical flow. IEICE Transactions, 74(10):3474-3483. [Padgett et al., 1996] Padgett, C., Cottrell, G., and Adolphs, R. (1996). Categorical perception in facial emotion classification. In Cottrell, G., editor, Proceedings of the 18th Annual Cognitive Science Conference, San Diego CA . [Pentland et al., 1994] Pentland, A. P., Moghaddam, B., and Starner, T. (1994). View-based and modular eigenspaces for face recognition. In IEEE Conference on Computer Vision fj Pattern Recognition. [Yacoob and Davis, 1996] Yacoob, Y. and Davis, L. (1996). Recognizing human facial expressions from long image sequences using optical flow. IEEE Transactions on Pattern Analysis and Machine Intelligence, 18:636-642.
1996
63
1,262
Separating Style and Content Joshua B. Tenenbaum Dept. of Brain and Cognitive Sciences Massachusetts Institute of Technology Cambridge, MA 02139 jbtGpsyche.mit.edu William T. Freeman MERL, Mitsubishi Electric Res. Lab. 201 Broadway Cambridge, MA 02139 freemanOmerl.com Abstract We seek to analyze and manipulate two factors, which we call style and content, underlying a set of observations. We fit training data with bilinear models which explicitly represent the two-factor structure. These models can adapt easily during testing to new styles or content, allowing us to solve three general tasks: extrapolation of a new style to unobserved content; classification of content observed in a new style; and translation of new content observed in a new style. For classification, we embed bilinear models in a probabilistic framework, Separable Mixture Models (SMMsj, which generalizes earlier work on factorial mixture models [7, 3]. Significant performance improvement on a benchmark speech dataset shows the benefits of our approach. 1 Introduction In many pattern analysis or synthesis tasks, the observed data are generated from the interaction of two underlying factors which we will generically call "style" and "content." For example, in a character recognition task, we might observe different letters in different fonts (see Fig. 1); with handwriting, different words in different writing styles; with speech, different phonemes in different accents; with visual images, the faces of different people under different lighting conditions. Such data raises a number of learning problems. Extracting a hidden two-factor structure given only the raw observations has received significant attention [7, 3], but unsupervised factorial learning of this kind has yet to prove tractable for our focus: real-world data with subtly interacting factors. We work in a more supervised setting, where labels for style or content may be available during training or testing. Figure 1 shows three problems we want to solve. Given a labelled training set of observations in multiple styles, we want to extrapolate a new style to unobserved content classes (Fig. la), classify content observed in a new style (Fig. Ib), and translate new content observed in a new style (Fig. lc). This paper treats these problems in a common framework, by fitting the training data with a separable model that can easily adapt during testing to new styles or content classes. We write an observation vector in style s and content class c as y3C. We seek to fit these observations with some model y3C = f(a 3 , bC; W), (1) where a particular functional form of f is assumed. We must estimate parameter vectors a 3 and b C describing style s and content c, respectively, and W, parameters Separating Style and Content 663 Extrapolation Classification Translation H B [ 0 E ) Jl '.8 C 'l> '£ A B C D E A r C II E. A B C D E H B I: 0 E V Jl 'B C 1J) '£ A B C D E A r c II e AB C D E R B [ 0 E ? ? ? Jl '.8 C'D '£ I I A B C D E I I A J C n E. I I A BC D E ? ? ? A B C ? ? I ED CB A ? - - r? F G H (a) (b) (c) Figure 1: Given observations of content (letters) in different styles (fonts), we want to extrapolate, classify, and translate observations from a new style or content class. for f that are independent of style and content but govern their interaction. In terms of Fig. 1 (and in the spirit of [8]), the model represents what the elements of each row have in common independent of column (a&), what the elements of each column have in common independent of row (b C), and what all elements have in common independent of row and column (W). With these three modular components, we can solve problems like those illustrated in Fig. 1. For example, we can extrapolate a new style to unobserved content classes (Fig. 1a) by combining content and interaction parameters learned during training with style parameters estimated from available data in the new style. 2 Bilinear models We propose to separate style and content using bilinear models - two-factor models that are linear in either factor when the other is held constant. These simple models are still complex enough to model subtle interactions of style and content. The empirical success of linear models in many pattern recognition applications with single-factor data (e.g. "eigenface" models of faces under varying identity but constant illumination and pose [15], or under varying illumination but constant identity and pose [5] ), makes bilinear models a natural choice when two such factors vary independently across the data set. Also, many of the computationally desirable properties of linear models extend to bilinear models. Model fitting (discussed in Section 3 below) is easy, based on efficient and well-known techniques such as the singular value decomposition (SVD) and the expectation-maximization (EM) algorithm. Model complexity can be controlled by varying model dimensionality to achieve a compromise between reproduction of the training data and generalization during testing. Finally, the approach extends to multilinear models [10], for data generated by three or more interacting factors. We have explored two bilinear models for Eq. 1. In the symmetric model (so called because it treats the two factors symmetrically), we assume f is a bilinear mapping given by yi/ = a&TWkb C = LaibjWijk. ij (2) The Wij k parameters represent a set of basis functions independent of style and content, which characterize the interaction between these two factors. Observations in style s and content c are generated by mixing these basis functions with coefficients 664 1. B. Tenenbaum and W T. Freeman given by the tensor product of a 3 and be vectors. The model exactly reproduces the observations when the dimensionalities of a 3 and be equal the number of styles N3 and content classes Ne observed. It finds coarser but more compact representations as these dimensionalities are decreased. Sometimes it may not be practical to represent both style and content with lowdimensional vectors. For example, a linear combination of a few basis styles learned during training may not describe new styles well. We can obtain more flexible, asymmetric bilinear models by letting the basis functions Wijk themselves depend on style or content. For example, if the basis functions are allowed to depend on style, the bilinear model from Eq. 2 becomes Yi/ = Lij at bj W/j k' This simplifies to Yi/ = Lj Aj kbj, by summing out the i index and identifying Aj k == Li at W/j k' In vector notation, we have y3C = A 3 b c , (3) where A 3 is a matrix of basis functions specific to style s (independent of content), and be is a vector of coefficients specific to content c (independent of style). Alternatively, the basis functions may depend on content, which gives (4) Asymmetric models do not parameterize the rendering function f independently of style and content, and so cannot translate across both factors simultaneously (Fig. lc). Further, a matrix representation of style or content may be too flexible and overfit the training data. But if overfitting is not a problem or can be controlled by some additional constraint, asymmetric models may solve extrapolation and classification tasks using less training data than symmetric models. Figure 2 illustrates an example of an asymmetric model used to separate style and content. We have collected a small database offace images, with 11 different people (content classes) in 15 different head poses (styles). The images are 22 x 32 pixels, which we treat as 704-dimensional vectors. A subset of the data is shown in Fig. 2a. Fig. 2b schematically depicts an asymmetric bilinear model of the data, with each pose represented by a set of basis vectors Apose (shown as images) and each person represented by a set of coefficients bperson. To render an image of a particular person in a particular pose, the pose-specific basis vectors are mixed according to the person-specific coefficients. Note that the basis vectors for each pose look like eigenfaces [15] in the appropriate style of each pose. However, the bilinear structure of the model ensures that corresponding basis vectors play corresponding roles across poses (e.g. the first vector holds (roughly) the mean face for that pose, the second may modulate overall head size, the third may modulate head thickness, etc.), which is crucial for adapting to new styles or content classes. 3 Model fitting All the tasks shown in Fig. 1 break down into a training phase and a testing phase; both involve some model fitting. In the training phase (corresponding to the first 5 rows and columns of Figs. la-c), we learn all the parameters of a bilinear model from a complete matrix of observations of Nc content classes in N3 styles. In the testing phase (corresponding to the final rows of Figs. la, b and the final row and last 3 columns of Fig. lc), we adapt the same model to data in a new style or content class (or both), estimating new parameters for the new style or content, clamping the other parameters. Then new and old parameters are combined to accomplish the desired classification, extrapolation, or translation task. This section focuses on the asymmetric model and its use in extrapolation and classification. Training and Separating Style and Content (a) Faces rendered from: y = APose lJPerson AUp-Right A Up-Straight A Up-Left (b) 665 Figure 2: An illustraton of the asymmetric model, with faces varying in identity and head pose. adaptation procedures for the symmetric model are similar and based on algorithms in [10, 11]. In [2], we describe these procedures and their application to extrapolation and translation tasks. 3.1 Training Let nSC be the number of observations in style s and content c, and let mSc = L: y8C be the sum of these observations. Then estimates of AS and b C that minimize the sum-of-squared-errors for the asymmetric model in Eq. 3 can be found by iterating the fixed point equations obtained by setting derivatives of the error equal to O. To ensure stability, we update the parameters according to AS = (l-"7)As+"7As and b C = (l-"7)bc+"7bc, typically using a stepsize 0.2 < "7 < 0.5. Replacing AS with BC and b C with as yields the .analogous procedure for training the model in Eq. 4. If the same number of observations are available for all style-content pairs, there exists a closed-form procedure to fit the asymmetric model using the SVD. Let the K-dimensional vector y8C denote the mean of the observed data generated by style s and content c, and stack these vectors into a single (K x Ns ) x Nc matrix y= (6) We compute the SVD ofY = USVT, and define the (K x Ns ) x J matrix A to be the first J columns of U, and the J x Nc matrix B to be the first J rows of SVT . Finally, we identify A and B as the desired parameter estimates in stacked form 666 J. B. Tenenbaum and W T. Freeman (see also [9, 14]), (7) The model dimensionality J can be chosen in various standard ways: by a priori considerations, by requiring a sufficiently good approximation to the data (as measured by mean squared error or some more subjective metric), or by looking for a gap in the singular value spectrum. 3.2 Testing It is straightforward to adapt the asymmetric model to an incomplete new style s* , in order to extrapolate that style to unseen content. We simply estimate A 3 from Eq. 5, using b C values learned during training and restricting the sums over e to those content classes observed in the new style. Then data in content e and style s* can be synthesized from A $- b C • Extrapolating incomplete new content to unseen styles is done similarly. Adapting the asymmetric model for classification in new styles is more involved, because the content class of the new data (and possibly its style as well) is unlabeled. To deal with this uncertainty, we embed the bilinear model within a gaussian mixture model to yield a separable mixture model (SMM), which can then be fit efficiently to data in new styles using the EM algorithm. Specifically, we assume that the probability of a new, unlabeled observation y being generated by style sand content c is given by a spherical gaussian centered at the prediction of the asymmetric bilinear model: p(Yls, e) <X exp{ -IIY - A$b C 112 j(2(T2)}. The total probability of Y is then p(y) = 2:3 cp(yls, e)p(s, e); we use equal priors p(s, e). We assume that the content vectors 'bc are known from training, but that new style matrices A $ must be found to explain the test data. The EM algorithm alternates between computing soft style and content-class assignments p(s, ely) = p(yls, e)p(s, e)jp(y) for each test vector y given the current style matrix estimates (E-step), and estimating new style matrices by setting A3 to maximize 2:y logp(y) (M-step). The M-step is solved in closed form using the update rule for A 3 from Eq. 5, with m 3C = 2:y p(s, elY) y and nu = 2:y p(s, ely)· Test vectors in new styles can now be classified by grouping each vector y with the content class e that maximizes p(cly) = 2:3 p(s, ely)· 4 Application: speaker-adaptive speech recognition This example illustrates our approach to style-adaptive classification on a real-world data set that is a benchmark for many connectionist learning algorithms. The data consist of 6 samples of each of 11 vowels uttered by 15 speakers of British English (originally collected by David Deterding, from the eMU neural-bench ftp archive). Each data vector consists of 10 parameters computed from a linear predictive analysis of the digitized speech. Robinson [13] compared many learning algorithms trained to categorize vowels from the first 8 speakers (4 male and 4 female) and tested on samples from the remaining 7 speakers (4 male and 3 female). Using the SVD-based procedure described above, we fit an asymmetric bilinear model to the training data, labeled by style (speaker) and content (vowel). We then used the learned vowel parameters b C in an SMM and tested classification performance with varying degrees of style information for the 7 new speakers' data: Separating Style and Content 667 both style and content labels missing for each test vector (SMMl), style labels present (indicating a change of speaker) but content labels missing (SMM2), and both labels missing but with the test data loglikelihood 2:y logp(y) augmented by a prior favoring temporal continuity of style assignments (SMM3). The few training styles makes this problem difficult and a good showcase for our approach. Robinson [13] obtained 51% correct vowel classification on the test set with a multi-layer perceptron and 56% with a I-nearest neighbor (l-NN) classifier, the best performance of the many standard techniques he tried. Hastie and Tibshirani [6] recently obtained 62% correct using their discriminant adaptive nearest neighbor algorithm, the best result we know of for an approach that does not model speaker style. We obtained 69% correct for SMMl, 77% for SMM2, and 76% for SMM3, using a model dimensionality of J = 4, model variance of (j2 = .5, and using the vowel class assignments of I-NN to initialize the E-step of EM. While good initial conditions were important for the EM algorithm, a range of model dimensionality and variance settings gave reasonable performance. We also applied these methods to the head pose data of Fig. 2a. We trained on 10 subjects in the 15 poses, and used SMM2 to learn a style model for a new person while simultaneously classifying the head poses. We obtained 81% correct pose categorization (averaged over all 11 test subjects), compared with 53% correct performance for I-NN matching. These results demonstrate that modeling style and content can substantially improve content classification in new styles even when no style information is available during testing (SMMl), and dramatically so when some style demarkation is available explicitly (SMM2) or implicitly (SMM3). Bilinear models offer an easy way to improve performance using style labels which are frequently available for many classification tasks. 5 Pointers to other work and conclusions We discuss the extrapolation and translation problems in [2]. Here we summarize results. Figure 3 shows extrapolation of a partially observed font (Monaco) to the unseen letters (see also the gridfont work of [8, 4]). During training, we presented all letters of the five fonts shown at the left. To accomodate many shape topologies, we described letters by the warps of black particles from a reference shape into the letter shape. During testing, we fit an asymmetric model style matrix to all the letters of the Monaco font except those shown in the figure. We used the best fitting linear combination of training fonts as a prior for the style matrix, in order to control model complexity. Using the fit style, we then synthesized the unseen letters of the Monaco font . These compare well with the actual letters in that style. Because the W weights of the symmetric model are independent of any particular style and content class, they allow translation of observations from unknown styles and content-classes to known ones. During training, we fit the symmetric model to the observations. For a test observation under a new style and content class, we find a 3 and be values using the known W numbers, iterating least squares fits of the two parameters. Typically, the resulting a 3 and be vectors are unique up to an uncertainty in scale. We have used this approach to translate across shape or lighting conditions for images of faces, and to translate across illumination color for color measurements (assuming small specular reflections). Our work naturally combines two current themes in the connectionist learning literature: factorial learning [7, 3] and learning a family of many related tasks [1, 12] 668 1. B. Tenenbaum and W. T. Freeman u I 1 Training fonts i R .111 A It A A A B 2l 8 J B 8 8 C C C C C C C 0 'D 0 D D D 0 E :& E £ E E E F 7 F F •• f F 6 ~ G a 0G G H :Ii H H H U H I I I I 1 I Figure 3: Style extrapolation in typography. The training data were all letters of the 5 fonts at left. The test data were all the Monaco letters except those shown at right. The synthesized Monaco letters compare well with the missing ones. to facilitate task transfer. Separable bilinear models provide a powerful framework for separating style and content by combining explicit representation of each factor with the computational efficiency of linear models. Acknowledgements We thank W. Richards, Y. Weiss, and M. Bernstein for helpful discussions. Joshua Tenenbaum is a Howard Hughes Medical Institute Predoctoral Fellow. References [1] R. Caruana. Learning many related tasks at the same time with backpropagation. In Adv. in Neural Info. Proc. Systems, volume 7, pages 657-674, 1995. [2] W. T. Freeman and J. B. Tenenbaum. Learning bilinear models for two-factor problems in vision. TR 96-37, MERL, 201 Broadway, Cambridge, MA 02139, 1996. [3] Z. Ghahramani. Factorial learning and the EM algorithm. In Adv. in Neural Info. Proc. Systems, volume 7, pages 617- 624, 1995. [4] I. Grebert, D. G. Stork, R. Keesing, and S. Mims. Connectionist generalization for production: An example from gridfont. Neural Networks, 5:699-710, 1992. [5] P. W. Hallinan. A low-dimensional representation of human faces for arbitrary lighting conditions. In Proc. IEEE CVPR, pages 995- 999, 1994. [6] T. Hastie and R. Tibshirani. Discriminant adaptive nearest neighbor classification. IEEE Pat. Anal. Mach. Intell., (18):607-616, 1996. [7] G. E. Hinton and R. Zemel. Autoencoders, minimum description length, and Helmholtz free energy. In Adv. in Neural Info. Proc. Systems, volume 6, 1994. [8] D. Hofstadter. Fluid Concepts and Creative Analogies. Basic Books, 1995. [9] J. J. Koenderink and A. J. van Doorn. The generic bilinear calibration-estimation problem. Inti. J. Compo Vis., 1997. in press. [10] J. R. Magnus and H. Neudecker. Matrix differential calculus with applications in statistics and econometrics. Wiley, 1988. [ll] D. H. Marimont and B. A. Wandell. Linear models of surface and illuminant spectra. J. Opt. Soc. Am. A, 9(1l):1905-1913, 1992. [12] S. M. Omohundro. Family discovery. In Adv. in Neural Info. Proc. Sys., vol. 8, 1995. [13] A. Robinson. Dynamic error propagation networks. PhD thesis, Cambridge University Engineering Dept., 1989. [14] C. Tomasi and T. Kanade. Shape and motion from image streams under orthography: a factorization method. Inti. J. Compo Vis., 9(2):137-154, 1992. [15] M. Turk and A. Pentland. Eigenfaces for recognition. J. Cog. Neurosci., 3(1), 1991.
1996
64
1,263
Contour Organisation with the EM Algorithm J. A. F. Leite and E. R. Hancock Department of Computer Science University of York, York, Y01 5DD, UK. Abstract This paper describes how the early visual process of contour organisation can be realised using the EM algorithm. The underlying computational representation is based on fine spline coverings. According to our EM approach the adjustment of spline parameters draws on an iterative weighted least-squares fitting process. The expectation step of our EM procedure computes the likelihood of the data using a mixture model defined over the set of spline coverings. These splines are limited in their spatial extent using Gaussian windowing functions. The maximisation of the likelihood leads to a set of linear equations in the spline parameters which solve the weighted least squares problem. We evaluate the technique on the localisation of road structures in aerial infra-red images. 1 Introduction Dempster, Laird and Rubin's EM (expectation and maximisation) [1] algorithm was originally introduced as a means of finding maximum likelihood solutions to problems posed in terms of incomplete data. The basic idea underlying the algorithm is to iterate between the expectation and maximisation modes until convergence is reached. Expectation involves computing a posteriori model probabilities using a mixture density specified in terms of a series of model parameters. In the maximisation phase, the model parameters are recomputed to maximise the expected value of the incomplete data likelihood. In fact, when viewed from this perspective, the updating of a posteriori probabilities in the expectation phase would appear to have much in common with the probabilistic relaxation process extensively exploited in low and intermediate level vision [9, 2]. Maximisation of the incomplete Contour Organisation with the EM Algorithm 881 data likelihood is reminiscent of robust estimation where outlier reject is employed in the iterative re-computation of model parameters [7]. It is these observations that motivate the study reported in this paper. We are interested in the organisation of the output of local feature enhancement operators into meaningful global contour structures [13, 2]. Despite providing one of the classical applications of relaxation labelling in low-level vision [9], successful solutions to the iterative curve reinforcement problem have proved to be surprisingly elusive [8, 12, 2]. Recently, two contrasting ideas have offered practical relaxation operators. Zucker et al [13] have sought biologically plausible operators which draw on the idea of computing a global curve organisation potential and locating consistent structure using a form of local snake dynamics [11]. In essence this biologically inspired model delivers a fine arrangement of local splines that minimise the curve organisation potential. Hancock and Kittler [2], on the other hand, appealed to a more information theoretic motivation [4]. In an attempt to overcome some of the well documented limitations of the original Rosenfeld, Hummel and Zucker relaxation operator [9] they have developed a Bayesian framework for relaxation labelling [4]. Of particular significance for the low-level curve enhancement problem is the underlying statistical framework which makes a clear-cut distinction between the roles of uncertain image data and prior knowledge of contour structure. This framework has allowed the output of local image operators to be represented in terms of Gaussian measurement densities, while curve structure is represented by a dictionary of consistent contour structures [2]. While both the fine-spline coverings of Zucker [13] and the dictionary-based relaxation operator of Hancock and Kittler [2] have delivered practical solutions to the curve reinforcement problem, they each suffer a number of shortcomings. For instance, although the fine spline operator can achieve quasi-global curve organisation, it is based on an essentially ad hoc local compatibility model. While being more information theoretic, the dictionary-based relaxation operator is limited by virtue of the fact that in most practical applications the dictionary can only realistically be evaluated over at most a 3x3 pixel neighbourhood. Our aim in this paper is to bridge the methodological divide between the biologically inspired fine-spline operator and the statistical framework of dictionary-based relaxation. We develop an iterative spline fitting process using the EM algorithm of Dempster et al [1] . In doing this we retain the statistical framework for representing filter responses that has been used to great effect in the initialisation of dictionary-based relaxation. However, we overcome the limited contour representation of the dictionary by drawing on local cubic splines. 2 Prerequisites The practical goal in this paper is the detection of line-features which manifest themselves as intensity ridges of variable width in raw image data. Each pixel is characterised by a vector of measurements, ?<i where i is the pixel-index. This measurement vector is computed by applying a battery of line-detection filters of various widths and orientations to the raw image. Suppose that the image data is indexed by the pixel-set I. Associated with each image pixel is a cubic spline parameterisation which represents the best-fit contour that couples it to adjacent feature pixels. The spline is represented by a vector of parameters denoted by 882 1. A. F. Leite and E. R. Hancock 9.; = (q?, q}, q; , qr) T . Let (Xi, Yi) represent the position co-ordinates of the pixel indexed i. The spline variable, Si,j = X i - Xj associated with the contour connecting the pixel indexed j is the horizontal displacement between the pixels indexed i and j. We can write the cubic spline as an inner product F(Si,j, g) = g[.Si,j where Si,j = (1, Si,j , S~, j' S~,j) T . Central to our EM algorithm will be the comparison of the predicted vertical spline displacement with its measured value Ti,j = Yi Yj. In order to initialise the EM algorithm, we require a set of initial spline probabilities which we denote by 7l'(q(O»). Here we use the multi-channel combination model -t recently reported by Leite and Hancock [5] to compute an initial multi-scale linefeature probability. Accordingly, if ~ is the variance-covariance matrix for the components of the filter bank, then 7l'(g~O») = l-exp [-~~r~-l~i] (1) The remainder of this paper outlines how these initial probabilities are iteratively refined using the EM algorithm. Because space is limited we only provide an algorithm sketch. Essential algorithm details such as the estimation of spline orientation and the local receptive gating of the spline probabilities are omitted for clarity. Full details can be found in a forthcoming journal article [6]. 3 Expectation Our basic model of the spline organisation process is as follows. Associated with each image pixel is a spline parameterisation. Key to our philosophy of exploiting a mixture model to describe the global contour structure of the image is the idea that the pixel indexed i can associate to each of the putative splines residing in a local Gaussian window N i . We commence by developing a mixture model for the conditional probability density for the filter response ~ i given the current global spline description. If ~(n) = {q(n), Vi E I} is the global spline description at -t iteration n of the EM process, then we can expand the mixture distribution over a set of putative splines that may associate with the image pixel indexed i p(~il~(n») = L p(~ilg;n»)7l'(g;n») (2) jENi The components of the above mixture density are the conditional measurement densities p(~ilq(n») and the spline mixing proportions 7l'(q(n»). The conditional -J -J measurement densities represent the likelihood that the datum ~i originates from the spline centred on pixel j. The mixing proportions, on the other hand, represent the fractional contribution to the data arising from the jth parameter vector i.e. q(n). -J Since we are interested in the maximum likelihood estimation of spline parameters, we turn our attention to the likelihood of the raw data, i.e. P(~i' Vi E II~(n») = IIp(~il~(n)) (3) iEI The expectation step of the EM algorithm is aimed at estimating the log-likelihood using the parameters of the mixture distribution. In other words, we need to average the likelihood over the space of potential pixel-spline assignments. In fact, Contour Organisation with the EM Algorithm 883 it was Dempster, Laird and Rubin [1] who observed that maximising the weighted log-likelihood was equivalent to maximising the conditional expectation of the likelihood for a new parameter set given an old parameter set. For our spline fitting problem, maximisation of the expectation of the conditional likelihood is equivalent to maximising the weighted log-likelihood function Q(~(n+l)I~(n») = L L p(g;n)l~i) lnp(~ilg;n+l») (4) iEI JEN, The a posteriori probabilities p(q(n)l~i) may be computed from the corresponding -1 components of the mixture density p(~ilqJn») using the Bayes formula P(~i Iq(n) )7r( q(n») p(g;n)l~i) = 2: (11 (n»)) ( (n») kEN. P ~t gk 7r gk (5) For notational convenience, and to make the weighting role of the a posteriori probabilities explicit we use the shorthand w~~) = p(g;n) I~i). Once updated parameter estimates q(n) become available through the maximisation of this criterion, im-t proved estimates of the mixture components may be obtained by substitution into equation (6). The updated mixing proportions, 7r(q(n+l»), required to determine -t the new weights w~~) are computed from the newly available density components using the following estimator p(z ·lq(n»)7r(q(n») 7r( (n+l») = "" -) -i -i gi ~ ~ (I (n») ( (n») JEN. L.JkEIP ~j gk 7r gk (6) In order to proceed with the development of a spline fitting process we require a model for the mixture components, i.e. p(~ilg;n»). Here we assume that the required model can be specified in terms of Gaussian distribution functions. In other words, we confine our attention to Gaussian mixtures. The physical variable of these distributions is the squared error residual for the position prediction of the ith datum delivered by the jth spline. Accordingly we write p(~ilg;n») = a exp [-,8 (ri,j - F(Si,j, g;n»)) 2] (7) where ,8 is the inverse variance of the fit residuals. Rather than estimating ,8, we use it in the spirit of a control variable to regulate the effect of fit residuals. Equations (5), (6) and (11) therefore specify a recursive procedure that iterates the weighted residuals to compute a new mixing proportions based on the quality of the spline fit. 4 Maximisation The maximisation step aims to optimize the quantity Q(~(n+l)I~(n») with respect to the spline parameters. Formally this corresponds to finding the set of spline parameters which satisfy the condition (8) 884 1. A. F. Leite and E. R. Hancock We find a local approximation to this condition by solving the following set of linear equations 8Q( <)(n+l) I<)(n») 8(qf)(n+l) = 0 (9) for each spline parameter (qf)(n+l) in turn, i.e. for k=O,1,2,3. Recovery of the splines is most conveniently expressed in terms of the following matrix equation for the components of the parameter-vector q(n) -t (10) The elements of the vector x(n) are weighted cross-moments between the parallel and perpendicular spline distances in the Gaussian window, i.e. x(n) = -t (11) The elements of the matrix A~n), on the other hand, are weighted moments computed purely in terms of the parallel distance Si ,j. If k and I are the row and column indices, then the (k, l)th element of the matrix A~n) is [A(n)]k I = " w(n) sk+I-2 t , L t,) t,l (12) JENi 5 Experiments We have evaluated our iterative spline fitting algorithm on the detection of linefeatures in aerial infra-red images. Figure 1a shows the original picture. The initial feature probabilities (i.e. 71" { q~O»)) assigned according to equation (1) are shown in Figure lb. Figure 1c shows the final contour-map after the EM algorithm has converged. Notice that the line contours exhibit good connectivity and that the junctions are well reconstructed. We have highlighted a subregion of the original image. There are two features in this subregion to which we would like to draw attention. The first of these is centred on the junction structure. The second feature is a neighbouring point on the descending branch of the road. Figure 2 shows the iterative evolution of the cubic splines at these two locations. The spline shown in Figure 2a adjusts to fit the upper pair of road segments. Notice also that although initially displaced, the final spline passes directly through the junction. In the case of the descending road-branch the spline shown in Figure 2b recovers from an initially poor orientation estimate to align itself with the underlying road structure. Figure 2c shows how the spline probabilities (i.e. 7I"(q~n»)) evolve with iteration number. Initially, the neighbourhood is highly ambiguous. Many neighbouring splines compete to account for the local image structure. As a result the detected junction is several pixels wide. However, as the fitting process iterates, the splines move from the inconsistent initial estimate to give a good local estimate which is less ambiguous. In other words the two splines illustrated in Figure 2 have successfully arranged themselfs to account for the junction structure. Contour Organisation with the EM Algorithm 885 (a) Original image. (b) Probability map. (c) Line map. Figure 1: Infra-red aerial picture with corresponding probability map showing region containing pixel under study and correspondent line map. 6 Conclusions We have demonstrated how the process of parallel iterative contour refinement can be realised using the classical EM algorithm of Dempster, Laird and Rubin [1]. The refinement of curves by relaxation operations has been a preoccupation in the literature since the seminal work of Rosenfeld, Hummel and Zucker [9]. However, it is only recently that successful algorithms have been developed by appealing to more sophisticated modelling methodologies [13, 2]. Our EM approach not only delivers comparable performance, it does so using a very simple underlying model. Moreover, it allows the contour re-enforcement process to be understood in a weighted least-squares optimisation framework which has many features in common with snake dynamics [11] without being sensitive on the initial positioning of control points. Viewed from the perspective of classical relaxation labelling [9, 4], the EM framework provides a natural way of evaluating support beyond the immediate object neighbourhood. Moreover, the framework for spline fitting in 2D is readily extendible to the reconstruction of surface patches in 3D [10]. References [1] Dempster A., Laird N. and Rubin D., "Maximum-likelihood from incomplete data via the EM algorithm", J. Royal Statistical Soc. Ser. B (methodological) ,39, pp 1-38, 1977. [2] Hancock E.R and Kittler J., "Edge Labelling using Dictionary-based Probabilistic Relaxation", IEEE PAMI, 12, pp. 161-185, 1990. [3] Jordan M.1. and Jacobs RA, "Hierarchical Mixtures of Experts and the EM Algorithm", Neural Computation, 6, pp. 181-214, 1994. [4] Kittler J. and Hancock, E.R, "Combining Evidence in Probabilistic relaxation", International Journal of Pattern Recognition and Artificial Intelligence, 3, N1, pp 29-51, 1989. [5] Leite J.A.F. and Hancock, E.R, " Statistically Combining and Refining Multichannel Information" , Progress in Image Analysis and Processing III: Edited by S Impedovo, World Scientific, pp. 193-200, 1994. [6] Leite J.A.F. and Hancock, E.R, "Iterative curve organisation with the EM algorithm", to appear in Pattern Recognition Letters, 1997. 886 1. A. F. Leite and E. R. Hancock ~ /' \ ~ ", (ll' Iili> :--I" ~ + (j) (i!) <tii> :f\ ~ ~ I,.) x Iv) (vi> + + ~ so hy) (v) ~ I ~'''' ~ (ril) , (YIIII """X (Ix) " :+ :+ :+ " , ) MI) (i,) ~ (Ie) x ~ ,l) :+ I» ~ (a) (b) (c) Figure 2: Evolution of the spline in the fitting process. The image in (a) is the junction spline while the image in (b) is the branch spline. The first spline is shown in (i), and the subsequent ones from (ii) to (xi). The evolution of the corresponding spline probabilities is shown in (c). [7] Meer P., Mintz D., Rosenfeld A. and Kim D.Y., "Robust Regression Methods for Computer Vision - A Review", International Journal of Computer vision, 6, pp. 59- 70, 1991. [8] Peleg S. and Rosenfeld A" "Determining Compatibility Coefficients for curve enhancement relaxation processes", IEEE SMC, 8, pp. 548-555, 1978. [9] Rosenfeld A., Hummel R.A. and Zucker S.W., "Scene labelling by relaxation operations", IEEE Transactions SMC, SMC-6, pp400-433, 1976. [10] Sander P.T. and Zucker S.W., "Inferring surface structure and differential structure from 3D images" , IEEE PAMI, 12, pp 833-854, 1990. [11] Terzopoulos D., "Regularisation of inverse problems involving discontinuities" , IEEE PAMI, 8, pp 129-139, 1986. [12] Zucker, S.W., Hummel R.A., and Rosenfeld A., "An application ofrelaxation labelling to line and curve enhancement", IEEE TC, C-26, pp. 394-403, 1977. [13] Zucker S., David C., Dobbins A. and Iverson L., "The organisation of curve qetection: coarse tangent fields and fine spline coverings" , Proceedings of the Second International Conference on Computer Vision, pp. 577-586, 1988.
1996
65
1,264
Combining Neural Network Regression Estimate1s with Regularized Linear Weights Christopher J. Merz and Michael J. Pazzani Dept. of Information and Computer Science University of California, Irvine, CA 92717-3425 U.S.A. { cmerz,pazzani }@ics.uci.edu Category: Algorithms and Architectures. Abstract When combining a set of learned models to form an improved estimator, the issue of redundancy or multicollinearity in the set of models must be addressed. A progression of existing approaches and their limitations with respect to the redundancy is discussed. A new approach, PCR *, based on principal components regression is proposed to address these limitations. An evaluation of the new approach on a collection of domains reveals that: 1) PCR* was the most robust combination method as the redundancy of the learned models increased, 2) redundancy could be handled without eliminating any of the learned models, and 3) the principal components of the learned models provided a continuum of "regularized" weights from which PCR * could choose. 1 INTRODUCTION Combining a set of learned modelsl to improve classification and regression estimates has been an area of much research in machine learning and neural networks [Wolpert, 1992, Merz, 1995, Perrone and Cooper, 1992, Leblanc and Tibshirani, 1993, Breiman, 1992, Meir, 1995, Krogh and Vedelsby, 1995, Tresp, 1995, Chan and Stolfo, 1995]. The challenge of this problem is to decide which models to rely on for prediction and how much weight to give each. 1 A learned model may be anything from a decision/regression tree to a neural network. Combining Neural Network Regression Estimates 565 The goal of combining learned models is to obtain a more accurate prediction than can be obtained from any single source alone. One major issue in combining a set of learned models is redundancy. Redundancy refers to the amount of agreement or linear dependence between models when making a set of predictions. The more the set agrees, the more redundancy is present. In statistical terms, this is referred to as the multicollinearity problem. The focus of this paper is to explore and evaluate the properties of existing methods for combining regression estimates (Section 2), and to motivate the need for more advanced methods which deal with multicollinearity in the set of learned models (Section 3). In particular, a method based on principal components regression (PCR, [Draper and Smith, 1981]) is described, and is evaluated emperically demonstrating the it is a robust and efficient method for finding a set of combining weights with low prediction error (Section 4). Finally, Section 5 draws some conclusions. 2 MOTIVATION The problem of combining a set of learned models is defined using the terminology of [Perrone and Cooper, 1992]. Suppose two sets of data are given: a training set 'DTrain = (xm, Ym) and a test set 'DTelt = (Xl, Yl). Now suppose 'DTrain is used to build a set of functions, :F = fi(X), each element of which approximates f(x). The goal is to find the best approximation of f(x) using :F. To date, most approaches to this problem limit the space of approximations of f( x) to linear combinations of the elements of :F, i.e., N j(x) = L Cidi(X) i=l where Cij is the coefficient or weight of fj(x). The focus of this paper is to evaluate and address the limitations of these approaches. To do so, a brief summary of these approaches is now provided progressing from simpler to more complex methods pointing out their limitations along the way. The simplest method for combining the members of :F is by taking the unweighted average, (i.e., Cij = 1/ N). Perrone and Cooper refer to this as the Basic Ensemble Method (BEM), written as N fBEM = I/NLfi(x) i=l This equation can also be written in terms of the misfit function for each fi(X). These functions describe the deviations of the elements of :F from the true solution and are written as mi(X) = f(x) -Ji(x). Thus, N fBEM = f(x) -1/NL mi(x). i=l Perrone and Cooper show that as long as the mi (x) are mutually independent with zero mean, the error in estimating f(x) can be made arbitrarily small by increasing the population size of :F. Since these assumptions break down in practice, 566 C. J. Merz and M. J. Pazzani they developed a more general approach which finds the "optimal,,2 weights while allowing the mi (x) 's to be correlated and have non-zero means. This Generalized Ensemble Method (GEM) is written as N N IGEM = LQ:di(X) = I(x) - LQ:imi(X) i=1 i=l where C is the symmetric sample covariance matrix for the misfit function and the goal is to minimize E7,; Q:iQ:jCii' Note that the misfit functions are calculated on the training data and I(x) is not required. The main disadvantage to this approach is that it involves taking the inverse of C which can be "unstable". That is, redundancy in the members of :F leads to linear dependence in the rows and columns of C which in turn leads to unreliable estimates of C- 1 • To circumvent this sensitivity redundancy, Perrone and Cooper propose a method for discarding member(s) of :F when the strength of its agreement with another member exceeds a certain threshold. Unfortunately, this approach only checks for linear dependence (or redundancy) between pairs of Ii (x) and two Ii (x) for i =1= j. In fact, Ii (x) could be a linear combination of several other members of :F and the instability problem would be manifest. Also, depending on how high the threshold is set, a member of :F could be discarded while still having some degree of uniqueness and utility. An ideal method for weighting the members of :F would neither discard any models nor suffer when there is redundancy in the model set. The next approach reviewed is linear regression (LR)3 which also finds the "optimal" weights for the Ii (x) with respect to the training data. In fact, G EM and LR are both considered "optimal" because they are closely related in that GEM is a form of linear regression with the added constraint that E~1 Q:i = 1. The weights for LR are found as follows4 , N hR = LQ:di(X) i=1 where Like GEM, LR and LRC are subject to the multicollinearity problem because finding the Q:i's involves taking the inverse of a matrix. That is, if the I matrix is composed of li(x) which strongly agree with other members of :F, some linear dependence will be present. 20ptimal here refers to weights which minimize mean square error for the training data. 3 Actually, it is a form of linear regression without the intercept term. The more general form, denote by LRC, would be formulated the same way but with member, fo which always predicts 1. According to [Leblanc and Tibshirani, 1993] having the extra constant term will not be necessary (i.e., it will equal zero) because in practice, E[fi(x)] = E[f(x)]. 4Note that the constraint, E;:'l ai = 1, for GEM is a form of regularization [Leblanc and Tibshirani, 1993]. The purpose of regularizing the weights is to provide an estimate which is less biased by the training sample. Thus, one would not expect GEM and LR to produce identical weights. Combining Neural Network Regression Estimates 567 Given the limitations of these methods, the goal of this research was to find a method which finds weights for the learned models with low prediction error without discarding any of the original models, and without being subject to the multicollinearity problem. 3 METHODS FOR HANDLING MULTICOLLINEARITY In the abovementioned methods, multicollinearity leads to inflation of the variance of the estimated weights, Ck. Consequently, the weights obtained from fitting the model to a particular sample may be far from their true values. To circumvent this problem, approaches have been developed which: 1) constrain the estimated regression coefficients so as to improve prediction performance (Le., ridge regression, RIDG E [Montgomery and Friedman 1993], and principal components regression), 2) search for the coefficients via gradient descent procedures (i.e., Widrow-Hofflearning, GD and EG+- [Kivinen and Warmuth, 1994]), or build models which make decorrelated errors by adjusting the bias of the learning algorithm [Opitz and Shavlik, 1995] or the data which it sees [Meir, 1995]. The third approach ameliorates, but does not solve, the problem because redundancy is an inherent part of the task of combining estimators. The focus of this paper is on the first approach. Leblanc and Tibshirani [Leblanc and Tibshirani, 1993] have proposed several ways of constraining or regularizing the weights to help produce estimators with lower prediction error: 1. Shrink a towards (1/ K, 1/ K, ... ,1/ K)T where K is the number of learned models. 2. 2:~1 Ckj = 1 3. Ckj ~ O,i = 1,2 ... K Breiman [Breiman, 1992] provides an intuitive justification for these constraints by pointing out that the more strongly they are satisfied, the more interpolative the weighting scheme is. In the extreme case, a uniformly weighted set of learned models is likely to produce a prediction between the maximum and minimum predicted values of the learned models. Without these constraints, there is no guarantee that the resulting predictor will stay near that range and generalization may be poor. The next subsection describes a variant of principal components regression and explains how it provides a continuum of regularized weights for the original learned models. 3.1 PRINCIPAL COMPONENTS REGRESSION When dealing with the above mentioned multicollinearity problem, principal components regression [Draper and Smith, 1981] may be used to summarize and extract the "relevant" information from the learned models. The main idea of PCR is to map the original learned models to a set of (independent) principal components in which each component is a linear combination of the original learned models, and then to build a regression equation using the best subset of the principal components to predict lex). The advantage of this representation is that the components are sorted according to how much information (or variance) from the original learned models for which they account. Given this representation, the goal is to choose the number of principal components to include in the final regression by retaining the first k which meet a preselected stopping criteria. The basic approach is summarized as follows: 568 C. J. Merz and M. J. Pazzani 1. Do a principal components analysis (PCA) on the covariance matrix of the learned models' predictions on the training data (i.e., do a PCA on the covariance matrix of M, where Mi,j is the j-th model's reponse for the i-th training example) to produce a set of principal components, PC = {PC1, ... ,PCN }. 2. Use a stopping criteria to decide on k, the number of principal components to use. 3. Do a least squares regression on the selected components (i.e., include PCi for i:::; k). 4. Derive the weights, fri, for the original learned models by expanding /peR* = i31PC1 + ... + i3"PC" according to PCi = ;i,O/O + ... + ;i,N /N, and simplifying for the coefficients of ". Note that ;i,j is the j-th coefficient of the i-th principal component. The second step is very important because choosing too few or too many principal components may result in underfitting or overfitting, respectively. Ten-fold crossvalidation is used to select k here. Examining the spectrum of (N) weight sets derived in step four reveals that PCR* provides a continuum of weight sets spanning from highly constrained (i.e., weights generated from PCR1 satisfy all three regularization constraints) to completely unconstrained (i.e., PCRN is equivalent to unconstrained linear regression). To see that the weights, fr, derived from PCR1 are (nearly) uniform, recall that the first principal component accounts for where the learned models agree. Because the learned models are all fairly accurate they agree quite often so their first principal component weights, ;1,* will be similar. The "Y-weights are in turn multiplied by a constant when PCR1 is regressed upon. Thus, the resulting fri'S will be fairly uniform. The later principal components serve as refinements to those already included producing less constrained weight sets until finally PCRN is included resulting in an unconstrained estimator much like LR, LRC and GEM. 4 EXPERIMENTAL RESULTS The set of learned models, :F, were generated using Backpropogation [Rumelhart, 1986]. For each dataset, a network topology was developed which gave good performance. The collection of networks built differed only in their initial weights5 . Three data sets were chosen: cpu and housing (from the UCI repository), and body/at (from the Statistics Library at Carnegie Mellon University). Due to space limitation, the data sets reported on were chosen because they were representative of the basic trends found in a larger collection of datasets. The combining methods evaluated consist of all the methods discussed in Sections 2 and 3, as well as PCRI and PCRN (to demonstrate PCR*'s most and least regularized weight sets, SThere was no extreme effort to produce networks with more decorrelated errors. Even with such networks, the issue of extreme multicollinearity would still exist because E[f;(x)] = E[fi(x)] for a.ll i and j. Combining Neural Network Regression Estimates 569 Table 1· Results Data bodyfat II cpu housing 11 N 10 50 10 50 10 50 BEM 1.03 1.04 38.57 38.62 2.79 2.77 GEM 1.02 0.86 46.59 227.54 2.72 2.57 LR 1.02 3.09 44.9 238.0 2.72 6.44 RIDGE 1.02 0.826 44.8 191.0 2.72 2.55 GD 1.03 1.04 38.9 38.8 2.79 2.77 EGPM 1.03 1.07 38.4 38.0 2.77 2.75 PCRl 1.04 1.05 39.0 39.0 2.78 2.76 PCRN 1.02 0.848 44.8 249.9 2.72 2.57 PCR* 0.99 0.786 40.3 40.8 2.70 2.56 respectively). The more computationally intense procedures based on stacking and bootstrapping proposed by [Leblanc and Tibshirani, 1993, Breiman, 1992] were not evaluated here because they required many more models (i.e., neural networks) to be generated for each of the elements of F. There were 20 trials run for each of the datasets. On each trial the data was randomly divided into 70% training data and 30% test data. These trials were rerun for varying sizes of F (i.e., 10 and 50, respectively). As more models are included the linear dependence amongst them goes up showing how well the multicollinearity problem is handled6 . Table 1 shows the average residual errors for the each of the methods on the three data sets. Each row is a particular method and each column is the size of F for a given data set. Bold-faced entries indicate methods which were not significantly different from the method with the lowest error (via two-tailed paired t-tests with p :::; 0.05). PCR* is the only approach which is among the leaders for all three data sets. For the body/at and housing data sets the weights produced by BEM, PCRI , GD, and EG+- tended to be too constrained, while the weights for LR tended to be too unconstrained for the larger collection of models. The less constrained weights of GEM, LR, RIDGE, and PCRN severely harmed performance in the cpu domain where uniform weighting performed better. The biggest demonstration of PCR*'s robustness is its ability to gravitate towards the more constrained weights produced by the earlier principal components when appropriate (i.e., in the cpu dataset). Similarly, it uses the less constrained principal components closer to PCRn when it is preferable as in the bodyfat and housing domains. 5 CONCLUSION This investigation suggests that the principal components of a set of learned models can be useful when combining the models to form an improved estimator. It was demonstrated that the principal components provide a continuum of weight sets from highly regularized to unconstrained. An algorithm, PCR*, was developed which attempts to automatically select the subset of these components which provides the lowest prediction error. Experiments on a collection of domains demonstrated PCR*'s ability to robustly handle redundancy in the set of learned models. Future work will be to improve upon PCR* and expand it to the classification task. 6This is verified by observing the eigenvalues of the principal components and values in the covariance matrix of the models in :F 570 C. 1. Merz and M. J. Pazzani References [Breiman et ai, 1984] Breiman, L., Friedman, J .H., Olshen, R.A. & Stone, C.J. (1984). Classification and Regression Trees. Belmont, CA: Wadsworth. [Breiman, 1992] Breiman, L. (1992). Stacked Regression. Dept of Statistics, Berkeley, TR No. 367. [Chan and Stolfo, 1995] Chan, P.K., Stolfo, S.J. (1995). A Comparative Evaluation of Voting and Meta-Learning on Partitioned Data Proceedings of the Twelvth International Machine Learning Conference (90-98). San Mateo, CA: Morgan Kaufmann. [Draper and Smith, 1981] Draper, N.R., Smith, H. (1981). Applied Regression Analysis. New York, NY: John Wiley and Sons. [Kivinen and Warmuth, 1994] Kivinen, J., and Warmuth, M. (1994). Exponentiated Gradient Descent Versus Gradient Descent for Linear Predictors. Dept. of Computer Science, UC-Santa Cruz, TR No. ucsc-crl-94-16. [Krogh and Vedelsby, 1995] Krogh, A., and Vedelsby, J. (1995). Neural Network Ensembles, Cross Validation, and Active Learning. In Advances in Neural Information Processing Systems 7. San Mateo, CA: Morgan Kaufmann. [Hansen and Salamon, 1990] Hansen, L.K., and Salamon, P. (1990). Neural Network Ensembles. IEEE Transactions on Pattern Analysis and Machine Intelligence, 12 (993-1001). [Leblanc and Tibshirani, 1993] Leblanc, M., Tibshirani, R. (1993) Combining estimates in regression and classification Dept. of Statistics, University of Toronto, TR. [Meir, 1995] Meir, R. (1995). Bias, variance and the combination of estimators. In Advances in Neural Information Processing Systems 7. San Mateo, CA: Morgan Kaufmann. [Merz, 1995] Merz, C.J. (1995) Dynamical Selection of Learning Algorithms. In Fisher, C. and Lenz, H. (Eds.) Learning from Data: Artificial Intelligence and Statistics, 5). Springer Verlag [Montgomery and Friedman 1993] Mongomery, D.C., and Friedman, D.J. (1993). Prediction Using Regression Models with Multicollinear Predictor Variables. lIE Transactions, vol. 25, no. 3 73-85. [Opitz and Shavlik, 1995] Opitz, D.W., Shavlik, J .W. (1996). Generating Accurate and Diverse Members of a Neural-Network Ensemble. Advances in Neural and Information Processing Systems 8. Touretzky, D.S., Mozer, M.C., and Hasselmo, M.E., eds. Cambridge MA: MIT Press. [Perrone and Cooper, 1992] Perrone, M. P., Cooper, L. N., (1993). When Networks Disagree: Ensemble Methods for Hybrid Neural Networks. Neural Networks for Speech and Image Processing, edited by Mammone, R. J .. New York: Chapman and Hall. [Rumelhart, 1986] Rumelhart, D. E., Hinton, G. E., & Williams, R. J . (1986). Learning Interior Representation by Error Propagation. Parallel Distributed Processing, 1 318-362. Cambridge, MASS.: MIT Press. [Tresp, 1995] Tresp, V., Taniguchi, M. (1995). Combining Estimators Using NonConstant Weighting Functions. In Advances in Neural Information Processing Systems 7. San Mateo, CA: Morgan Kaufmann. [Wolpert, 1992] Wolpert, D. H. (1992). Stacked Generalization. Neural Networks, 5, 241-259.
1996
66
1,265
Triangulation by Continuous Embedding Marina MeiHl and Michael I. Jordan {mmp, jordan }@ai.mit.edu Center for Biological & Computational Learning Massachusetts Institute of Technology 45 Carleton St. E25-201 Cambridge, MA 02142 Abstract When triangulating a belief network we aim to obtain a junction tree of minimum state space. According to (Rose, 1970), searching for the optimal triangulation can be cast as a search over all the permutations of the graph's vertices. Our approach is to embed the discrete set of permutations in a convex continuous domain D. By suitably extending the cost function over D and solving the continous nonlinear optimization task we hope to obtain a good triangulation with respect to the aformentioned cost. This paper presents two ways of embedding the triangulation problem into continuous domain and shows that they perform well compared to the best known heuristic. 1 INTRODUCTION. WHAT IS TRIANGULATION? Belief networks are graphical representations of probability distributions over a set of variables. In what follows it will be always assumed that the variables take values in a finite set and that they correspond to the vertices of a graph. The graph's arcs will represent the dependencies among variables. There are two kinds of representations that have gained wide use: one is the directed acyclic graph model, also called a Bayes net, which represents the joint distribution as a product of the probabilities of each vertex conditioned on the values of its parents; the other is the undirected graph model, also called a Markov field, where the joint distribution is factorized over the cliques! of an undirected graph. This factorization is called a junction tree and optimizing it is the subject of the present paper. The power of both models lies in their ability to display and exploit existent marginal and conditional independencies among subsets of variables. Emphasizing independencies is useful 1 A clique is a fully connected set of vertices and a maximal clique is a clique that is not contained in any other clique. 558 M. Meilii and M. /. Jordan from both a qualitative point of view (it reveals something about the domain under study) and a quantitative one (it makes computations tractable). The two models differ in the kinds of independencies they are able to represent and often times in their naturalness in particular tasks. Directed graphs are more convenient for learning a model from data; on the other hand, the clique structure of undirected graphs organizes the information in a way that makes it immediately available to inference algorithms. Therefore it is a standard procedure to construct the model of a domain as a Bayes net and then to convert it to a Markov field for the purpose of querying it. This process is known as decomposition and it consists of the following stages: first, the directed graph is transformed into an undirected graph by an operation called moralization. Second, the moralized graph is triangulated. A graph is called triangulated if any cycle of length> 3 has a chord (i.e. an edge connecting two nonconsecutive vertices). If a graph is not triangulated it is always possible to add new edges so that the resulting graph is triangulated. We shall call this procedure triangulation and the added edges the fill-in. In the final stage, the junction tree (Kjrerulff, 1991) is constructed from the maximal cliques of the triangulated graph. We define the state space of a clique to be the cartesian product of the state spaces of the variables associated to the vertices in the clique and we call weight of the clique the size of this state space. The weight of the junction tree is the sum of the weights of its component cliques. All further exact inference in the net takes place in the junction tree representation. The number of computations required by an inference operation is proportional to the weight of the tree. For each graph there are several and usually a large number of possible triangulations, with widely varying state space sizes. Moreover, triangulation is the only stage where the cost of inference can be influenced. It is therefore critical that the triangulation procedure produces a graph that is optimal or at least "good" in this respect. Unfortunately, this is a hard problem. No optimal triangulation algorithm is known to date. However, a number of heuristic algorithms like maximum cardinality search (Tarjan and Yannakakis, 1984), lexicographic search (Rose et al., 1976) and the minimum weight heuristic (MW) (Kjrerulff, 1990) are known. An optimization method based on simulated annealing which performs better than the heuristics on large graphs has been proposed in (Kjrerulff, 1991) and recently a "divide and conquer" algorithm which bounds the maximum clique size of the triangulated graph has been published (Becker and Geiger, 1996). All but the last algorithm are based on Rose's (Rose, 1970) elimination procedure: choose a node v of the graph, connect all its neighbors to form a clique, then eliminate v and all the edges incident to it and proceed recursively. The resulting filled-in graph is triangulated. It can be proven that the optimal triangulation can always be obtained by applying Rose's elimination procedure with an appropriate ordering of the nodes. It follows then that searching for an optimal triangulation can be cast as a search in the space of all node permutations. The idea of the present work is the following: embed the discrete search space of permutations of n objects (where n is the number of vertices) into a suitably chosen continuous space. Then extend the cost to a smooth function over the continuous domain and thus transform the discrete optimization problem into a continuous nonlinear optimization task. This allows one to take advantage of the thesaurus of optimization methods that exist for continuous cost functions. The rest of the paper will present this procedure in the following sequence: the next section introduces and discusses the objective function; section 3 states the continuous version of the problem; section 4 discusses further aspects of the optimization procedure and presents experimental results and section 5 concludes Triangulation by Continuous Embedding 559 the paper. 2 THE OBJECTIVE In this section we introduce the objective function that we used and we discuss its relationship to the junction tree weight. First, some notation. Let G = (V, E) be a graph, its vertex set and its edge set respectively. Denote by n the cardinality of the vertex set, by ru the number of values of the (discrete) variable associated to vertex v E V, by # the elimination ordering of the nodes, such that #v = i means that node v is the i-th node to be eliminated according to ordering #, by n(v) the set of neighbors of v E V in the triangulated graph and by Cu = {v} U {u E n( v) I #u > #v}. 2 Then, a result in (Golumbic, 1980) allows us to express the total weight of the junction tree obtained with elimination ordering # as (1) where ismax(Cu ) is a variable which is 1 when Cu is a maximal clique and 0 otherwise. As stated, this is the objective of interest for belief net triangulation. Any reference to optimality henceforth will be made with respect to J* . This result implies that there are no more than n maximal cliques in a junction tree and provides a method to enumerate them. This suggests defining a cost function that we call the raw weight J as the sum over all the cliques Cu (thus possibly including some non-maximal cliques): J(#) = I: II ru (2) uEV uECv J is the cost function that will be used throughout this paper. A reason to use it instead of J* in our algorithm is that the former is easier to compute and to approximate. How to do this will be the object of the next section. But it is natural to ask first how well do the two agree? Obviously, J is an upper bound for J*. Moreover, it can be proved that if r = min ru (3) and therefore that J is less than a fraction 1/(r - 1) away from J* . The upper bound is attained when the triangulated graph is fully connected and all ru are equal. In other words, the differece between J and J* is largest for the highest cost triangulation. We also expect this difference to be low for the low cost triangulation. An intuitive argument for this is that good triangulations are associated with a large number of smaller cliques rather than with a few large ones. But the former situation means that there will be only a small number of small size non-maximal cliques to contribute to the difference J - J* , and therefore that the agreement with J* is usually closer than (3) implies. This conclusion is supported by simulations (Meila. and Jordan, 1997). 2Both n(v) and CO) depend on # but we chose not to emphasize this in the notation for the sake of readability. 560 M. Meilii and M. I. Jordan 3 THE CONTINUOUS OPTIMIZATION PROBLEM This section shows two ways of defining J over continuous domains. Both rely on a formulation of J that eliminates explicit reference to the cliques Gu ; we describe this formulation here. Let us first define new variables J.tUIl and eUIl , U, v = 1, .. , n. For any permutation # J.tuu { 1 if #u ~ #v o otherwise { 1 if the edge (u,v) E EUF# o otherwise where F # is the set of fill-in edges. In other words, J.t represent precedence relationships and e represent the edges between the n vertices. Therefore, they will be called precedence variables and edge variables respectively. With these variables, J can be expressed as J(#) = I: IT r~vuevu (4) uEV uEV In (4), the product J.tuueuu acts as an indicator variable being 1 iff "u E Gil" is true. For any given permutation, finding the J.t variables is straightforward. Computing the edge variables is possible thanks to a result in (Rose et al., 1976). It states that an edge (u, v) is contained in F# iff there is a path in G between u and v containing only nodes w for which #w < mine #u, #v). Formally, eUIl = euu = 1 iff there exists a path P = (u, WI, W2, ... v) such that IT J.tw,uJ.tw,u = 1 WoEP So far, we have succeeded in defining the cost J associated with any permutation in terms of the variables J.t and e. In the following, the set of permutations will be embedded in a continuous domain. As a consequence, J.t and e will take values in the interval [0,1] but the form of J in (4) will stay the same. The first method, called J.t-continuous embedding (J.t-CE) assumes that the variables J.tuu E [0,1] represent independent probabilities that #u < #v. For any permutation, the precedence variables have to satisfy the transitivity condition. Transitivity means that if #u < #v and #v < #w, then #u < #w, or, that for any triple (J.tuu, J.tIlW, J.twu) the assignments (0, 0,0) and (1,1,1) are forbidden. According to the probabilistic interpretation of J.t we introduce a term that penalizes the probability of a transitivity violation: L P[(u, v, w) nontransitive] (5) U<u<W I: [J.tUUJ.tIlWJ.tWU + (1 - J.tuu)(l - J.tuw)(l- J.twu)] (6) U<Il<W > P[assignment non transitive] (7) In the second approach, called O-continuous embedding (O-CE), the permutations are directly embedded into the set of doubly stochastic matrices. A doubly stochastic matrix () is a matrix for which the elements in a row or column sum to one. I:0ij I:0ij = 1 Oij ~ 0 for i,j = 1, .. n. (8) j Triangulation by Continuous Embedding 561 When Oij are either 0 or 1, implying that there is exactly one nonzero element in each row or column, the matrix is called a permutation matrix. Oij = 1 and #i = j both mean that the position of object i is j in the given permutation. The set of doubly stochastic matrices e is a convex polytope of dimension (n - 1)2 whose extreme points are the permutation matrices (Balinski and Russakoff, 1974). Thus, every doubly stochastic matrix can be represented as a convex combination of permutation matrices. To constrain the optimum to be a an extreme point, we use the penalty term R(O) = I: Oij (1 - Oij) ij The precedence variables are defined over e as J.luv 1 - J.lvu 1 (9) Now, for both embeddings, the edge variables can be computed from J.l as follows { I max ITwEP J.lwuJ.lwv PE {path& u-v} for (u, v) E E or u = v otherwise The above assignments give the correct values for J.l and e for any point representing a permutation. Over the interior of the domain, e is a continuous, piecewise differentiable function. Each euv , (u, v) ftE can be computed by a shortest path algorithm between u and v, with the length of (WI,W2) E E defined as (-logJ.lwluJ.lw:>v). O-CE is an interior point method whereas in J.l-CE the current point, although inside [0,I]n(n-I)/2, isn't necessarily in the convex hull of the hypercube's corners that represent permutations. The number of operation required for one evaluation of J and its gradient is as follows: O(n4) operations to compute J.l from 0, O(n3 10gn) to compute e, O(n3 ) for ~: and O(n2 ) for ~~ and ~~ afterwards. Since computing J.l is the most computationally intensive step, J.l-CE is a clear win in terms of computation cost. In addition, by operating directly in the J.l domain, one level of approximation is eliminated, which makes one expect J.l-CE to perform better than O-CE. The results in the following section will confirm this. 4 EXPERIMENTAL RESULTS To assess the performance of our algorithms we compared their results with the results of the minimum weight heuristic (MW), the heuristic that scored best in empirical tests (Kjrerulff, 1990). The lowest junction tree weight obtained in 200 runs of MW was retained and denoted by JMW' Tests were run on 6 graphs of different sizes and densities: graph h9 h12 dlO m20 a20 d20 n= IVI 9 12 10 20 20 20 density .33 .25 .6 .25 .45 .6 r min/r max/r avf1, 2/2/2/ 3/3/3 6/15/10 2/8/5 6/15/10 6/15/10 10giO JMW 2.43 2.71 7.44 5.47 12.75 13.94 The last row of the table shows the 10giO JMW' We ran 11 or more trials of each of our two algorithms on each graph. To enforce the variables to converge to a permutation, we minimized the objective J + >.R, where>. > 0 is a parameter 562 100 30 10 3 0.3 -=........- h9 h12 d10 m20 820 d20 a 20 10 5 2 .5 .2 M. Meilii and M. l Jordan .1~~--~--~------~--~----~ h9 h12 d10 m20 a20 d20 b Figure 1: Minimum, maximum (solid line) and median (dashed line) values of J1· MW obtained by O-CE (a) and JL-CE (b). that was progressively increased following a deterministic annealing schedule and R is one of the aforementioned penalty terms. The algorithms were run for 50150 optimization cycles, usually enough to reach convergence. However, for the JL-embedding on graph d20, there were several cases where many JL values did not converge to 0 or 1. In those cases we picked the most plausible permutation to be the answer. The results are shown in figure 1 in terms of the ratio of the true cost obtained by the continuous embedding algorithm (denoted by J*) and J'Mw. For the first two graphs, h9 and h12, J1-w is the optimal cost; the embedding algorithms reach it most trials. On the remaining graphs, JL-CE clearly outperforms O-CE, which also performs poorer than MW on average. On dIO, a20 and m20 it also outperforms the MW heuristic, attaining junction tree weights that are 1.6 to 5 times lower on average than those obtained by MW. On d20, a denser graph, the results are similar for MW and JL-CE in half of the cases and worse for JL-CE otherwise. The plots also show that the variability of the results is much larger for CE than for MW. This behaviour is not surprising, given that the search space for CE, although continuous, comprises a large number of local minima. This induces dependence on the initial point and, as a consequence, nondeterministic behaviour of the algorithm. Moreover, while the number of choices that MW has is much lower than the upper limit of n!, the "choices" that CE algorithms consider, although soft, span the space of all possible permutations. 5 CONCLUSION The idea of continuous embedding is not new in the field of applied mathematics. The large body of literature dealing with smooth (sygmoidal) functions instead of hard nonlinearities (step functions) is only one example. The present paper shows a nontrivial way of applying a similar treatment to a new problem in a new field. The results obtained by it-embedding are on average better than the standard MW heuristic. Although not directly comparable, the best results reported on triangulation (Kjrerulff, 1991; Becker and Geiger, 1996) are only by little better than ours. Therefore the significance of the latter goes beyond the scope of the present problem. They are obtained on a hard problem, whose cost function has no feature to ease its minimization (J is neither linear, nor quadratic, nor is it additive Triangulation by Continuous Embedding 563 w.r.t. the vertices or the edges) and therefore they demonstrate the potential of continuous embedding as a general tool. Colaterally, we have introduced the cost function J, which is directly amenable to continuous approximations and is in good agreement with the true cost r. Since minimizing J may not be NP-hard, this opens a way for investigating new triangulation methods. Acknowledgements The authors are grateful to Tommi Jaakkola for many discussions and to Ellie Bonsaint for her invaluable help in typing the paper. References Balinski, M. and Russakoff, R. (1974). On the assignment polytope. SIAM Rev. Becker, A. and Geiger, D. (1996). A sufficiently fast algorithm for finding close to optimal junction trees. In UAI 96 Proceedings. Golumbic, M. (1980). Algorithmic Graph Theory and Perfect Graphs. Academic Press, New York. Kjrerulff, U. (1990). Triangulation of graphs-algorithms giving small total state space. Technical Report R 90-09, Department of Mathematics and Computer Science, Aalborg University, Denmark. Kjrerulff, U. (1991). Optimal decomposition of probabilistic networks by simulated annealing. Statistics and Computing. Meila., M. and Jordan, M. I. (1997). An objective function for belief net triangulation. In Madigan, D., editor, AI and Statistics, number 7. (to appear). Rose, D. J. (1970). Triangulated graphs and the elimination process. Journal of Mathematical Analysis and Applications. Rose, D. J., Tarjan, R. E., and Lueker, E. (1976). Algorithmic aspects of vertex elimination on graphs. SIAM J. Comput. Tarjan, R. and Yannakakis, M. (1984). Simple linear-time algorithms to test chordality of graphs, test acyclicity of hypergraphs, and select reduced acyclic hypergraphs. SIAM 1. Comput.
1996
67
1,266
Bayesian Model Comparison by Monte Carlo Chaining David Barber D.Barber~aston.ac.uk Christopher M. Bishop C.M.Bishop~aston.ac.uk Neural Computing Research Group Aston University, Birmingham, B4 7ET, U.K. http://www.ncrg.aston.ac.uk/ Abstract The techniques of Bayesian inference have been applied with great success to many problems in neural computing including evaluation of regression functions, determination of error bars on predictions, and the treatment of hyper-parameters. However, the problem of model comparison is a much more challenging one for which current techniques have significant limitations. In this paper we show how an extended form of Markov chain Monte Carlo, called chaining, is able to provide effective estimates of the relative probabilities of different models. We present results from the robot arm problem and compare them with the corresponding results obtained using the standard Gaussian approximation framework. 1 Bayesian Model Comparison In a Bayesian treatment of statistical inference, our state of knowledge of the values of the parameters w in a model M is described in terms of a probability distribution function. Initially this is chosen to be some prior distribution p(wIM), which can be combined with a likelihood function p( Dlw, M) using Bayes' theorem to give a posterior distribution p(wID, M) in the form ( ID M) = p(Dlw,M)p(wIM) p w , p(DIM) (1) where D is the data set. Predictions of the model are obtained by performing integrations weighted by the posterior distribution. 334 D. Barber and C. M. Bishop The comparison of different models Mi is based on their relative probabilities, which can be expressed, again using Bayes' theorem, in terms of prior probabilities P(Mi) to give P(MiID ) p(MjID) p(DIMdp(Mi) p(DIMj )p(Mj) (2) and so requires that we be able to evaluate the model evidence p(DIMi), which corresponds to the denominator in (1). The relative probabilities of different models can be used to select the single most probable model, or to form a committee of models, weighed by their probabilities. It is convenient to write the numerator of (1) in the form exp{ -E(w)}, where E(w) is an error function. Normalization of the posterior distribution then requires that p(DIM) = J exp{ -E(w)} dw. (3) Generally, it is straightforward to evaluate E(w) for a given value of w, although it is extremely difficult to evaluate the corresponding model evidence using (3) since the posterior distribution is typically very small except in narrow regions of the high-dimensional parameter space, which are unknown a-priori. Standard numerical integration techniques are therefore inapplicable. One approach is based on a local Gaussian approximation around a mode of the posterior (MacKay, 1992). Unfortunately, this approximation is expected to be accurate only when the number of data points is large in relation to the number of parameters in the model. In fact it is for relatively complex models, or problems for which data is scarce, that Bayesian methods have the most to offer. Indeed, Neal (1996) has argued that, from a Bayesian perspective, there is no reason to limit the number of parameters in a model, other than for computational reasons. We therefore consider an approach to the evaluation of model evidence which overcomes the limitations of the Gaussian framework. For additional techniques and references to Bayesian model comparison, see Gilks et al. (1995) and Kass and Raftery (1995). 2 Chaining Suppose we have a simple model Mo for which we can evaluate the evidence analytically, and for which we can easily generate a sample wi (where I = 1, ... , L) from the corresponding distribution p(wID, Mo). Then the evidence for some other model M can be expressed in the form p(DIM) p(DIMo) J exp{-E(w) + Eo(w)}p(wID, Mo)dw 1 L L :E exp{ -E(w1) + Eo(w1)}. 1=1 (4) Unfortunately, the Monte Carlo approximation in (4) will be poor if the two error functions are significantly different, since the exponent is dominated by regions where E is relatively small, for which there will be few samples unless Eo is also small in those regions. A simple Monte Carlo approach will therefore yield poor results. This problem is equivalent to the evaluation of free energies in statistical physics, Bayesian Model Comparison by Monte Carlo Chaining 335 which is known to be a challenging problem, and where a number of approaches have been developed Neal (1993). Here we discuss one such approach to this problem based on a chain of J{ successive models Mi which interpolate between Mo and M, so that the required evidence can be written as p(DIMl) p(DIM2) p(DIM) p(DIM) = p(DIMo) p(DIMo) p(DIMt} ... p(DIMK)· (5) Each of the ratios in (5) can be evaluated using (4). The goal is to devise a chain of models such that each successive pair of models has probability distributions which are reasonably close, so that each of the ratios in (5) can be evaluated accurately, while keeping the total number of links in the chain fairly small to limit the computational costs. We have chosen the technique of hybrid Monte Carlo (Duane et ai., 1987; Neal, 1993) to sample from the various distributions, since this has been shown to be effective for sampling from the complex distributions arising with neural network models (Neal, 1996). This involves introducing Hamiltonian equations of motion in which the parameters ware augmented by a set of fictitious 'momentum' variables, which are then integrated using the leapfrog method. At the end of each trajectory the new parameter vector is accepted with a probability governed by the Metropolis criterion, and the momenta are replaced using Gibbs sampling. As a check on our software implementation of chaining, we have evaluated the evidence for a mixture of two non-isotropic Gaussian distributions, and obtained a result which was within 10% of the analytical solution. 3 Application to Neural Networks We now consider the application of the chaining method to regression problems involving neural network models. The network corresponds to a function y(x, w), and the data set consists of N pairs of input vectors Xn and corresponding targets tn where n = 1, ... , N. Assuming Gaussian noise on the target data, the likelihood function takes the form ( f3)N/2 {f3 N } p(Dlw, M) = 211" exp -2" ~ Ily(xn ; w) - t n l1 2 (6) where f3 is a hyper-parameter representing the inverse of the noise variance. We consider networks with a single hidden layer of 'tanh' units, and linear output units. Following Neal (1996) we use a diagonal Gaussian prior in which the weights are divided into groups Wk, where k = 1, ... ,4 corresponding to input-to-hidden weights, hidden-unit biases, hidden-to-output weights, and output biases. Each group is governed by a separate 'precision' hyper-parameter ak, so that the prior takes the form p(wl{a,}) = L exp { -~ ~ a,wfw, } (7) where Zw is the normalization coefficient. The hyper-parameters {ad and f3 are themselves each governed by hyper-priors given by Gamma distributions of the form p( a) ex: a$ exp( -as /2w) (8) 336 D. Barber and C. M. Bishop in which the mean wand variance 2w2 / s are chosen to give very broad hyper-priors in reflection of our limited prior knowledge of the values of the hyper-parameters. We use the hybrid Monte Carlo algorithm to sample from the joint distribution of parameters and hyper-parameters. For the evaluation of evidence ratios, however, we consider only the parameter samples, and perform the integrals over hyperparameters analytically, using the fact that the gamma distribution is conjugate to the Gaussian. In order to apply chaining to this problem, we choose the prior as our reference distribution, and then define a set of intermediate distributions based on a parameter A which governs the effective contribution from the data term, so that E(A, w) = A</>(W) + Eo(w) (9) where </>(w) arises from the likelihood term (6) while Eo(w) corresponds to the prior (7). We select a set of 18 values of A which interpolate between the reference distribution (A = 0) and the desired model distribution (A = 1). The evidence for the prior alone is easily evaluated analytically. 4 Gaussian Approximation As a comparison against the method of chaining, we consider the framework of MacKay (1992) based on a local Gaussian approximation to the posterior distribution. This approach makes use of the evidence approximation in which the integration over hyper-parameters is approximated by setting them to specific values which are themselves determined by maximizing their evidence functions. This leads to a hierarchical treatment as follows. At the lowest level, the maximum w of the posterior distribution over weights is found for fixed values of the hyperparameters by minimizing the error function. Periodically the hyper-parameters are re-estimated by evidence maximization, where the evidence is obtained analytically using the Gaussian approximation. This gives the following re-estimation formulae 1 f3 (10) where 'Yk = Wk - Uk Trk(A -1), Wk is the total number of parameters in group k, A = \7\7 E(w), 'Y = Lk 'Yk. and Trk(-) denotes the trace over the kth group of parameters. The weights are updated in an inner loop by minimizing the error function using a conjugate gradient optimizer, while the hyper-parameters are periodically re-estimated using (10)1. Once training is complete, the model evidence is evaluated by making a Gaussian approximation around the converged values of the hyper-parameters, and integrating over this distribution analytically. This gives the model log evidence as Inp(DIM) = -E(w) ~ In IAI + ~ L Wk lnuk + k N 1 1 2Inf3+lnh!+2Inh+ 2 ~ln(2hk) + 2 In (2/(N -'Y». (11) 1 Note that we are assuming that the hyper-priors (8) are sufficiently broad that they have no effect on the location of the evidence maximum and can therefore be neglected. Bayesian Model Comparison by Monte Carlo Chaining 337 Here h is the number of hidden units, and the terms In h! + 2ln h take account of the many equivalent modes of the posterior distribution arising from sign-flip and hidden unit interchange symmetries in the network model. A derivation of these results can be found in Bishop (1995; pages 434-436). The result (11) corresponds to a single mode of the distribution. If we initialize the weight optimization algorithm with different random values we can find distinct solutions. In order to compute an overall evidence for the particular network model with a given number of hidden units, we make the assumption that we have found all of the distinct modes of the posterior distribution precisely once each, and then sum the evidences to arrive at the total model evidence. This neglects the possibility that some of the solutions found are related by symmetry transformations (and therefore already taken into account) or that we have missed important modes. While some attempt could be made to detect degenerate solutions, it will be difficult to do much better than the above within the framework of the Gaussian approximation. 5 Results: Robot Arm Problem As an illustration of the evaluation of model evidence for a larger-scale problem we consider the modelling of the forward kinematics for a two-link robot arm in a two-dimensional space, as introduced by MacKay (1992). This problem was chosen as MacKay reports good results in using the Gaussian approximation framework to evaluate the evidences, and provides a good opportunity for comparison with the chaining approach. The task is to learn the mapping (Xl, X2) -+ (Yl, Y2) given by where the data set consists of 200 input-output pairs with outputs corrupted by zero mean Gaussian noise with standard deviation u = 0.05. We have used the original training data of MacKay, but generated our own test set of 1000 points using the same prescription. The evidence is evaluated using both chaining and the Gaussian approximation, for networks with various numbers of hidden units. In the chaining method, the particular form of the gamma priors for the precision variables are as follows: for the input-to-hidden weights and hidden-unit biases, w = 1, s = 0.2; for the hidden-to-output weights, w = h, s = 0.2; for the output biases, w = 0.2, s = 1. The noise level hyper-parameters were w = 400, s = 0.2. These settings follow closely those used by Neal (1996) for the same problem. The hidden-to-output precision scaling was chosen by Neal such that the limit of an infinite number of hidden units is well defined and corresponds to a Gaussian process prior. For each evidence ratio in the chain, the first 100 samples from the hybrid Monte Carlo run, obtained with a trajectory length of 50 leapfrog iterations, are omitted to give the algorithm a chance to reach the equilibrium distribution. The next 600 samples are obtained using a trajectory length of 300 and are used to evaluate the evidence ratio. In Figure 1 (a) we show the error values of the sampling stage for 24 hidden units, where we see that the errors are largely uncorrelated, as required for effective Monte Carlo sampling. In Figure 1 (b), we plot the values of In{p(DIMi)/p(DIMi_l)} against .Ai i = 1..18. Note that there is a large change in the evidence ratios at the beginning of the chain, where we sample close to the reference distribution. For this 338 D. Barber and C. M. Bishop (a) 6 (b) -4~--~----~--~----~--~ o 100 200 300 400 500 600 o 0.2 0.4 0.6 0.8 A 1 Figure 1: (a) error E(>. = 0.6,w) for h = 24, plotted for 600 successive Monte Carlo samples. (b) Values of the ratio In{p(DIM.)jp(DIM.-d} for i = 1, ... ,18 for h = 24. reason, we choose the Ai to be dense close to A = O. We are currently researching more principled approaches to the partitioning selection. Figure 2 (a) shows the log model evidence against the number of hidden units. Note that the chaining approach is computationally expensive: for h=24, a complete chain takes 48 hours in a Matlab implementation running on a Silicon Graphics Challenge L. We see that there is no decline in the evidence as the number of hidden units grows. Correspondingly, in Figure 2 (b), we see that the test error performance does not degrade as the number of hidden units increases. This indicates that there is no over-fitting with increasing model complexity, in accordance with Bayesian expectations. The corresponding results from the Gaussian approximation approach are shown in Figure 3. We see that there is a characteristic 'Occam hill' whereby the evidence shows a peak at around h = 12, with a strong decrease for smaller values of h and a slower decrease for larger values. The corresponding test set errors similarly show a minimum at around h = 12, indicating that the Gaussian approximation is becoming increasingly inaccurate for more complex models. 6 Discussion We have seen that the use of chaining allows the effective evaluation of model evidences for neural networks using Monte Carlo techniques. In particular, we find that there is no peak in the model evidence, or the corresponding test set error, as the number of hidden units is increased, and so there is no indication of overfitting. This is in accord with the expectation that model complexity should not be limited by the size of the data set, and is in marked contrast to the conventional 70.-----~----~------~----~ 1.4r-----~-----,-------~----~ 50 (a) 60 40 1.3~ 1.2' ~ 1.1 1!r------&----~ (b) 30L-----1~0------1~5------~20~--~ h 10 15 20 h Figure 2: (a) Plot of Inp(DIM) for different numbers of hidden units. (b) Test error against the number of hidden units. Here the theoretical minimum value is 1.0. For h = 64 the test error is 1.11 Bayesian Model Comparison by Monte Carlo Chaining 85or-----------~----~----~ 800 o o o o 00 0 00 0 (a) 2.5 0 o o 2 0 @ o 1.5 339 (b) o o o 7505L-----l~0----~15------2O~----2......J5 h 1L-----~~--~----~----~ 5 10 15 20 25 h Figure 3: (a) Plot of the model evidence for the robot arm problem versus the number of hidden units, using the Gaussian approximation framework. This clearly shows the characteristic 'Occam hill' shape. Note that the evidence is computed up to an additive constant, and so the origin of the vertical axis has no significance. (b) Corresponding plot of the test set error versus the number of hidden units. Individual points correspond to particular modes of the posterior weight distribution, while the line shows the mean test set error for each value of h. maximum likelihood viewpoint. It is also consistent with the result that, in the limit of an infinite number of hidden units, the prior over network weights leads to a well-defined Gaussian prior over functions (Williams, 1997). An important advantage of being able to make accurate evaluations of the model evidence is the ability to compare quite distinct kinds of model, for example radial basis function networks and multi-layer perceptIOns. This can be done either by chaining both models back to a common reference model, or by evaluating normalized model evidences explicitly. Acknowledgements We would like to thank Chris Williams and Alastair Bruce for a number of useful discussions. This work was supported by EPSRC grant GR/J75425: Novel Developments in Learning Theory for Neural Networks. References Bishop, C. M. (1995). Neural Networks for Pattern Recognition. Oxford University Press. Duane, S., A. D. Kennedy, B. J. Pendleton, and D. Roweth (1987). Hybrid Monte Carlo. Physics Letters B 195 (2), 216-222. Gilks, W. R., S. Richardson, and D. J. Spiegelhalter (1995). Markov Chain Monte Carlo in Practice. Chapman and Hall. Kass, R. E. and A. E. Raftery (1995). Bayes factors. J. Am. Statist. Ass. 90, 773-795. MacKay, D. J. C. (1992). A practical Bayesian framework for back-propagation networks. Neural Computation 4 (3), 448- 472. Neal, R. M. (1993). Probabilistic inference using Markov chain Monte Carlo methods. Technical Report CRG-TR-93-1, Department of Computer Science, University of Toronto, Cananda. Neal, R. M. (1996). Bayesian Learning for Neural Networks. Springer. Lecture Notes in Statistics 118. Williams, C. K. I. (1997). Computing with infinite networks. This volume.
1996
68
1,267
Practical confidence and prediction intervals Tom Heskes RWCP Novel Functions SNN Laboratory; University of Nijmegen Geert Grooteplein 21, 6525 EZ Nijmegen, The Netherlands tom@mbfys.kun.nl Abstract We propose a new method to compute prediction intervals. Especially for small data sets the width of a prediction interval does not only depend on the variance of the target distribution, but also on the accuracy of our estimator of the mean of the target, i.e., on the width of the confidence interval. The confidence interval follows from the variation in an ensemble of neural networks, each of them trained and stopped on bootstrap replicates of the original data set. A second improvement is the use of the residuals on validation patterns instead of on training patterns for estimation of the variance of the target distribution. As illustrated on a synthetic example, our method is better than existing methods with regard to extrapolation and interpolation in data regimes with a limited amount of data, and yields prediction intervals which actual confidence levels are closer to the desired confidence levels. 1 STATISTICAL INTERVALS In this paper we will consider feedforward neural networks for regression tasks: estimating an underlying mathematical function between input and output variables based on a finite number of data points possibly corrupted by noise. We are given a set of Pdata pairs {ifJ, tfJ } which are assumed to be generated according to t(i) = f(i) + e(i) , (1) where e(i) denotes noise with zero mean. Straightforwardly trained on such a regression task, the output of a network o(i) given a new input vector i can be RWCP: Real World Computing Partnership; SNN: Foundation for Neural Networks. Practical Confidence and Prediction Intervals 177 interpreted as an estimate of the regression f(i) , i.e., of the mean of the target distribution given input i. Sometimes this is all we are interested in: a reliable estimate of the regression f(i). In many applications, however, it is important to quantify the accuracy of our statements. For regression problems we can distinguish two different aspects: the accuracy of our estimate of the true regression and the accuracy of our estimate with respect to the observed output. Confidence intervals deal with the first aspect, i.e., consider the distribution of the quantity f(i) - o(i), prediction intervals with the latter, i.e., treat the quantity t(i) - o(i). We see from t(i) - o(i) = [f(i) - o(i)] + ~(i) , (2) that a prediction interval necessarily encloses the corresponding confidence interval. In [7] a method somewhat similar to ours is introduced to estimate both the mean and the variance of the target probability distribution. It is based on the assumption that there is a sufficiently large data set, i.e., that their is no risk of overfitting and that the neural network finds the correct regression. In practical applications with limited data sets such assumptions are too strict. In this paper we will propose a new method which estimates the inaccuracy of the estimator through bootstrap resampling and corrects for the tendency to overfit by considering the residuals on validation patterns rather than those on training patterns. 2 BOOTSTRAPPING AND EARLY STOPPING Bootstrapping [3] is based on the idea that the available data set is nothing but a particular realization of some unknown probability distribution. Instead of sampling over the "true" probability distribution, which is obviously impossible, one defines an empirical distribution. With so-called naive bootstrapping the empirical distribution is a sum of delta peaks on the available data points, each with probability content l/Pdata. A bootstrap sample is a collection of Pdata patterns drawn with replacement from this empirical probability distribution. This bootstrap sample is nothing but our training set and all patterns that do not occur in the training set are by definition part of the validation set. For large Pdata, the probability that a pattern becomes part of the validation set is (1 l/Pdata)Pdata ~ lie ~ 0.37. When training a neural network on a particular bootstrap sample, the weights are adjusted in order to minimize the error on the training data. Training is stopped when the error on the validation data starts to increase. This so-called early stopping procedure is a popular strategy to prevent overfitting in neural networks and can be viewed as an alternative to regularization techniques such as weight decay. In this context bootstrapping is just a procedure to generate subdivisions in training and validation set similar to k-fold cross-validation or subsampling. On each of the nrun bootstrap replicates we train and stop a single neural network. The output of network i on input vector i IJ is written oi(ilJ ) == or. As "the" estimate of our ensemble of networks for the regression f(i) we take the average outputl 1 nrun m(i) == L.: oi(i). n run i=l lThis is a so-called "bagged" estimator [2]. In [5] it is shown that a proper balancing of the network outputs can yield even better results. 178 T. Heskes 3 CONFIDENCE INTERVALS Confidence intervals provide a way to quantify our confidence in the estimate m(i) of the regression f(i), i.e., we have to consider the probability distribution P(f(x)lm(i)) that the true regression is f(x) given our estimate m(i). Our line of reasoning goes as follows (see also [8]). We assume that our ensemble of neural networks yields a more or less unbiased estimate for f(x), i.e., the distribution P(f(i)lm(i)) is centered around m(i). The truth is that neural networks are biased estimators. For example, neural networks trained on a finite number of examples will always have a tendency (as almost any other model) to oversmooth a sharp peak in the data. This introduces a bias, which, to arrive at asymptotically correct confidence intervals, should be taken into account. However, if it would be possible to compute such a bias correction, one should do it in the first place to arrive at a better estimator. Our working hypothesis here is that the bias component of the confidence intervals is negligible in comparison with the variance component. There do exist methods that claim to give confidence intervals that are "secondorder correct" , i.e., up to and including terms of order l/P!~~a (see e.g. the discussion after [3]). Since we do not know how to handle the bias component anyways, such precise confidence intervals, which require a tremendous amount of bootstrap samples, are too ambitious for our purposes. First-order correct intervals up to and including terms of order l/Pdata are always symmetric and can be derived by assuming a Gaussian distribution P(f(i)lm(i)). The variance of this distribution can be estimated from the variance in the outputs of the nrun networks: 1 nrun ----1 2: [Oi(X) - m(iW· n run i=l (3) This is the crux of the bootstrap method (see e.g. [3]). Since the distribution of P(f(i)lm(i)) is a Gaussian, so is the "inverse" distribution P(m(i)lf(i)) to find the regression m( i) by randomly drawing data sets consisting of Pdata data points according to the prescription (1). Not knowing the true distribution of inputs and corresponding targets2 , the best we can do is to define the empirical distribution as explained before and estimate P(m(i)lf(i)) from the distribution P(o(i)lm(i)). This then yields the estimate (3). So, following this bootstrap procedure we arrive at the confidence intervals m(i) - Cconfidenceu(i) ~ f(i) ~ m(i) + Cconfidenceu(i) , where Cconfidence depends on the desired confidence level 1-o'. The factors Cconfidence can be taken from a table with the percentage points of the Student's t-distribution with number of degrees of freedom equal to the number of bootstrap runs nrun . A more direct alternative is to choose Cconfidence such that for no more than 1000'% of all nrun x Pdata network predictions lor - m'" I 2: Cconfidence u"'. 2In this paper we assume that both the inputs and the outputs are stochastic. For the case of deterministic input variables other bootstrapping techniques (see e.g. [4]) are more appropriate, since the statistical intervals resulting from naive bootstrapping may be too conservative. Practical Confidence and Prediction Intervals 179 4 PREDICTION INTERVALS Confidence intervals deal with the accuracy of our prediction of the regression, Le., of the mean of the target probability distribution. Prediction intervals consider the accuracy with which we can predict the targets themselves, i.e., they are based on estimates of the distribution P(t( x) Im(i)). We propose the following method. The two noise components f(i) - m(x) and ~(x) in (2) are independent. The variance of the first component has been estimated in our bootstrap procedure to arrive at confidence intervals. The remaining task is to estimate the noise inherent to the regression problem. We assume that this noise is more or less Gaussian such that it again suffices to compute its variance which may however depend on the input x. In mathematical symbols, Of course, we are interested in prediction intervals for new points x for which we do not know the targets t. Suppose that we had left aside a set of test patterns {Xli, til} that we had never used for training nor for validating our neural networks. Then we could try and estimate a model X2 (x) to fit the remaining residuals (4) using minus the loglikelihood as the error measure: (5) Of course, leaving out these test patterns is a waste of data and luckily our bootstrap procedure offers an alternative. Each pattern is in about 37% of all bootstrap runs not part of the training set. Let us write qf = 1 if pattern j.t is in the validation set of run i and qf = 0 otherwise. If we, for each pattern j.t, use the average nrun / nrun mvalidation (xJJ) = z= qf of z= qf , i=l JJ=l instead of the average m(xJJ) we get as close as possible to an unbiased estimate for the residual on independent test patterns as we can, without wasting any training data. So, summarizing, we suggest to find a function X(x) that minimizes the error (5), yet not by leaving out test patterns, which would be a waste of data, nor by straightforwardly using the training data, which would underestimate the error, but by exploiting the information about the residuals on the validation patterns. Once we have found the function X(i), we can compute for any x both the mean m(x) and the deviation s(x) which are combined in the prediction interval m(x) - CpredictionS(X) ~ t(x) ~ m(x) + CpredictionS(X) . Again, the factor Cprediction can be found in a Student's t-table or chosen such that for no more than 100a% of all Pdata patterns It JJ - mvalidation (xJJ) I ~ Cprediction s( xJJ) . The function X2(x) may be modelled by a separate neural network, similar to the method proposed in [7] with an exponential instead of a linear transfer function for the output unit to ensure that the variance is always positive. 180 "5 1.5 .-....----r----.-----,~-....___, 1 0.5 (a) .ea :::J o .s= -0.5 -1 -1.5 L--""'--_--'-__ _'_____--'-__ '---' -1 -0.5 a 0.5 input 0.2 ..--....---.-----.--~---....--n 0.15 \ (c) I I I I ~ 0.1 \ \ \ -'" 0.05 _ _ . __ . _ . _ . _ . _ . _ . I -1 -0.5 a 0.5 input T. Heskes 0.06 0.05 (b) Q) 0.04 (.) o c .~0.03 ..... to > 0.02 0.01 a -1 -0.5 a 0.5 1 input 1.5 1 (d) -1 _1.5L-----~-~--~-~.....J -1 -0.5 a 0.5 1 input Figure 1: Prediction intervals for a synthetic problem. (a) Training set (crosses), true regression (solid line), and network prediction (dashed line). (b) Validation residuals (crosses), training residuals (circles), true variance (solid line), estimated variance based on validation residuals (dashed line) and based on training residuals (dash-dotted line). (c) Width of standard error bars for the more advanced method (dashed line), the simpler procedure (dash-dotted line) and what it should be (solid line). (d) Prediction intervals (solid line) , network prediction (dashed line), and 1000 test points (dots) . 5 ILLUSTRATION We consider a synthetic problem similar to the one used in [7]. With this example we will demonstrate the desirability to incorporate the inaccuracy of the regression estimator in the prediction intervals. Inputs x are drawn from the interval [-1,1] with probability density p(x) = lxi, i.e., more examples are drawn at the boundary than in the middle. Targets t are generated according to t = sin(lI'x) cos(511'x/4) + ~(x) with (e(x)) = 0.005 + 0.005 [1 + sin(lI'x)]2 . The regression is the solid line in Figure 1 (a), the variance of the target distribution the solid line in Figure 1 (b). Following this prescription we obtain a training set of Pdata = 50 data points [the crosses in Figure l(a)] on which we train an ensemble of nrun = 25 networks, each having 8 hidden units with tanh-transfer function and one linear output unit. The average network output m(x) is the dashed line Practical Confidence and Prediction Intervals IBI in Figure l(a) and (d). In the following we compare two methods to arrive at prediction intervals: the more advanced method described in Section 4, Le., taking into account the uncertainty of the estimator and correcting for the tendency to overfit on the training data, and a simpler procedure similar to [7] which disregards both effects. We compute the (squared) "validation residuals" (mealidation t~)2 [crosses in Figure l(b)], based on runs in which pattern J.t was part of the validation set, and the "training residuals" (mrrain t~)2 (circles), based on runs in which pattern J.t was part of the training set. The validation residuals are most of the time somewhat larger than the training residuals. For our more advanced method we substract the uncertainty of our model from the validation residuals as in (4) . The other procedure simply keeps the training residuals to estimate the variance of the target distribution. It is obvious that the distribution of residuals in Figure l(b) does not allow for a complex model. Here we take a feedforward network with one hidden unit: The parameters {vo, VI, V2, V3} are found through minimization of the error (5). Both for the advanced method (dashed line) and for the simpler procedure (dashdotted line) the variance of the target distribution is estimated to be a step function. The former, being based on the validation residuals minus the uncertainty of the estimator, is slightly more conservative than the latter, being based on the training residuals. Both estimates are pretty far from the truth (solid line), especially for o < x < 0.5, yet considering such a limited amount of noisy residuals we can hardly expect anything better. Figure l(c) considers the width of standard error bars, i.e., of prediction intervals for error level a ~ 0.32. For the simpler procedure the width of the prediction interval [dash-dotted line in Figure l(c)] follows directly from the estimate of the variance of the target distribution. Our more advanced method adds the uncertainty of the estimator to arrive at the dashed line. The correct width of the prediction interval, i.e., the width that would include 68% of all targets for a particular input, is given by the solid line. The prediction intervals obtained through the more advanced procedure are displayed in Figure l(d) together with a set of 1000 test points visualizing the probability distribution of inputs and corresponding targets. The method proposed in Section 4 has several advantages. The prediction intervals of the advanced method include 65% of the test points in Figure l(d), pretty close to the desired confidence level of 68%. The simpler procedure is too liberal with an actual confidence level of only 58%. This difference is mainly due to the use of validation residuals instead of training residuals. Incorporation of the uncertainty of the estimator is important in regions of input space with just a few training data. In this example the density of training data affects both extrapolation and interpolation. For Ixl > 1 the prediction intervals obtained with the advanced method become wider and wider whereas those obtained through the simpler procedure remain more or less constant. The bump in the prediction interval (dashed line) near the origin is a result of the relatively large variance in the network predictions in this region. It shows that our method also incorporates the effect that the density of training data has on the accuracy of interpolation. 182 T. Heskes 6 CONCLUSION AND DISCUSSION We have presented a novel method to compute prediction intervals for applications with a limited amount of data. The uncertainty of the estimator itself has been taken into account by the computation of the confidence intervals. This explains the qualitative improvement over existing methods in regimes with a low density of training data. Usage of the residuals on validation instead of on training patterns yields prediction intervals with a better coverage. The price we have to pay is in the computation time: we have to train an ensemble of networks on about 20 to 50 different bootstrap replicates [3, 8]. There are other good reasons for resampling: averaging over networks improves the generalization performance and early stopping is a natural strategy to prevent overfitting. It would be interesting to see how our ''frequentist'' method compares with Bayesian alternatives (see e.g. [1, 6]). Prediction intervals can also be used for the detection of outliers. With regard to the training set it is straightforward to point out the targets that are not enclosed by a prediction interval of error level say a = 0.05. A wide prediction interval for a new test pattern indicates that this test pattern lies in a region of input space with a low density of training data making any prediction completely unreliable. A weak point in our method is the assumption of unbiased ness in the computation of the confidence intervals. This assumption makes the confidence intervals in general too liberal. However, as discussed in [8], such bootstrap methods tend to perform better than other alternatives based on the computation of the Hessian matrix, partly because they incorporate the variability due to the random initialization. Furthermore, when we model the prediction interval as a function of the input x we will, to some extent, repair this deficiency. But still, incorporating even a somewhat inaccurate confidence interval ensures that we can never severely overestimate our accuracy in regions of input space where we have never been before. References [1] C. Bishop and C. Qazaz. Regression with input-dependent noise: a Bayesian treatment. These proceedings, 1997. [2] L. Breiman. Bagging predictors. Machine Learning, 24: 123-140, 1996. [3] B. Efron and R. Tibshirani. An Introduction to the Bootstrap. Chapman & Hall, London, 1993. [4] W. HardIe. Applied Nonparametric Regression. Cambridge University Press, 1991. [5] T. Heskes. Balancing between bagging and bumping. These proceedings, 1997. [6] D. MacKay. A practical Bayesian framework for backpropagation. Neural Computation, 4:448-472, 1992. [7] D. Nix and A. Weigend. Estimating the mean and variance of the target probability distribution. In Proceedings of the [JCNN '94, pages 55-60. IEEE, 1994. [8] R. Tibshirani. A comparison of some error estimates for neural network models. Neural Computation, 8:152-163,1996.
1996
69
1,268
Minimizing Statistical Bias with Queries David A. Cohn Adaptive Systems Group Harlequin, Inc. One Cambridge Center Cambridge, MA 02142 cOhnCharlequin.com Abstract I describe a querying criterion that attempts to minimize the error of a learner by minimizing its estimated squared bias. I describe experiments with locally-weighted regression on two simple problems, and observe that this "bias-only" approach outperforms the more common "variance-only" exploration approach, even in the presence of noise. 1 INTRODUCTION In recent years, there has been an explosion of interest in "active" machine learning systems. These are learning systems that make queries, or perform experiments to gather data that are expected to maximize performance. When compared with "passive" learning systems, which accept given, or randomly drawn data, active learners have demonstrated significant decreases in the amount of data required to achieve equivalent performance. In industrial applications, where each experiment may take days to perform and cost thousands of dollars, a method for optimally selecting these points would offer enormous savings in time and money. An active learning system will typically attempt to select data that will minimize its predictive error. This error can be decomposed into bias and variance terms. Most research in selecting optimal actions or queries has assumed that the learner is approximately unbiased, and that to minimize learner error, variance is the only thing to minimize (e.g. Fedorov [1972]' MacKay [1992]' Cohn [1996], Cohn et al., [1996], Paass [1995]). In practice, however, there are very few problems for which we have unbiased learners. Frequently, bias constitutes a large portion of a learner's error; if the learner is deterministic and the data are noise-free, then bias is the only source of error. Note that the bias term here is a statistical bias, distinct from the inductive bias discussed in some machine learning research [Dietterich and Kong, 1995]. 418 D.A. Cohn In this paper I describe an algorithm which selects actions/ queries designed to minimize the bias of a locally weighted regression-based learner. Empirically, "varianceminimizing" strategies which ignore bias seem to perform well, even in cases where, strictly speaking, there is no variance to minimize. In the tasks considered in this paper, the bias-minimizing strategy consistently outperforms variance minimization, even in the presence of noise. 1.1 BIAS AND VARIANCE Let us begin by defining P(x, y) to be the unknown joint distribution over x and y, and P( x) to be the known marginal distribution of x (commonly called the input distribution). We denote the learner's output on input x, given training set D as y(x; D). We can then write the expected error of the learner as 1 E [(y(x;D) - y(x))2Ix] P(x)dx, (1) where E[·] denotes the expectation over P and over training sets D. The expectation inside the integral may be decomposed as follows (Geman et al., 1992): E [(y(x;D) - y(x))2Ix] E [(y(x) - E[ylx]?] (2) + (Ev [y(x; D)] - E[ylx])2 +Ev [(y(x;D) - Ev[y(x;D)])2] where Ev [.] denotes the expectation over training sets. The first term in Equation 2 is the variance of y given x - it is the noise in the distribution, and does not depend on our learner or how the training data are chosen. The second term is the learner's squared bias, and the third is its variance; these last two terms comprise the expected squared error of the learner with respect to the regression function E[Ylx]. Most research in active learning assumes that the second term of Equation 2 is approximately zero, that is, that the learner is unbiased. If this is the case, then one may concentrate on selecting data so as to minimize the variance of the learner. Although this "all-variance" approach is optimal when the learner is unbiased, truly unbiased learners are rare. Even when the learner's representation class is able to match the target function exactly, bias is generally introduced by the learning algorithm and learning parameters. From the Bayesian perspective, a learner is only unbiased if its priors are exactly correct. The optimal choice of query would, of course, minimize both bias and variance, but I leave that for future work. For the purposes of this paper, I will only be concerned with selecting queries that are expected to minimize learner bias. This approach is justified in cases where noise is believed to be only a small component of the learner's error. If the learner is deterministic and there is no noise, then strictly speaking, there is no error due to variance all the error must be due to learner bias. In cases with non-determinism or noise, all-bias minimization, like all-variance minimization, becomes an approximation of the optimal approach. The learning model discussed in this paper is a form of locally weighted regression (LWR) [Cleveland et al., 1988], which has been used in difficult machine learning tasks, notably the "robot juggler" of Schaal and Atkeson [1994]. Previous work [Cohn et al., 1996] discussed all-variance query selection for LWR; in the remainder of this paper, I describe a method for performing all-bias query selection. Section 2 describes the criterion that must be optimized for all-bias query selection. Section 3 describes the locally weighted regression learner used in this paper and describes Minimizing Statistical Bias with Queries 419 how the all-bias criterion may be computed for it. Section 4 describes the results of experiments using this criterion on several simple domains. Directions for future work are discussed in Section 5. 2 ALL-BIAS QUERY SELECTION Let us assume for the moment that we have a source of noise-free examples (Xi, Yi) and a deterministic learner which, given input X, outputs estimate Y(X).l Let us also assume that we have an accurate estimate of the bias of y which can be used to estimate the true function y(x) = y(x) - bias(x). We will break these rather strong assumptions of noise-free examples and accurate bias estimates in Section 4, but they are useful for deriving the theoretical approach described below. Given an accurate bias estimate, we must force the biased estimator into the best approximation of y(x) with the fewest number of examples. This, in effect, transforms the query selection problem into an example filter problem similar to that studied by Plutowski and White [1993] for neural networks. Below, I derive this criterion for estimating the change in error at X given a new queried example at x. Since we have (temporarily) assumed a deterministic learner and noise-free data, the expected error in Equation 2 simplifies to: E [(Y( X; 'D) - y( x))2Ix, 'D] (Y(x; 'D) - y(x))2 (3) We want to select a new x such that when we add (x, f)), the resulting squared bias is minimized: (Y' - y? == (y(x; 'D U (x, f))) - y(x))2 . (4) I will, for the remainder of the paper, use the "'" to indicate estimates based on the initial training set plus the additional example (x, y). To minimize Expression 4, we need to compute how a query at x will change the learner's bias at x. If we assume that we know the input distribution,2 then we can integrate this change over the entire domain (using Monte Carlo procedures) to estimate the resulting average change, and select a x such that the expected squared bias is minimized. Defining bias == y - y and f:,.y == y' - y, we can write the new squared bias as: bias,2 (y' - y)2 = (Y + f:,.y _ y)2 f:,.y2 + 2f:,.y . bias + bias2 (5) Note that since bias as defined here is independent of x, minimizing the bias is equivalent to minimizing f:,.y2 + 2f:,.y . bias. The estimate of bias' tells us how much our bias will change for a given x. We may optimize this value over x in one of a number of ways. In low dimensional spaces, it is often sufficient to consider a set of "candidate" x and select the one promising the smallest resulting error. In higher dimensional spaces, it is often more efficient to search for an optimal x with a response surface technique [Box and Draper, 1987], or hill climb on abias,2 / ax. Estimates of bias and f:,.y depend on the specific learning model being used. In Section 3, I describe a locally weighted regression model, and show how differentiable estimates of bias and f:,.y may be computed for it. 1 For clarity, I will drop the argument :z; except where required for disambiguation. I will also denote only the univariate case; the results apply in higher dimensions as well. 2This assumption is contrary to the assumption norma.lly made in some forms of learning, e.g. PAC-learning, but it is appropriate in many domains. 420 D. A. Cohn 2.1 AN ASIDE: WHY NOT JUST USE Y - Mas? If we have an accurate bias estimate, it is reasonable to ask why we do not simply use the corrected y - C;;;S as our predictor. The answer has two parts, the first of which is that for most learners, there are no perfect bias estimators they introduce their own bias and variance, which must be addressed in data selection. Second, we can define a composite learner Ye == Y - C:;;;S. Given a random training sample then, we would expect Ye to outperform y. However, there is no obvious way to select data for this composite learner other than selecting to maximize the performance of its two components. In our case, the second component (the bias estimate) is non-analytic, which leaves us selecting data so as to maximize the performance of the first component (the uncorrected estimator). We are now back to our original problem: we can select data so as to minimize either the bias or variance of the uncorrected LWR-based learner. Since the purpose of the correction is to give an unbiased estimator, intuition suggests that variance minimization would be the more sensible route in this case. Empirically, this approach does not appear to yield any benefit over uncorrected variance minimization (see Figure 1). 3 LOCALLY WEIGHTED REGRESSION The type of learner I consider here is a form of locally weighted regression (LWR) that is a slight variation on the LOESS model of Cleveland et al. [1988] (see Cohn et al., [1996] for details). The LOESS model performs a linear regression on points in the data set, weighted by a kernel centered at x. The kernel shape is a design parameter: the original LOESS model uses a "tricubic" kernel; in my experiments I use the more common Gaussian where Ie is a smoothing parameter. For brevity, I will drop the argument x for hi(x), and define n = 2:i hi. We can then write the weighted means and covariances as: "" Xi J.l:r; = L..J hi -, . n , Yi J.ly = L h,-, . n , "" (Xi - X)(Yi - J.ly) U:r;y = L..J hi . . n , We use these means and covariances to produce an estimate Y at the x around which the kernel is centered, with a confidence term in the form of a variance estimate: In all the experiments discussed in this paper, the smoothing parameter Ie was set so as to minimize u2. The low cost of incorporating new training examples makes this form of locally weighted regression appealing for learning systems which must operate in real time, or with time-varying target functions (e.g. [Schaal and Atkeson 1994]). Minimizing Statistical Bias with Queries 421 3.1 COMPUTING D..y FOR LWR If we know what new point (x, y) we're going to add, computing D..y for LWR is straightforward. Defining h as the weight given to x, and n as n + h we can write I A A A I A I U xy ( I ) U xy ( ) ~y = y - y = J.L + x J.L J.L - x - J.Lx Y U/2 x y u2 x x h (Y ___ J.Ly) _ uxy (x _ J.Lx) + (x _ n~x _ ~x) . nuXY_ + h . ~x -:xKii - J.Ly) n u; n n nu;+h·(x-J.Lx)2 Note that computing D..y requires us to know both the x and y of the new point. In practice, we only know x. If we assume, however, that we can estimate the learner's bias at any x, then we can also estimate the unknown value y ~ y(x) - bias(x) . Below, I consider how to compute the bias estimate. 3.2 ESTIMATING BIAS FOR LWR The most common technique for estimating bias is cross-validation. Standard crossvalidation however, only gives estimates of the bias at our specific training points, which are usually combined to form an average bias estimate. This is sufficient if one assumes that the training distribution is representative of the test distribution (which it isn't in query learning) and if one is content to just estimate the bias where one already has training data (which we can't be). In the query selection problem, we must be able to estimate the bias at all possible x. Box and Draper [1987] suggest fitting a higher order model and measuring the difference. For the experiments described in this paper, this method yielded poor results; two other bias-estimation techniques, however, performed very well. One method of estimating bias is by bootstrapping the residuals of the training points. One produces a "bootstrap sample" of the learner's residuals on the training data, and adds them to the original predictions to create a synthetic training set. By averaging predictions over a number of bootstrapped training sets and comparing the average prediction with that of the original predictor, one arrives at a first-order bootstrap estimate of the predictor's bias [Connor 1993; Efron and Tibshirani, 1993]. It is known that this estimate is itself biased towards zero; a standard heuristic is to divide the estimate by 0.632 [Efron, 1983]. Another method of estimating bias of a learner is by fitting its own cross-validated residuals. We first compute the cross-validated residuals on the training examples. These produce estimates of the learner's bias at each of the training points. We can then use these residuals as training examples for another learner (again LWR) to produce estimates of what the cross-validated error would be in places where we don't have training data. 4 EMPIRICAL RESULTS In the previous two sections, I have explained how having an estimate of D..y and bias for a learner allows one to compute the learner's change in bias given a new query, and have shown how these estimates may be computed for a learner that uses locally weighted regression. Here, I apply these results to two simple problems and demonstrate that they may actually be used to select queries that minimize the statistical bias (and the error) of the learner. The problems involve learning the kinematics of a planar two-jointed robot arm: given the shoulder and elbow joint angles, the learner must predict the tip position. 422 D.A. Cohn 4.1 BIAS ESTIMATES I tested the accuracy of the two bias estimators by observing their correlations on 64 reference inputs, given 100 random training examples from the planar arm problem. The bias estimates had a correlation with actual biases of 0.852 for the bootstrap method, and 0.871 for the cross-validation method. 4.2 BIAS MINIMIZATION I ran two sets of experiments using the bias-minimizing criterion in conjunction with the bias estimation technique of the previous section on the planar arm problem. The bias minimization criterion was used as follows: At each time step, the learner was given a set of 64 randomly chosen candidate queries and 64 uniformly chosen reference points. It evaluated E' (x) for each reference point given each candidate point and selected for its next query the candidate point with the smallest average E' (x) over the reference points. I compared the bias-minimizing strategy (using the cross-validation and bootstrap estimation techniques) against random sampling and the variance-minimizing strategy discussed in Cohn et al. [1996]. On a Sparc 10, with m training examples, the average evaluation times per candidate per reference point were 58 + 0.16m J.lseconds for the variance criterion, 65 + 0.53m J.lseconds for the cross-validation-based bias criterion, and 83 + 3. 7m J.lseconds for the bootstrapbased bias criterion (with 20x resampling). To test whether the bias-only assumption was robust against the presence of noise, 1 % Gaussian noise was added to the input values of the training data in all experiments. This simulates noisy position effectors on the arm, and results in nonGaussian noise in the output coordinate system. In the first series of experiments, the candidate shoulder and elbow joint angles were drawn uniformly over (U[O, 271"], U[O, 71")). In unconstrained domains like this, random sampling is a fairly good default strategy. The bias minimization strategies still significantly outperform both random sampling and the variance minimizing strategy in these experiments (see Figure 1). 10 -1 g '" 'il ,0-2 :a ' '\ ~ c: -3 ".'~ ~10 . random variance-min o cross-val-min ~ x bootstrap-min 10 0 100 200 300 trainlno set size 1 I, 10 \\ g 0 -"'10 }1O-1 ~ -2 ,- 10 .: ~~&e-.!)jni(llIZI~ , 10-3 ~ ~~~~rmiWngariOlmizing o 1 00 200 300 400 trainino set size Iheta 1 Figure 1: (left) MSE as a function of number of noisy training examples for the unconstrained arm problem. Errors are averaged over 10 runs for the bootstrap method and 15 runs for all others. One run with the cross-validation-based method was excluded when k failed to converge to a reasonable value. (center) MSE as a function of number of noisy training examples for the constrained arm problem. The bias correction strategy discussed in Section 2.1 does no better than the uncorrected variance-minimizing strategy, and much worse than the bias-minimization strategy. (right) Sample exploration trajectory in joint-space for the constrained arm problem, explored according to the bias minimizing criterion. In the second series of experiments, candidates were drawn uniformly from a region Minimizing Statistical Bias with Queries 423 local to the previously selected query: (01 ± 0.217", O2 ± 0.117"). This corresponds to restricting the arm to local motions. In a constrained problem such as this, random sampling is a poor strategy; both the bias and variance-reducing strategies outperform it at least an order of magnitude. Further, the bias-minimization strategy outperforms variance minimization by a large margin (Figure 1). Figure 1 also shows an exploration trajectory produced by pursuing the bias-minimizing criterion. It is noteworthy that, although the implementation in this case was a greedy (one-step) minimization, the trajectory results in globally good exploration. 5 DISCUSSION I have argued in this paper that, in many situations, selecting queries to minimize learner bias is an appropriate and effective strategy for active learning. I have given empirical evidence that, with a LWR-based learner and the examples considered here, the strategy is effective even in the presence of noise. Beyond minimizing either bias or variance, an important next step is to explicitly minimize them together. The bootstrap-based estimate should facilitate this, as it produces a complementary variance estimate with little additional computation. By optimizing over both criteria simultaneously, we expect to derive a criterion that that, in terms of statistics, is truly optimal for selecting queries. REFERENCES Box, G., & Draper, N. (1987). Empirical model-building and response surfaces, Wiley, New York. Cleveland, W., Devlin, S., & Grosse, E. (1988) . Regression by local fitting. Journal of Econometrics, 37, 87-114. Cohn, D. (1996) Neural network exploration using optimal experiment design. Neural Networks, 9(6):1071-1083. Cohn, D., Ghahramani, Z., & Jordan, M. (1996). Active learning with statistical models. Journal of Artificial Inteligence Research 4:129-145. Connor, J. (1993). Bootstrap Methods in Neural Network Time Series Prediction. In J . Alspector et al., eds., Proc. of the Int. Workshop on Applications of Neural Networks to Telecommunications, Lawrence Erlbaum, Hillsdale, N.J. Dietterich, T., & Kong, E. (1995). Error-correcting output coding corrects bias and variance. In S. Prieditis and S. Russell, eds., Proceedings of the 12th International Conference on Machine Learning. Efron, B. (1983) Estimating the error rate of a prediction rule: some improvements on cross-validation. J. Amer. Statist. Assoc. 78:316-331. Efron, B. & Tibshirani, R. (1993). An introduction to the bootstrap. Chapman & Hall, New York. Fedorov, V. (1972). Theory of Optimal Experiments. Academic Press, New York. Geman, S., Bienenstock, E., & Doursat, R. (1992). Neural networks and the bias/variance dilemma. Neural Computation, 4, 1-58. MacKay, D. (1992). Information-based objective functions for active data selection, Neural Computation, 4, 590-604. Paass, G., and Kindermann, J. (1994). Bayesian Query Construction for Neural Network Models. In G. Tesauro et al., eds., Advances in Neural Information Processing Systems 7, MIT Press. Plutowski, M., & White, H. (1993). Selecting concise training sets from clean data. IEEE Transactions on Neural Networks, 4, 305-318. Schaal, S. & Atkeson, C. (1994). Robot Juggling: An Implementation of Memory-based Learning. Control Systems 14, 57-71.
1996
7
1,269
Consistent Classification, Firm and Soft Yoram Baram* Department of Computer Science Technion, Israel Institute of Technology Haifa 32000, Israel baram@cs.technion.ac.il Abstract A classifier is called consistent with respect to a given set of classlabeled points if it correctly classifies the set. We consider classifiers defined by unions of local separators and propose algorithms for consistent classifier reduction. The expected complexities of the proposed algorithms are derived along with the expected classifier sizes. In particular, the proposed approach yields a consistent reduction of the nearest neighbor classifier, which performs "firm" classification, assigning each new object to a class, regardless of the data structure. The proposed reduction method suggests a notion of "soft" classification, allowing for indecision with respect to objects which are insufficiently or ambiguously supported by the data. The performances of the proposed classifiers in predicting stock behavior are compared to that achieved by the nearest neighbor method. 1 Introduction Certain classification problems, such as recognizing the digits of a hand written zipcode, require the assignment of each object to a class. Others, involving relatively small amounts of data and high risk, call for indecision until more data become available. Examples in such areas as medical diagnosis, stock trading and radar detection are well known. The training data for the classifier in both cases will correspond to firmly labeled members of the competing classes. (A patient may be ·Presently a Senior Research Associate of the National Research Council at M. S. 210-9, NASA Ames Research Center, Moffett Field, CA 94035, on sabbatical leave from the Technion. Consistent Classification, Firm and Soft 327 either ill or healthy. A stock price may increase, decrease or stay the same). Yet, the classification of new objects need not be firm. (A given patient may be kept in hospital for further observation. A given stock need not be bought or sold every day). We call classification of the first kind "firm" and classification of the second kind "soft". The latter is not the same as training the classifier with a "don't care" option, which would be just another firm labeling option, as "yes" and "no", and would require firm classification. A classifier that correctly classifies the training data is called "consistent". Consistent classifier reductions have been considered in the contexts of the nearest neighbor criterion (Hart, 1968) and decision trees (Holte, 1993, Webb, 1996). In this paper we present a geometric approach to consistent firm and soft classification. The classifiers are based on unions of local separators, which cover all the labeled points of a given class, and separate them from the others. We propose a consistent reduction of the nearest neighbor classifier and derive its expected design complexity and the expected classifier size. The nearest neighbor classifier and its consistent derivatives perform "firm" classification. Soft classification is performed by unions of maximal- volume spherical local separators. A domain of indecision is created near the boundary between the two sets of class-labeled points, and in regions where there is no data. We propose an economically motivated benefit function for a classifier as the difference between the probabilities of success and failure. Employing the respective benefit functions, the advantage of soft classification over firm classification is shown to depend on the rate of indecision. The performances of the proposed algorithms in predicting stock behavior are compared to those of the nearest neighbor method. 2 Consistent Firm Classification Consider a finite set of points X = {X(i), i = 1, ... , N} in some subset of Rn, the real space of dimension n. Suppose that each point of X is assigned to one of two classes, and let the corresponding subsets of X, having N1 and N2 points, respectively, be denoted Xl and X 2 • We shall say that the two sets are labeled L1 and L 2 , respectively. It is desired to divide Rn into labeled regions, so that new, . unlabeled points can be assigned to one of the two classes. We define a local separator of a point x of Xl with respect to X 2 as a convex set, s(xI2), which contains x and no point of X2. A separator family is defined as a rule that produces local separators for class- labeled points. We call the set of those points of Rn that are closer to a point x E Xl than to any point of X2 the minimum-distance local separator of x with respect to X2. We define the local clustering degree, c, of the data as the expected fraction of data points that are covered by a local minimum- distance separator. The nearest neighbor criterion extends the class assignment of a point x E Xl to its minimum-distance local separator. It is clearly a consistent and firm classifier whose memory size is O(N). Hart's Condensed Nearest Neighbor (CNN) classifier (Hart, 1968) is a consistent subset of the data points that correctly classifies the entire data by the nearest neighbor method. It is not difficult to show that the complexity of the algorithm 328 Y. Baram proposed by Hart for finding such a subset is O(N3). The expected memory requirement (or classifier size) has remained an open question. We propose the following Reduced Nearest Neighbor (RNN) classifier: include a labeled point in the consistent subset only if it is not covered by the minimumdistance local separator of any of the points of the same class already in the subset. It can be shown (Baram, 1996) that the complexity of the RNN algorithm is O(N2). and that the expected classifier size is O(IOgl/(I-C) N). It can also be shown that the latter bounds the expected size of the CNN classifier as well. It has been suggested that the utility of the Occam's razor in classification would be (Webb, 1996): "Given a choice between two plausible classifiers that perform identically on the data set, the simpler classifier is expected to classify correctly more objects outside the training set". The above statement is disproved by the CNN and the RNN classifiers, which are strict consistent reductions of the nearest neighbor classifier, likely to produce more errors. 3 Soft Classification: Indecision Pays, Sometimes When a new, unlabeled, point is closely surrounded by many points of the same class, its assignment to the same class can be said to be unambiguously supported by the data. When a new point is surrounded by points of different classes, or when it is relatively far from any of the labeled points, its assignment to either class can be said to be unsupported or ambiguously supported by the data. In the latter cases, it may be more desirable to have a certain indecision domain, where new points will not be assigned to a class. This will translate into the creation of indecision domains near the boundary between the two sets of labeled points and where there is no data. We define a separntor S(112) of Xl with respect to X2 as a set that includes Xl and excludes X2. Given a separator family, the union of local separators S(x(i) 12) of the points X(i), i = 1 ... , N I , of Xl with respect to X 2 , is a separator of Xl with respect to X2. It consists of NI local separators. Let XI,c be a subset of Xl. The set (1) (2) will be called a consistent separator of Xl with respect to X2 if it contains all the points of X 1. The set XI,c will then be called a consistent subset with respect to the given separator family. Let us extend the class assignment of each of the labeled points to a local separator of a given family and maximize the volume of each of the local separators without Consistent Classification, Finn and Soft 329 including in it any point of the competing class. Let Sc(112) and Sc(211) be consistent separators of the two sets, consisting of maximal-volume (or, simply, maximaQ local separators of labeled points of the corresponding classes. The intersection of Sc(112) and Sc(211) defines a conflict and will be called a domain of ambiguity of the first kind. A region uncovered by either separator will be called a domain of ambiguity of the second kind. The union of the domains of ambiguity will be designated the domain of indecision. The remainders of the two separators, excluding their intersection, define the conflict-free domains assigned to the two classes. The resulting "soft" classifier rules out hard conflicts, where labeled points of one class are included in the separator of the other. Yet, it allows for indecision in areas which are either claimed by both separators or claimed by neither. Let the true class be denoted y (with possible values, e.g., y=1 or y=2) and let the classification outcome be denoted y. Let the probabilities of decision and indecision by the soft classifier be denoted Pd and Pid, respectively (of course, P id = 1 - Pd), and let the probabilities of correct and incorrect decisions by the firm and the soft classifiers be denoted Pfirm {y = y}, Pfirm {y =P y}, P soft {y = y} and Psoft {y =P y}, respectively. Finally, let the joint probabilities of a decision being made by the soft classifier and the correctness or incorrectness of the decision be denoted, respectively, Psoft { d, Y = y} and P soft { d, Y =P y} and let the corresponding conditional probabilities be denoted Psoft {y = y I d} and Psoft {y =P y I d}, respectively. We define the benefit of using the firm classifier as the difference between the probability that a point is classified correctly by the classifier and the probability that it is misclassified: This definition is motivated by economic consideration: the profit produced by an investment will be, on average, proportional to the benefit function. This will become more evident in a later section, were we consider the problem of stock trading. For a soft classifier, we similarly define the benefit as the difference between the probability of a correct classification and that of an incorrect one (which, in an economic context, assumes that indecision has no cost, other than the possible loss of profit). Now, however, these probabilities are for the joint events that a classification is made, and that the outcome is correct or incorrect, respectively: (4) Soft classification will be more beneficial than firm classification if Bsoft > Bfirm' which may be written as Pfirm{Y = y} - 0.5 Pid < 1- Psoft{Y =y I d} -0.5 (5) For the latter to be a useful condition, it is necessary that Pfirm {y = y} > 0.5, Psofdy = y I d} > 0.5 and Psoft {y = y I d} > Pfirm {y = y}. The latter will be normally satisfied, since points of the same class can be expected to be denser under the corresponding separator than in the indecision domain. In other words, 330 Y. Baram the error ratio produced by the soft classifier on the decided cases can be expected to be smaller than the error ratio produced by the firm classifier, which decides on all the cases. The satisfaction of condition (5) would depend on the geometry of the data. It will be satisfied for certain cases, and will not be satisfied for others. This will be numerically demonstrated for the stock trading problem. The maximal local spherical separator of x is defined by the open sphere centered at x, whose radius r(xI2) is the distance between x and the point of X2 nearest to x. Denoting by s(x, r) the sphere of radius r in Rn centered at x, the maximal local separator is then sM(xI2) = s(x, r(xI2)). A separator construction algorithm employing maximal local spherical separators is described below. Its complexity is clearly O(N2). Let Xl = Xl. For each of the points xci) of Xl, find the minimal distance to the points of X 2 • Call it r(x(i) 12). Select the point x(i) for which r(x(i) 12) 2: r(x(j) 12), j f: i, for the consistent subset. Eliminate from Xl all the points that are covered by SM(X(i) 12). Denote the remaining set Xl. Repeat the procedure while Xl is non-empty. The union of the maximal local spherical separators is a separator for Xl with respect to X 2 . 4 Example: Firm and soft prediction of stock behaviour Given a sequence of k daily trading ("close") values of a stock, it is desired to predict whether the next day will show an increase or a decrease with respect to the last day in the sequence. Records for ten different stocks, each containing, on average, 1260 daily values were used. About 60 percent of the data were used for training and the rest for testing. The CNN algorithm reduced the data by 40% while the RNN algorithm reduced the data by 35%. Results are show in Fig. 1. It can be seen that, on average, the nearest neighbor method has produced the best results. The performances of the CNN and the RNN classifiers (the latter producing only slightly better results) are somewhat lower. It has been argued that performance within a couple of percentage points by a reduced classifier supports the utility of Occam's razor (Holte, 1993). However, a couple of percentage points can be quite meaningful in stock trading. In order to evaluate the utility of soft classification in stock trading, let the prediction success rate of a firm classifier, be denoted f and that of a soft classifier for the decided cases s. For a given trade, let the gain or loss per unit invested be denoted q, and the rate of indecision of the soft classifier ir. Suppose that, employing the firm classifier, a stock is traded once every day (say, at the "close" value), and that, employing the soft classifier, it is traded on a given day only if a trade is decided by the classifier (that is, the input does not fall in the indecision domain). The expected profit for M days per unit invested is 2(1 - 0.5)qM for the firm classifier and 2(s - 0.5)q(l- ir)M for the soft classifier (these values disregard possible commission and slippage costs). The soft classifier will be preferred over the firm one if the latter quantity is greater than the former, that is, if . 1 f - 0.5 ~r < - =---s - 0.5 (6) which is the sample representation of condition (5) for the stock trading problem. Consistent Classification, Firm and Soft 331 NN CNN RNN ~ni . ~llI~~ifip.· ..... . . ~llCce~~ bene.fit xaip 61.4% 58.4% 57.0% 48.4% 70.1% xaldnf 51.7% 52.3% 50.1% 74.6% 51.3% xddddf 52.0% 49.6% 51.8% 44.3% 53.3% xdssi 48.3% 47.7% 48.6% 43.6% 52.6% + xecilf 53.0% 50.9% 52.6% 47.6% 48.8% xedusf 80.7% 74.7% 76.3% 30.6% 89.9% xelbtf 53.7% 55.6% 52.5% 42.2% 50.1% xetz 66.0% 61.0% 61.0% 43.8% 68.6% xelrnf 51.5% 49.0% 49.2% 39.2% 56.0% + xelt 85.6% 82.7% 84.2% 32.9% 93.0% Average 60.4% 58.2% 58.3% 44.7% 63.4% Figure 1: Success rates in the prediction of rize and fall in stock values. Results for the soft classifier, applied to the stock data, are presented in Fig. 1. The indecision rates and the success rates in the decided cases are then specified along with a benefit sign. A positive benefit represents a satisfaction of condition (6), with ir, f and s replaced by the corresponding sample values given in the table. This indicates a higher profit in applying the soft classifier over the application of the nearest neighbor classifier. A negative benefit indicates that a higher profit is produced by the nearest neighbor classifier. It can be seen that for two of the stocks (xdssi and xelrnf) soft classification has produced better results than firm classification, and for the remaining eight stocks finn classification by the nearest neighbor method has produced better results. 5 Conclusion Solutions to the consistent classification problem have been specified in tenns of local separators of data points of one class with respect to the other. The expected complexities of the proposed algorithms have been specified, along with the expected sizes of the resulting classifiers. Reduced consistent versions of the nearest neighbor classifier have been specified and their expected complexities have been derived. A notion of "soft" classification has been introduced an algorithm for its implementation have been presented and analyzed. A criterion for the utility of such classification has been presented and its application in stock trading has been demonstrated. Acknowledgment The author thanks Dr. Amir Atiya of Cairo University for providing the stock data used in the examples and for valuable discussions of the corresponding results. 332 y. Baram References Baram Y. (1996) Consistent Classification, Firm and Soft, CIS Report No. 9627, Center for Intelligent Systems, Technion, Israel Institute of Technology, Haifa 32000, Israel. Baum, E. B. (1988) On the Capabilities of Multilayer Perceptrons, J. Complexity, Vol. 4, pp. 193 - 215. Hart, P. E. (1968) The Condensed Nearest Neighbor Rule, IEEE Trans. on Information Theory, Vol. IT -14, No.3, pp. 515 - 516. Holte, R. C. (1993) Very Simple Classification Rules Perform Well on Most Commonly Used databases, Machine Learning, Vol. 11, No. 1 pp. 63 - 90. Rosenblatt, F. (1958) The Perceptron: A Probabilistic Model for Information Storage and Organization in the Brain, Psychological Review, Vol. 65, pp. 386 - 408. Webb, G. 1. (1996) Further Experimental Evidence against the Utility of Occam's Razor, J. of Artificial Intelligence Research 4, pp. 397 - 147.
1996
70
1,270
Neural Models for Part-Whole Hierarchies Maximilian Riesenhuber Peter Dayan Department of Brain & Cognitive Sciences Massachusetts Institute of Technology Cambridge, MA 02139 {max,dayan}~ai.mit.edu Abstract We present a connectionist method for representing images that explicitly addresses their hierarchical nature. It blends data from neuroscience about whole-object viewpoint sensitive cells in inferotemporal cortex8 and attentional basis-field modulation in V43 with ideas about hierarchical descriptions based on microfeatures.5,11 The resulting model makes critical use of bottom-up and top-down pathways for analysis and synthesis.6 We illustrate the model with a simple example of representing information about faces. 1 Hierarchical Models Images of objects constitute an important paradigm case of a representational hierarchy, in which 'wholes', such as faces, consist of 'parts', such as eyes, noses and mouths. The representation and manipulation of part-whole hierarchical information in fixed hardware is a heavy millstone around connectionist necks, and has consequently been the inspiration for many interesting proposals, such as Pollack's RAAM.l1 We turned to the primate visual system for clues. Anterior inferotemporal cortex (IT) appears to construct representations of visually presented objects. Mouths and faces are both objects, and so require fully elaborated representations, presumably at the level of anterior IT, probably using different (or possibly partially overlapping) sets of cells. The natural way to represent the part-whole relationship between mouths and faces is to have a neuronal hierarchy, with connections bottom-up from the mouth units to the face units so that information about the mouth can be used to help recognize or analyze the image of a face, and connections top-down from the face units to the mouth units expressing the generative or synthetic knowledge that if there is a face in a scene, then there is (usually) a mouth too. There is little We thank Larry Abbott, Geoff Hinton, Bruno Olshausen, Tomaso Poggio, Alex Pouget, Emilio Salinas and Pawan Sinha for discussions and comments. 18 M. Riesenhuberand P. Dayan empirical support for or against such a neuronal hierarchy, but it seems extremely unlikely on the grounds that arranging for one with the correct set of levels for all classes of objects seems to be impossible. There is recent evidence that activities of cells in intermediate areas in the visual processing hierarchy (such as V4) are influenced by the locus of visual attention.3 This suggests an alternative strategy for representing part-whole information, in which there is an interaction, subject to attentional control, between top-down generative and bottom-up recognition processing. In one version of our example, activating units in IT that represent a particular face leads, through the top-down generative model, to a pattern of activity in lower areas that is closely related to the pattern of activity that would be seen when the entire face is viewed. This activation in the lower areas in turn provides bottom-up input to the recognition system. In the bottom-up direction, the attentional signal controls which aspects of that activation are actually processed, for example, specifying that only the activity reflecting the lower part of the face should be recognized. In this case, the mouth units in IT can then recognize this restricted pattern of activity as being a particular sort of mouth. Therefore, we have provided a way by which the visual system can represent the part-whole relationship between faces and mouths. This describes just one of many possibilities. For instance, attentional control could be mainly active during the top-down phase instead. Then it would create in VI (or indeed in intermediate areas) just the activity corresponding to the lower portion of the face in the first place. Also the focus of attention need not be so ineluctably spatial. The overall scheme is based on an hierarchical top-down synthesis and bottom-up analysis model for visual processing, as in the Helmholtz machine6 (note that "hierarchy" here refers to a processing hierarchy rather than the part-whole hierarchy discussed above) with a synthetic model forming the effective map: 'object' 18) 'attentional eye-position' -t 'image' (1) (shown in cartoon form in figure 1) where 'image' stands in for the (probabilities over the) activities of units at various levels in the system that would be caused by seeing the aspect of the 'object' selected by placing the focus and scale of attention appropriately. We use this generative model during synthesis in the way described above to traverse the hierarchical description of any particular image. We use the statistical inverse of the synthetic model as the way of analyzing images to determine what objects they depict. This inversion process is clearly also sensitive to the attentional eye-position - it actually determines not only the nature of the object in the scene, but also the way that it is depicted (ie its instantiation parameters) as reflected in the attentional eye position. In particular, the bottom-up analysis model exists in the connections leading to the 2D viewpoint-selective image cells in IT reported by Logothetis et al8 which form population codes for all the represented images (mouths, noses, etc). The top-down synthesis model exists in the connections leading in the reverse direction. In generalizations of our scheme, it may, of course, not be necessary to generate an image all the way down in VI. The map (1) specifies a top-down computational task very like the bottom-up one addressed using a multiplicatively controlled synaptic matrix in the shifter model Neural Modelsfor Part-Whole Hierarchies layer 3 o 2 m 1 p "- ... ,'iIL.~ \ ';:111 .. , ,~ '-" 19 attentional eye position e =(c;, tyl t%) Figure 1: Cartoon of the model. In the top-down, generative, direction, the model generates images of faces, eyes, mouths or noses based on an attentional eye position and a selection of a single top-layer unit; the bottom-up, recognition, direction is the inverse of this map. The response of the neurons in the middle layer is modulated sigmoidally (as illustrated by the graphs shown inside the circles representing the neurons in the middle layer) by the attentional eye position. See section 2 for more details. of Olshausen et al.9 Our solution emerges from the control the attentional eye position exerts at various levels of processing, most relevantly modulating activity in V4.3 Equivalent modulation in the parietal cortex based on actual (rather than attentional) eye position! has been characterized by Pouget & Sejnowski13 and Salinas & Abbott15 in terms of basis fields. They showed that these basis fields can be used to solve the same tasks as the shifter model but with neuronal rather than synaptic multiplicative modulation. In fact, eye-position modulation almost certainly occurs at many leve~s in the system, possibly including VIP Our sch~me clearly requires that the modulating attentional eye-position must be able to become detached from the spatial eye-position - Connor et al.3 collected evidence for part of this hypothesis; although the coordinate system(s) of the modulation is not entirely clear from their data. Bottom-up and top-down mappings are learned taking the eye-position modulation into account. In the experiments below, we used a version of the wake-sleep algorithm,6 for its conceptual and computational simplicity. This requires learning the bottom-up model from generated imagery (during sleep) and learning the topdown model from assigned explanations (during observation of real input during wake). In the current version, for simplicity, the eye position is set correctly during recognition, but we are also interested in exploring automatic ways of doing this. 2 Results We have developed a simple model that illustrates the feasibility of the scheme presented above in the context of recognizing and generating cartoon drawings of a face and its parts. Recognition involves taking an image of a face or a part thereof (the mouth, nose or one of the eyes) at an arbitrary position on the retina, 20 M. Riesenhuber and P. Dayan a) b) faoo nose ~ .} ::0 GJ ::Li] ..., .. ~ .. , , ~ f~. _,...... __ ...... .:c.. _no .. _ mouth eye [iJ ··'0 [!J ·'0 •.• l. ~ •.• '~.' ~J:: .... J •. 0 • . ~. ,__ ..--..th noe__ __ ............ _ __ Figure 2: a) Recognition: the left column of each pair shows the stimuli; the right shows the resulting activations in the top layer (ordered as face, mouth, nose and eye). The stimuli are faces at random positions in the retina. Recognition is performed by setting the attentional eye position in the image and setting the attentional scale, which creates a window of attention around the attended to position, shown by a circle of corresponding size and position. b) Generation: each panel shows the output of the generative pathway for a randomly chosen attentional eye position on activating each of the top layer units in turn. The focus of attention is marked by a circle whose size reflects the attentional scale. The name of the object whose neuronal representation in the top layer was activated is shown above each panel. and setting the appropriate top level unit to 1 (and the remaining units to zero). Generation involves imaging either a whole face or of one of its parts (selected by the active unit in the top layer) at an arbitrary position on the retina. The model (figure 1) consists of three layers. The lowest layer is a 32 x 32 'retina'. In the recognition direction, the retina feeds into a layer of 500 hidden units. These project to the top layer, which has four neurons. In the generative direction, the connectivity is reversed. The network is fully connected in both directions. The activity of each neuron based on input from the preceding (for recognition) or the following layer (for generation) is a linear function (weight matrices wr, Vr in the recognition and vg, W g in the generative direction). The attentional eye position influences activity through multiplicative modulation of the neuronal responses in the hidden layer. The linear response ri = (Wrp)i or ri = (VgO)i of each neuron i in the middle layer based on the bottom-up or top-down connections is multiplied by ~i = ¢i(ex)¢f(ey)¢f(es), where ¢!x,y,s} are the tuning curves in each dimension of the attentional eye position e = (eX, eY , eS), coding the x- and y- coordinates and the scale of the focus of attention, respectively. Thus, for the activity mi of hidden neuron i we have mi = (Wrp)i ·~i in the recognition pathway and mi = (VgO)i ·~i in the generative pathway. The tuning curves of the ~i are chosen to be sigmoid with random centers Ci and random directions di E {-I, I}, eg ¢! = u( 4 * d! * (e S - cn). In other implementations, we have also used Gaussian tuning functions. In fact, the only requirement regarding the shape of the tuning functions is that through a superposition of them one can construct functions that show a peaked dependence on the attentional eye position. In the recognition direction, the attentional eye position also has an influence on the activity in the input layer by defining a 'window of attention',7 which we implemented using a Gaussian window centered at the attentional eye position with its size given by the attentional scale. This is to allow the system to learn models of parts based on experience with images of whole faces. To train the model, we employ a variant of the unsupervised wake-sleep algorithm. 6 In this algorithm, the generative pathway is trained during a wake-phase, in which Neural Models/or Part-Whole Hierarchies 21 stimuli in the input layer (the retina, in our case) cause activation of the neurons in the network through the recognition pathway, providing an error signal to train the generative pathway using the delta rule. Conversely, in the sleep-phase, random activation of a top layer unit (in conjunction with a randomly chosen attentional eye-position) leads, via the generative connections, to the generation of activation in the middle layer and consequently an image in the input layer that is then used to adapt the recognition weights, again using the delta rule. Although the delta rule in wake-sleep is fine for the recognition direction, it leads to a poor generative model - in our simple case, generation is much more difficult than recognition. As an interim solution, we therefore train the generative weights using back-propagation, which uses the activity in the top layer created by the recognition pathway as the input and the retinal activation pattern as the target signal. Hence, learning is still unsupervised (except that appropriate attentional eye-positions are always set during recognition). We have also experimented with a system in which the weights wr and w g are preset and only the weights between layers 2 and 3 are trained. For this model, training could be done with the standard wake-sleep algorithm, ie using the local delta-rule for both sets of weights. Figure 2a shows several examples of the performance of the recognition pathway for the different stimuli after 300,000 iterations. The network is able to recognize the stimuli accurately at different positions in the visual field. Figure 2b shows several examples of the output of the generative model, illustrating its capacity to produce images of faces or their parts at arbitrary locations. By imaging a whole face and then focusing the attention on eg an area around its center, which activates the 'nose' unit through the recognition pathway, the relationship that, eg a nose is part of a face can be established in a straightforward way. 3 Discussion Representing hierarchical structure is a key problem for connectionism. Visual images offer a canonical example for which it seems possible to elucidate some of the underlying neural mechanisms. The theory is based on 2D view object selective cells in anterior IT, and attentional eye-position modulation of the firing of cells in V 4. These work in the context of analysis by synthesis or recognition and generative models such that the part-whole hierarchy of an object such as a face (which contains eyes, which contain pupils, etc) can be traversed in the generative direction by choosing to view the object through a different effective eye-position, and in the recognition direction by allowing the real and the attentional eye-positions to be decoupled to activate the requisite 2D view selective cells. The scheme is related to Pollack's Recursive Auto-Associative Memory (RAAM) system.l1 RAAM provides a way of representing tree-structured information - for instance to learn an object whose structure is {{A,B},{C,D}}, a standard threelayer auto-associative net would be taught AB, leading to a pattern of hidden unit activations 0:; then it would learn CD leading to (3; and finally 0:(3 leading to I, which would itself be the representation of the whole object. The compression operation (AB -t 0:) and its expansion inverse are required as explicit methods for manipulating tree structure. Our scheme for representing hierarchical information is similar to RAAM, using the notion of an attentional eye-position to perform its compression and expansion 22 M. Riesenhuberand P. Dayan operations. However, whereas RAAM normally constructs its own codes for intermediate levels of the trees that it is fed, here, images of faces are as real and as available as those, for instance, of their associated mouths. This not only changes the learning task, but also renders sensible a notion of direct recognition without repeated RAAMification of the parts. Various aspects of our scheme require comment: the way that eye position affects recognition; the coding of different instances of objects; the use of top-down information during bottom-up recognition; variants of the scheme for objects that are too big or too geometrically challenging to 'fit' in one go into a single image; and hierarchical objects other than images. We are also working on a more probabilistically correct version, taking advantage of the statistical soundness of the Helmholtz machine. Eye position information is ubiquitous in visual processing areas,12 including the LG N and VI, 17 as well as the parietal cortex 1 and V 4.3 Further, it can be revealed as having a dramatic effect on perception, as in Ramachandran et al'sl4 study on intermittent exotropes. This is a form of squint in which the two eyes are normally aligned, but in which the exotropic eye can deviate (voluntarily or involuntarily) by as much as 60°. The study showed that even if an image is 'burnt' on the retina in this eye as an afterimage, and so is fixed in retinal coordinates, at least one component of the percept moves as the eye moves. This argues that information about eye position dramatically effects visual processing in a manner that is consistent with the model presented here of shifts based on modulation. This is also required by Bridgeman et al's2 theory of perceptual stability across fixations, that essentially builds up an impression of a scene in exactly the form of mapping (1). In general, there will be many instances for an object, e.g., many different faces. In this general case, the top level would implement a distributed code for the identity and instantiation parameters of the objects. We are currently investigating methods of implementing this form of representation into the model. A key feature of the model is the interaction of the synthesis and analysis pathways when traversing the part-whole hierarchies. This interaction between the two pathways can also aid the system when performing image analysis by integrating information across the hierarchy. Just as in RAAM, the extra feature required when traversing a hierarchy is short term memory. For RAAM, the memory stores information about the various separate sub-trees that have already been decoded (or encoded). For our system, the memory is required during generative traversal to force 'whole' activity on lower layers to persist even after the activity on upper layers has ceased, to free these upper units to recognize a 'part'. Memory during recognition traversal is necessary in marginal cases to accumulate information across separate 'parts' as well as the 'whole'. This solution to hierarchical representation inevitably gives up the computational simplicity of the naive neuronal hierarchical scheme described in the introduction which does not require any such accumulation. Knowledge of images that are too large to fit naturally in a single view4 at a canonical location and scale, or that theoretically cannot fit in a view (like 360° information about a room) can be handled in a straightforward extension of the scheme. All this requires is generalizing further the notion of eye-position. One can explore one's generative model of a room in the same way that one can explore one's generative model of a face. Neural Models/or Part-Whole Hierarchies 23 We have described our scheme from the perspective of images. This is convenient because of the substantial information available about visual processing. However, images are not the only examples of hierarchical structure - this is also very relevant to words, music and also inferential mechanisms. We believe that our mechanisms are also more general - proving this will require the equivalent of the attentional eye-position that lies at the heart of the method. References [1] Andersen, R, Essick, GK & Siegel, RM (1985). Encoding of spatial location by posterior parietal neurons. Science, 230, 456-458. [2] Bridgeman, B, van der Hejiden, AHC & Velichkovsky, BM (1994). A theory of visual stability across saccadic eye movements. Behavioral and Brain Sciences, 17, 247-292. [3] Connor, CE, Gallant, JL, Preddie, DC & Van Essen, DC (1996). Responses in area V 4 depend on the spatial relationship between stimulus and attention. Journal of Neurophysiology, 75, 1306-1308. [4] Feldman, JA (1985). Four frames suffice: A provisional model of vision and space. The Behavioral and Brain Sciences, 8, 265-289. [5] Hinton, GE (1981). Implementing semantic networks in parallel hardware. In GE Hinton & JA Anderson, editors, Parallel Models of Associative Memory. Hillsdale, NJ: Erlbaum, 161-188. [6] Hinton, GE, Dayan, P, Frey, BJ & Neal, RM (1995). The wake-sleep algorithm for unsupervised neural networks. Science, 268, 1158-1160. [7] Koch, C & Ullmann, S (1985). Shifts in selective visual attention: towards the underlying neural circuitry. Human Neurobiology, 4, 219-227. [8] Logothetis, NK, Pauls, J, & Poggio, T (1995). Shape representation in the inferior temporal cortex of monkeys. Current Biology, 5, 552-563. [9] Olshausen, BA, Anderson, CH & Van Essen, DC (1993). A neurobiological model of visual attention and invariant pattern recognition based on dynamic routing of information. Journal of Neuroscience, 13, 4700-4719. [10] Pearl, J (1988). Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. San Mateo, CA: Morgan Kaufmann. [11] Pollack, JB (1990). Recursive distributed representations. Artificial Intelligence, 46, 77-105. [12] Pouget, A, Fisher, SA & Sejnowski, TJ (1993). Egocentric spatial representation in early vision. Journal of Cognitive Neuroscience, 5, 150-161. [13] Pouget, A & Sejnowski, TJ (1995). Spatial representations in the parietal cortex may use basis functions. In G Tesauro, DS Touretzky & TK Leen, editors, Advances in Neural Information Processing Systems 7, 157-164. [14] Ramachandran, VS, Cobb, S & Levi, L (1994). The neural locus of binocular rivalry and monocular diplopia in intermittent exotropes. Neuroreport, 5, 1141-1144. [15] Salinas, E & Abbott LF (1996). Transfer of coded information from sensory to motor networks. Journal of Neuroscience, 15, 6461-6474. [16] Sung, K & Poggio, T (1995). Example based learning for view-based human face detection. AI Memo 1521, CBCL paper 112, Cambridge, MA: MIT. [17] Trotter, Y, Celebrini, S, Stricanne, B, Thorpe, S & Imbert, M (1992). Modulation of neural stereoscopic processing in primate area VI by the viewing distance. Science, 257, 1279-1281. PART II NEUROSCIENCE
1996
71
1,271
Bayesian Unsupervised Learning of Higher Order Structure Michael S. Lewicki Terrence J. Sejnowski levicki~salk.edu terry~salk.edu The Salk Institute Howard Hughes Medical Institute Computational Neurobiology Lab 10010 N. Torrey Pines Rd. La Jolla, CA 92037 Abstract Multilayer architectures such as those used in Bayesian belief networks and Helmholtz machines provide a powerful framework for representing and learning higher order statistical relations among inputs. Because exact probability calculations with these models are often intractable, there is much interest in finding approximate algorithms. We present an algorithm that efficiently discovers higher order structure using EM and Gibbs sampling. The model can be interpreted as a stochastic recurrent network in which ambiguity in lower-level states is resolved through feedback from higher levels. We demonstrate the performance of the algorithm on benchmark problems. 1 Introduction Discovering high order structure in patterns is one of the keys to performing complex recognition and discrimination tasks. Many real world patterns have a hierarchical underlying structure in which simple features have a higher order structure among themselves. Because these relationships are often statistical in nature, it is natural to view the process of discovering such structures as a statistical inference problem in which a hierarchical model is fit to data. Hierarchical statistical structure can be conveniently represented with Bayesian belief networks (Pearl, 1988; Lauritzen and Spiegelhalter, 1988; Neal, 1992). These 530 M. S. Lewicki and T. 1. Sejnowski models are powerful, because they can capture complex statistical relationships among the data variables, and also mathematically convenient, because they allow efficient computation of the joint probability for any given set of model parameters. The joint probability density of a network of binary states is given by a product of conditional probabilities (1) where W is the weight matrix that parameterizes the model. Note that the probability of an individual state Si depends only on its parents. This probability is given by P(Si = 1lpa[Si], W) = h(L SjWji) j where Wji is the weight from Sj to Si (Wji = 0 for j < i). (2) The weights specify a hierarchical prior on the input states, which are the fixed subset of states at the lowest layer of units. The active parents of state Si represent the underlying causes of that state. The function h specifies how these causes are combined to give the probability of Si. We assume h to be the "noisy OR" function, h(u) = 1 - exp( -u), u >= O. 2 Learning Objective The learning objective is to adapt W to find the most probable explanation of the input patterns. The probability of the input data is (3) n P(Dn IW) is computed by marginalizing over all states of the network P(DnIW) = L P(DnISk, W)P(SkIW) (4) k Because the number of different states, Sk, is exponential in the number of units, computing the sum exactly is intractable and must be approximated. The nature of the learning tasks discussed here, however, allow us to make accurate approximations. A desirable property for representations is that most patterns have just one or a few possible explanations. In this case, all but a few terms P(Dn ISk, W) will be zero, and, as described below, it becomes feasible to use sampling based methods which select Sk according to P(SkIDn, W). 3 Inferring the Internal Representation Given the input data, finding its most likely explanation is an inference process. Although it is simple to calculate the probability of any particular network state, there is no simple way to determine the most probable state given input D. A general approach to this problem is Gibbs sampling (Pearl, 1988; Neal, 1992). Bayesian Unsupervised Learning of Higher Order Structure 531 In Gibbs sampling, each state Si of the network is updated iteratively according to the probability of Si given the remaining states in the network. This conditional probability can be computed using P(SiISj : j::l i, W) oc P(Silpa[Si], W) II P(Sjlpa[Sj],Si' W) (5) jEch[S.] where ch[Sil indicates the children of state Si. In the limit, the ensemble of states obtained by this procedure will be typical samples from P(SID, W). More generally, any subset of states can be fixed and the rest sampled. The Gibbs equations have an interpretation in terms of a stochastic recurrent neural network in which feedback from higher levels influences the states at lower levels. For the models defined here, the probability of Si changing state given the remaining states is (6) The variable D-xi indicates how much changing the state Si changes the probability of the network state ~xi=logh(ui;l-Si)-logh(ui;Si)+ L logh(uj+bij;Sj)-logh(uj;Sj) (7) jEch[S;] where h(u; a) = h(u) if a = 1 and 1 - h(u) if a = o. The variable Ui is the causal input to Si, given by l:k SkWki. The variable bj specifies the change in Uj for a change in Si: bij = +SjWij if Si = 0 and -SjWij if Si = 1. The first two terms in (7) can be interpreted as the feedback from higher levels. The sum can be interpreted as the feedforward input from the children of Si. Feedback allows the lower level units to use information only computable at higher levels. The feedforward terms typically dominate the expression, but the feedback becomes the determining factor when the feedforward input is ambiguous. For general distributions, Gibbs sampling can require many samples to achieve a representative samples. But if there is little ambiguity in the internal representation, as is the goal, Gibbs sampling can be as efficient as a single feedforward pass. One potential problem is that Gibbs sampling will not work before the weights have adapted, when the representations are highly ambiguous. We show below, however, that it is not necessary to sample for long periods in order for good representations to be learned. As learning proceeds, the internal representations obtained with limited Gibbs sampling become increasingly accurate. 4 Adapting the Weights The complexity of the model is controlled by placing a prior on the weights. For the form of the noisy OR function in which all weights are constrained to be positive, we assume the prior to be the product of independent gamma distributions parameterized by 0: and {3. The objective function becomes C = P(D 1:N IW)P(Wlo:, {3) (8) A simple and efficient EM-type formula for adapt the weights can be derived by setting aCjwij to zero and solving for Wij. Using the transformations iij = 1 532 M. S. Lewicki and T. 1. Sejnowski exp( -Wij) and 9i = 1 - exp( -Ui) we obtain (9) where s(n) is the state obtained with Gibbs sampling for the nth input pattern. The variable Iij can be interpreted as the frequency of state Sj given cause Si. The sum in the above expression is a weighted average of the number of times Sj was active when Si was active. The ratio hj /9j weights each term in the sum inversely according to the number of different causes for Sj. If Si is the unique cause of Sj then hj = 9j and the term would have full weight. A straightforward application of the learning algorithm would adapt all the weights at the same time. This does not produce good results, however, because there is nothing to prevent the model from learning overly strong priors. This can be prevented by adapting the weights in the upper levels after the weights in the lower levels have stabilized. This allows the higher levels to adapt to structure that is actually present in the data. We have obtained good results from both the naive method of adapting the lowest layers first and from more sophisticated methods where stability was based on how often a unit changed during the Gibbs sampling. 5 Examples In the following examples, the weight prior was specified with a = 1.0 and {3 = 1.5. Weights were set to random values between 0.05 and 0.15. Gibbs sampling was stopped if the maximum state change probability was less than 0.05 or after 15 sweeps through the units. Weights were reestimated after blocks of 200 patterns. Each layer of weights was adapted for 10 epochs before adapting the next layer. A High Order Lines Problem. The first example illustrates that the algorithm can discover the underlying features in complicated patterns and that the higher layers can capture interesting higher order structure. The first dataset is a variant of the lines problem proposed by Foldiak (1989). The patterns in the dataset are composed of horizontal and vertical lines as illustrated in figure 1. Note that, although B Figure 1: Dataset for the high order lines problem. (A) Patterns are generated by selecting one of the pattern types according to the probabilities next to the arrows. Top patterns are copied to the input. The horizontal and vertical lines on the left are selected with probability 0.3. (B) Typical input patterns. Bayesian Unsupervised Learning of Higher Order Structure 533 the datasets are displayed on a 2-D grid, the network makes no assumptions about topography. Because the network is fully connected, all spatial arrangements of the inputs are identical. The weights learned by the network are shown in figure 2. I'~ 1.111:1 ~·I·II:I:l:I:I:II.I·~tl.II:~:I·I:1 __ -.... IIlImmll 1111111111 Figure 2: The weights in a 25-10-5 network after training. Blocks indicate the weights to each unit. Square size is proportional to the weight values. Second layer units capture the structure of the horizontal and vertical lines. Third layer units capture the correlations among the lines. The first unit in the third layer is active when the 'II' is present. The second, fourth, and fifth units have learned to represent the '+', '=', and '0' respectively, with the remaining unit acting as a bias. The Shifter Problem. The shifter problem (Hinton and Sejnowski, 1986), explained in figure 3, is important because the structure that must be discovered is in the higher order input correlations. This example also illustrates the importance of allowing high level states to influence low level states to determine the most probable internal representation. The units in the second layer can only capture second order statistics and cannot determine the direction of the shift. The only way these units can be disambiguated is to use the feedback from the units in the third layer which detect the direction of the shift by integrating the output of the units in the second layer. This allows the representation in the second layer to be "cleaned up" and makes it easier to discover the higher order structure of the global shift. The speed and reliability of the learning was tested by learning from random initial conditions. The results are shown in figure 4. Note that the best solutions have a cost of about one bit higher than the optimal cost of less than 9 bits, because top units cannot capture the fact that they are mutually exclusive. 6 Discussion The methods we have described work well on these simple benchmark problems and scale well to larger problems such as the handwritten digits example used in (Hinton et al., 1995). We believe there are two main reasons why the algorithm described here runs considerably faster than other Gibbs sampling based methods. The first is that there is no need to collect state statistics for each pattern. The weight values are reestimated using just one sampled internal state per pattern. The second is that weights that are not connected to informative units are not updated. This prevents the model from learning what are effectively overly strong priors and allows the weights in upper layers to adapt to structure actually in the data. Gibbs sampling allows internal representations to be selected according to their true posterior probability. This was shown to be effective in cases where the resulting 534 M. S. Lewicki and T. 1. Sejnowski A B m··· _· .. .. . .. . . . . .. . . ' ... . . . . . . . ---------------Figure 3: The shifter problem. (A) Input patterns are generated by generating a random binary vector in the bottom row. This pattern is shifted either left or right (with wrap around) with equal probability and copied to the top row. The input rows are duplicated to add redundancy as in Dayan et al. (1995). (B) The weights of a 32-20-2 after learning. The second layer of units learn to detect either local left shifts or right shifts in the data. These units cannot determine the shift direction alone, however, and require feedback from third layer units which integrate the outputs of all the units that represent a common shift (note that there is no overlap in the weights for the two third-layer units). This feedback turns off units that are inconsistent with the direction of shift. The weights that are close to zero for both third layer units effectively remove redundant second layer units that are not required to represent the input patterns. ~w-----~------~----~-------.------. ~~-----I~O------~15------~ro~-----2~5----~~ Number of Epochs Figure 4: The graph shows 10 runs on the shifter problem from random initial conditions. The average bits per pattern is computed by -log(C)/(Nlog2). Each epoch used 200 input randomly generated input patterns. Two additional epochs were performed with 1000 random patterns to obtain accurate estimates of the average bits per pattern. The network converges rapidly and reliably. The best solutions, like the one shown in figure 3b, were found in 4/10 runs and had costs of approximately 10 bits at epoch 30. In this example, the network can get caught in local minima if too many units learn to represent the same local shifts. Bayesian Unsupervised Learning of Higher Order Structure 535 representation has little ambiguity, i.e. each pattern has only a small number of probable explanations. If the causal structure to be learned is inherently ambiguous, e.g. in modeling the causal structure of medical symptoms, Gibbs sampling will be slow and better performance can be obtained with wake-sleep learning (Hinton et al., 1995; Frey et al., 1995) or mean field approximations (Saul et al., 1996). There are many natural situations when there is ambiguity in low level features. This ambiguity can only be resolved by integrating the contextual information which itself is derived from the ambiguous simple features. This problem is common in the case of noisy input patterns and in feature grouping problems such as figureground separation. Feedback is crucial for ensuring that low-level representations are consistent within the larger context. Some systems, such as the Helmholtz machine (Dayan et al., 1995; Hinton et al., 1995), arrive at the internal state through a feedforward process. It possible that this ambiguity in lower-level representations could be resolved by circuitry in the higherlevel representations, but if multiple higher-level modules make use of the same lowlevel representations, the additional circuitry would have to be duplicated in each module. It seems more parsimonious to use feedback to influence the formation of the lower-level representations. References Dayan, P., Hinton, G. E., Neal, R. M., and Zemel, R. S. (1995). The Helmholtz machine. Neural Computation, 7:889-904. F61dilik, P. (1989). Adaptive network for optimal linear feature extraction. In Proceedings of the International Joint Conference on Neural Networks, volume I, pages 401-405, Washington, D. C. Frey, B. J., Hinton, G. E., and Dayan, P. (1995). Does the wake-sleep algorithm produce good density estimators? In Touretzky, D. S., Mozer, M., and Hasselmo, M., editors, Advances in Neural Information Processing Systems, volume 8, pages 661-667, San Mateo. Morgan Kaufmann. Hinton, G. E., Dayan, P., Frey, B. J ., and Neal, R. M. (1995). The wake-sleep algorithm for unsupervised neural networks. Science, 268(5214):1158-116l. Hinton, G. E. and Sejnowski, T. J. (1986). Learning and relearning in Boltzmann machines. In Rumelhart, D. E. and McClelland, J. L., editors, Parallel Distributed Processing, volume 1, chapter 7, pages 282-317. MIT Press, Cambridge. Lauritzen, S. L. and Spiegelhalter, D. J. (1988). Local computations with probabilities on graphical structures and their application to expert systems. J. Royal Statistical Soc. Series B Methodological, 50(2):157- 224. Neal, R. M. (1992). Connectionist learning of belief networks. Artificial Intelligence, 56(1):71-113. Pearl, J. (1988). Probabilistic Reasoning in Intelligent Systems. Morgan Kaufmann, San Mateo. Saul, L. K., Jaakkola, T., and Jordan, M. I. (1996). Mean field theory for sigmoid belief networks. J. Artificial Intelligence Research, 4:61-76.
1996
72
1,272
An Architectural Mechanism for Direction-tuned Cortical Simple Cells: The Role of Mutual Inhibition Silvio P. Sabatini silvio@dibe.unige.it Fabio Solari fabio@dibe.unige.it Giacomo M. Bisio bisio@dibe.unige.it Department of Biophysical and Electronic Engineering PSPC Research Group Genova, 1-16145, Italy Abstract A linear architectural model of cortical simple cells is presented. The model evidences how mutual inhibition, occurring through synaptic coupling functions asymmetrically distributed in space, can be a possible basis for a wide variety of spatio-temporal simple cell response properties, including direction selectivity and velocity tuning. While spatial asymmetries are included explicitly in the structure of the inhibitory interconnections, temporal asymmetries originate from the specific mutual inhibition scheme considered. Extensive simulations supporting the model are reported. 1 INTRODUCTION One of the most distinctive features of striate cortex neurons is their combined selectivity for stimulus orientation and the direction of motion. The majority of simple cells, indeed, responds better to sinusoidal gratings that are moving in one direction than to the opposite one, exhibiting also a narrower velocity tuning with respect to that of geniculate cells. Recent theoretical and neurophysiological studies [1] [2] pointed out that the initial stage of direction selectivity can be related to the linear space-time receptive field structure of simple cells. A large class of simple cells has a very specific space-time behavior in which the spatial phase of the receptive field changes gradually as a function of time. This results in receptive field profiles that are tilted in the space-time domain. To account for the origin of this particular spatio-temporal inseparability, numerous models have been proposed postulating the existence of structural asymmetries of the geniculo-cortical projections both in the temporal and in the spatial domains (for a review, see [3] An Architectural Mechanism/or Direction-tuned Cortical Simple Cells inhibitory [~i: h01 pool ! : m Layer 1 ' T oJ eq Layer 2 {;l4.\D------t ez (a) eo ge~iculate mput I ~ el k1d + ~ k1a m KZd I I~I !ez (b) 105 eO Figure 1: (a) A schematic neural circuitry for the mutual inhibition; (b) equivalent block diagram representation. [4]). Among them, feed-forward inhibition along the non-preferred direction, and the combination of lagged and non-lagged geniculate inputs to the cortex have been commonly suggested as the major mechanisms. In this paper, within a linear field theory framework, we propose and analyse an architectural model for dynamic receptive field formation, based on intracortical interactions occurring through asymmetric mutual inhibition schemes. 2 MODELING INTRACORTICAL PROCESSING The computational characteristics of each neuron are not independent of the ones of other neurons laying in the same layer, rather, they are often the consequence of a collective behavior of neighboring cells. To understand how intracortical circuits may affect the response properties of simple cells one can study their structure and function at many levels of organization, from subcellular, driven primarily by biophysical data, to systemic, driven by functional considerations. In this study, we present a model at the intermediate abstraction level to combine both functional and neurophysiological descriptions into an analytic model of cortical simple cells. 2.1 STRUCTURE OF THE MODEL Following a linear neural field approach [5] [6], we regard visual cortex as a continuous distribution of neurons and synapses. Accordingly, the geniculo-cortical pathway is modeled by a multi-layer network interconnected through feed-forward and feedback connections, both inter- and intra-layers. Each location on the cortical plane represents a homogeneous population of cells, and connections represent average interactions among populations. Such connections can be modeled by spatial coupling functions which represent the spread of the synaptic influence of a 106 S. P. Sabatini, F. Solari and G. M. Bisio population on its neighbors, as mediated by local axonal and dendritic fields. From an architectural point of view, we assume the superposition of feed-forward (i.e., geniculate) and intracortical contributions which arise from inhibitory pools whose activity is also primed by a geniculate excitatory drive. A schematic diagram showing the "building blocks" of the model is depicted in Fig. 1. The dynamics of each population is modeled as first-order low-pass filters characterized by time constants T'S . For the sake of simplicity, we restrict our analysis to 1-0 case, assuming that such direction is orthogonal to the preferred direction of the receptive field [7]. This 1-0 model would produce spatia-temporal results that are directly compared with the spatio-temporal plots usually obtained when an optimal stimulus is moved along the direction orthogonal to the preferred direction of the receptive field. Geniculate contributions eo(x, t) are modeled directly by a spatiotemporal convolution of the visual input s(x, t) and a separable kernel ho(x, t) characteri~d in the spatial domain by a Gaussian shape with spatial extent 0'0 and, in the temporal domain, by a first-order temporal low-pass filter with time constant TO. The output el(x, t) of the inhibitory neuron population results from the mutual inhibitory scheme through spatially organized pre- and post-synaptic sites, modeled by the kernels kIa(x -~) and kId(X ~), respectively: del(X,t) J Tl dt =-el(x,t)+ kId(x-~)[eo(-~,t)-bm(-~,t)]d~ (1) m(x,t) = J k~a(X-~)el(-~)~ (2) where the function m(x, t) describes the spatia-temporal mutual inhibitory interactions, and b is the inhibition strength. The layer 2 cortical excitation e2(x, t) is the result of feed-forward contributions collected (k2d) from the inhibitory loop, at axonal synaptic sites, and the geniculate input (eo(x, t)) . To focus the attention on the inhibitory loop, in the following we assume a one-to-one mapping from layer 1 to layer 2, i.e., k2d(x -~) = 6(x ~), consequently: de2( x, t) T2 dt = -e2(x, t) + eo(x, t) - bm(x, t) (3) where TI and T2 are the time constants associated to layer 1 and layer 2, respectively. 2.2 AVERAGE INTRACORTICAL CONNECTIVITY When assessing the role of intracortical circuits on the receptive field properties of cortical cells, one important issue concerns the spatial localization of inhibitory and excitatory influences. In a previous work [8] we evidenced how the steadystate solution of Eqs. (1 )-(3) can give rise to highly structured Gabor-like receptive field profiles, when inhibition arises from laterally distributed clusters of cells. In this case, the effective intrinsic kernel kl(x), defined as kl(x _~) ~f J J k1a( -x', -e)k1d( x - x', ~ - e)dx'de, can be modeled as the sum of two Gaussians symmetrically offset with respect to the target cell (see Fig. 2): 1 ( WI [ 2 2 W2 [ 2 2 ) k1(x) = z= exp -(x - dl ) /20'd + exp -(x + d2) /20'2] . v2~ ~ 0'2 (4) This work is aimed to investigate how spatial asymmetries in the intracortical coupling function lead to non-separable space-time interactions within the resulting discharge field of the simple cells. To this end, we varied systematically the geometrical parameters (0', W, d) of the inhibitory kernel to consider three different An Architectural Mechanism/or Direction-tuned Cortical Simple Cells 107 w Figure 2: The basic inhibitory kernel used k1(x -e). The cell in the center receives inhibitory contributions from laterally distributed clusters of cells. The asymmetric kernels used in the model derive from this basic kernel by systematic variations of its geometrical parameters (see Table 1). types of asymmetries: (1) different spatial spread of inhibition (i.e., 0'1 i= 0'2); (2) different amount of inhibition (W1 i= W2); (3) different spatial offset (d1 i= d2). A more rigorous treatment should take care also of the continuous distortion of the topographic map [9]. In our analysis this would result in a continuous deformation of the inhibitory kernel, but for the small distances within which inhibition occurs, the approximation of a uniform mapping produces only a negligible error. Architectural parameters were determined from reliable measured values of receptive fields of simple cells [10] [11]. Concerning the spatial domain, we fixed the size (0'0) of the initial receptive field (due to geniculate contributions) for an "average" cortical simple cell with a resultant discharge field of"" 20 ; and we adjusted, accordingly, the parameters of the inhibitory kernel in order to account for spatial interactions only within the receptive field. Considering the temporal domain, one should distinguish the time constant T1, caused by network interactions, from the time constants TO and T2 caused by temporal integration at a single cell membrane. In any case, throughout all simulations, we fixed TO and T2 to 20ms, whereas we"varied T1 in the range 2 - lOOms. 3 RESULTS Since visual cortex is visuotopically organized, a direct correspondence exists between the spatial organization of intracortical connections and the resulting receptive field topography. Therefore, the dependence of cortical surface activity e2(x, t) on the visual input sex, t) can be formulated as e2(x, t) = h(x,t) * sex, t), where the symbol * indicates a spatia-temporal convolution, and hex, t) is the equivalent receptive field interpreted as the spatia-temporal distribution of the signs of all the effects of cortical interactions. In this context, hex, t) reflects the whole spatio-temporal couplings and not only the direct neuroanatomical connectivity. To test the efficacy of the various inhibitory schemes, we used a drifting sine wave grating s(x,t) = Ccos[211'{fxx ± Itt)] where C is the contrast, Ix and It are the spatial and temporal frequency, respectively. The direction selectivity index (DSI) and the optimal velocity (vopt) obtained from the various inhibitory kernels of Fig.2 are summarized in Table 1, for different values of T1 and b. The direction selectivity index is defined as DS] = ~:~~::, where Rp is the maximum response amplitude for preferred direction, and Rnp is the maximum amplitude for non-preferred direction. The optimal velocity is defined as I;pt / lx, where Ix is chosen to match the spatial structure of the receptive field, and I?t is the frequency which elicits the maximum cell's response. As expected, increasing the parameter b enhances the effects of inhibition, thus resulting in larger DSI and higher optimal velocities. However, for stability reason, b should "remain below a theshold value strictly reJ08 S. P. Sabatini, F. Solari and G. M. Bisio Table 1: T1 = 2 Tl = 10 Tl = 20 T1 = 100 II b II DSI I Vopt II DSI I vopt II DSI I Vopt II DSI I Vopt II ASY-IA 0.25 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.60 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.80 0.08 1.82 0.24 1.82 0.00 0.00 0.00 0.00 0.91 0.17 1.82 0.34 1.82 0.00 0.00 0.00 0.00 0.50 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 ASY-lB 0.85 0.04 1.82 0.17 1.82 0.25 1.82 0.00 0.00 1.00 0.06 1.82 0.19 1.82 0.28 1.82 0.00 0.00 1.30 0.07 1.82 0.28 1.82 0.37 1.82 0.00 0.00 0.50 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 ASY-2A 1.00 0.00 0.00 0.00 0.00 0.09 1.82 0.16 1.82 5.00 0.02 1.82 0.05 1.82 0.09 1.82 0.39 3.64 9.00 0.01 1.82 0.03 1.82 0.06 1.82 0.38 3.64 0.25 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 ASY-2B 0.50 0.06 2.07 0.20 2.07 0.32 2.07 0.00 0.00 0.60 0.06 2.07 0.39 4.14 0.40 2.07 0.00 0.00 0.72 0.07 2.00 0.66 6.00 0.65 4.00 0.00 0.00 0.25 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.60 0.00 00.0 0.00 0.00 0.00 0.00 0.00 0.00 0.80 0.08 1.82 0.23 1.82 0.00 0.00 0.00 0.00 ASY-3A A ::::,. ;. t$LA 0.88 0.14 1.82 0.26 1.82 0.00 0.00 0.00 0.00 0.50 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 ASY-3B 0.85 0.04 1.82 0.16 1.82 0.23 1.82 0.00 0.00 1.00 0.05 1.82 0.18 1.82 0.26 1.82 0.00 0.00 1.33 0.06 1.82 0.26 1.82 0.35 1.82 0.00 0.00 lated to the inhibitory kernel considered. Moreover, we observe that, except for ASY-2A, the strongest direction selectivity can be obtained when the intracortical time constant Tl· has values in the range of 10 - 20 ms, i.e., comparable to TO and T2. Larger values of Tl would result, indeed, in a recrudescence of the velocity low-pass behavior. For each asymmetry, Figs. 3 show the direction tuning curves and the x-t plots, respectively, for the best cases considered (cf. bold-faced values in Table 1). We have evidenced that appreciable DSI can be obtained when inhibition arises from cortical sites at different distance from the target cell (i.e., ASY-2B, dl i= d2). In such conditions we obtained a DSI as high as 0.66 and an optimal velocity up to '" 6°/s, as could be inferred also from the spatia-temporal plot which present a marked motion-type (i.e., oriented) non-separability (see Fig. 3ASY-2B). 4 DISCUSSION AND CONCLUSIONS As anticipated in the Introduction, direction selectivity mechanisms usually relies upon asymmetric alteration of the spatial and temporal response characteristics of the geniculate input, which are presumably mediated by intracortical circuits. In the architectural model presented in this study, spatial asymmetries were included explicitly in the extension of the inhibitory interconnections, but no explicit asymmetric temporal mechanisms were introduced. It is worth evidencing how temporal asymmetries originate from the specific mutual inhibition scheme considered, which operates, regarding temporal domain, like a quadrature model [12] [13]. This can An Architectural Mechanism/or Direction-tuned Cortical Simple Cells 109 " co .. ~ 1.0 ~ 1.0 ASY-IB ~ 1.0 0 & 0 e- 0.8 :l O.B g. 0.8 ~ co " 0.6 ~ 0.6 i 06 ..... _--_. __ .' ... ,. co co ----" " -----' . • !::! 0.4 ~ 0.4 ~ 0.4 "6 0 '. 0 ". E 0.2 E 0.2 E 0.2 0 0 0 Z 00 Z 0.0 Z 0.0 0 . 1 1 10 100 0.1 1 10 100 0 .1 1 10 100 Velocity deg/s Velocity deg/s Velocity deg/s " co co ~ 1.0 ~ 1.0 ASY-3A f! 1.0 0 0 0 ~ 0.8 ~ 0.8 -.. ~ ........ -. a- 0.8 " ~ ~ ~ 0.6 .., 0.6 " 0.6 co co " .!::' 0.4 .!::! 0.4 ~ 0.4 '0 '0 0 E 0.2 E 0.2 E 0.2 0 0 0 Z 0 .0 Z 0.0 Z 0.0 0.1 1 10 100 0. 1 1 10 100 0. 1 1 10 100 Velocity deg/s Velocity deg/s Velocity deg/s 400 400 400 ......... '<il '<il en e: e: e: "-' "-' "-' s ~ ~ S S ':;j ':;j ':;j 0 0 0 0 space (deg) 2 0 space (deg) 2 0 space (deg) 2 400 400 400 ......... ......... ......... en en en S S S "-' '-' '-' s s ~ s 'J: ':;j 'J: 0 0 0 0 space (deg) 2 0 space (deg) 2 0 space (deg) 2 Figure 3: Results from the model related to the bold-typed values indicated in Table 1, for each asymmetry considered. (Top) direction tuning curve; (Bottom) spatia-temporal plots. We can evidence a marked direction tuning for ASY-2B, i.e., when inhibition arises from two differentially offset Gaussians be inferred by the examination of the equivalent transfer function H (w x , Wt) in the Fourier domain: H( ) Ho(wx ,wt) ( 1 . 1 ) Wx , Wt =. . + JWt T1 -~.:-----:---:--...,... 1 + JWtT2 1 + JWtT1 + bK1(wx ) 1 + JWtT1 + bK1(wx ) (5) where upper case letters indicate Fourier transforms, and j is the complex variable. The terms in parentheses in Eq. (5) can be interpreted as the sum of temporal components that are approximately arranged in temporal quadrature. Furthermore, one can observe that a direct monosynaptic influence (el) from the inhibitory neurons of layer 1 to the excitatory cells of layer 2, would result in the cancellation of the quadrature component in Eq. (5). Further improvement of the model should take into account also transmission delays between spatially separated interacting cells, theshold non-linearities, and ON-OFF interactions. 110 S. P. Sabatini, F. Solari and G. M. Bisio Acknowledgements This research was partially supported by CEC-Esprit CORMORANT 8503, and by MURST 40%-60%. References [1] R.C. Reid, R.E. Soodak, and R.M. Shapley. Directional selectivity and spatiotemporal structure of receptive fields of simple cells in cat striate cortex. 1. Neurophysiol., 66:505-529, 1991. [2] D.B. Hamilton, D.G. Albrecht, and W.S. Geisler. Visual cortical receptive fields in monkey and cat: spatial and temporal phase transfer function. Vision Res., 29(10):1285-1308, 1989. [3] K. Nakayama. Biological image motion processing: a review. Vision Res., 25:625-660, 1985. [4] E.C. Hildreth and C. Koch. The analysis of visual motion: From computational theory to neuronal mechanisms. Ann. Rev. Neurosci., 10:477-533, 1987. [5] G. Krone, H. Mallot, G. Palm, and A. Schiiz. Spatiotemporal receptive fields: A dynamical model derived from cortical architectonics. Proc. R. Soc. London Bioi, 226:421-444, 1986. [6] H.R. Wilson and J.D. Cowan. A mathematical theory of the functional dynamics of cortical and thalamic nervous tissue. Kibernetik, 13:55-80, 1973. [7] G.C. De Angelis, I. Ohzawa, and R.D. Freeman. Spatiotemporalorganization of simple-cell receptive fields in the cat's striate cortex.!. General characteristics and postnatal development. J. Neurophysiol., 69:1091-1117, 1993. [8] S.P. Sabatini, 1. Raffo, and G.M. Bisio. Functional periodic intracortical couplings induced by structured lateral inhibition in a linear cortical network. Neural Computation, 9(3):525-531, 1997. [9] H.A. Mallot, W. von Seelen, and F. Giannakopoulos. Neural mapping and space variant image processing. Neural Networks, 3:245-263, 1990. [10] K. Albus. A quantitative study of the projection area of the central and the paracentral visual field in area 17 of the cat. Exp. Brain Res., 24:159-202, 1975. [11] J . Jones and 1. Palmer. The two-dimensional spatial structure of simple receptive fields in cat striate cortex. J. Neurophysiol., 58:1187-1211, 1987. [12] A.B. Watson and A.J. Ahumada. Model of human visual-motion sensing. J. Opt. Soc. Amer., 2:322-341, 1985. [13] E.H. Adelson and J .R. Bergen. Spatiotemporal energy models for the perception of motion. J. Opt. Soc. Amer., 2:284-321, 1985.
1996
73
1,273
Complex-Cell Responses Derived from Center-Surround Inputs: The Surprising Power of Intradendritic Computation Bartlett W. Mel and Daniel L. Ruderman Department of Biomedical Engineering University of Southern California Los Angeles, CA 90089 Kevin A. Archie Neuroscience Program University of Southern California Los Angeles, CA 90089 Abstract Biophysical modeling studies have previously shown that cortical pyramidal cells driven by strong NMDA-type synaptic currents and/or containing dendritic voltage-dependent Ca++ or Na+ channels, respond more strongly when synapses are activated in several spatially clustered groups of optimal size-in comparison to the same number of synapses activated diffusely about the dendritic arbor [8]- The nonlinear intradendritic interactions giving rise to this "cluster sensitivity" property are akin to a layer of virtual nonlinear "hidden units" in the dendrites, with implications for the cellular basis of learning and memory [7, 6], and for certain classes of nonlinear sensory processing [8]- In the present study, we show that a single neuron, with access only to excitatory inputs from unoriented ON- and OFF-center cells in the LGN, exhibits the principal nonlinear response properties of a "complex" cell in primary visual cortex, namely orientation tuning coupled with translation invariance and contrast insensitivity_ We conjecture that this type of intradendritic processing could explain how complex cell responses can persist in the absence of oriented simple cell input [13]84 B. W. Mel, D. L. Ruderman and K. A. Archie 1 INTRODUCTION Simple and complex cells were first described in visual cortex by Hubel and Wiesel [4]. Simple cell receptive fields could be subdivided into ON and OFF subregions, with spatial summation within a subregion and antagonism between subregions; cells of this type have historically been modeled as linear filters followed by a thresholding nonlinearity (see [13]). In contrast, complex cell receptive fields cannot generally be subdivided into distinct ON and OFF subfields, and as a group exhibit a number of fundamentally nonlinear behaviors, including (1) orientation tuning across a receptive field much wider than an optimal bar, (2) larger responses to thin bars than thick bars-in direct violation of the superposition principle, and (3) sensitivity to both light and dark bars across the receptive field. The traditional Hubel-Wiesel model for complex cell responses involves a hierarchy, consisting of center-surround inputs that drive simple cells, which in turn provide oriented, phase-dependent input to the complex cell. By pooling over a set of simple cells with different positions and phases, the complex cell could respond selectively to stimulus orientation, while generalizing over stimulus position and contrast. A pure hierarchy involving simple cells is challenged, however, by a variety of more recent experimental results indicating many complex cells receive monosynaptic input from LGN cells [3], or do not depend on simple cell input [10, 5, 1]. It remains unknown how complex cell responses might derive from intracortical network computations that do no depend on simple cells, or whether they could originate directly from intracellular computations. Previous biophysical modeling studies have indicated that the input-output function of a dendritic tree containing excitatory voltage-dependent membrane mechanisms can be abstracted as low-order polynomial function, i.e. a big sum of little products (see [9] for review). The close match between this type of computation and "energy" models for complex cells [12, 11, 2] suggested that a single-cell origin of complex cell responses was possible. In the present study, we tested the hypothesis that local nonlinear processing in the dendritic tree of a single neuron, which receives only excitatory synaptic input from unoriented center-surround LGN cells, could in and of itself generate nonlinear complex cell response properties, including orientation selectivity, coupled with position and contrast invariance. 2 METHODS 2.1 BIOPHYSICAL MODELING Simulations of a layer 5 pyramidal cell from cat visual cortex (fig. 1) were carried out in NEURON1 . Biophysical parameters and other implementation details were as in [8] andj or shown in Table 2, except dendritic spines were not modeled here. The soma contained modified Hodgkin-Huxley channels with peak somatic conductances of UNa and 90R 0.20 Sjcm2 and 0.12 Sjcm2, respectively; dendritic membrane was electrically passive. Each synapse included both an NMDA and AMPA-type INEURON simulation environment courtesy Michael Hines and John Moore; synaptic channel implementations courtesy Alan Destexhe and Zach Mainen. Complex-cell Responses Derived from Center-surround Inputs 85 • Figure 1: Layer 5 pyramidal neuron used in the simulations, showing 100 synaptic contacts. Morphology courtesy Rodney Douglas and Kevan Martin. excitatory conductances (see Table 1). Conductances were scaled by an estimate of the local input resistance, to keep local EPSP size approximately uniform across the dendritic tree. Inhibitory synapses were not modeled. 2.2 MAPPING VISUAL STIMULI ONTO THE DENDRITIC TREE A stimulus image consisted of a 64 x 64 pixel array containing a light or dark bar (pixel value ±1 against a background of 0). Bars of length 45 and width 7 were presented at various orientations and positions within the image. Images were linearly filtered through difference-of-Gaussian receptive fields (center width: 0.6, surround width: 1.2, with no DC response). Filtered images were then mapped onto 64 x 64 arrays of ON-center and OFF-center LGN cells, whose outputs were thresholded at ±0.02 respectively. In a crude model of gain control, only a random subset of 100 of the LGN neurons remained active to drive the modeled cortical cell. Each LGN neuron gave rise to a single synapse onto the cortical cell's dendritic tree. In a given run, excitatory synapses originating from the 100 active LGN cells were activated asynchronously at 40 Hz, while all other synapses remained silent. The spatial arrangement of connections from LGN cells onto the pyramidal cell dendrites was generated automatically, such that pairs of LGN cells which are coactive during presentations of optimally oriented bars formed synapses at nearby sites in the dendritic tree. The activity of the LGN cell array to an optimally oriented bar is shown in fig. 3. Frequently co-activated pairs of LGN neurons are hereafter referred to as "friend-pairs", and lie in a geometric arrangement as shown in fig. 4. Correlation-based clustering of friend-pairs was achieved by (1) choosing a random LGN cell and placing it at the next available dendritic site, (2) randomly 86 B. W. Me~ D. L. Ruderman and K. A. Archie Parameter Value Rm lOkncm~ Ra 200ncm em 1.0JLF/cm~ Vrest -70 mV Somatic 9Na 0.20 S/cm~ Somatic !/DR 0.12 S/cm~ Synapse count 100 Stimulus frequency 40 Hz TAMPA (on, off) 0.5 ms, 3 ms 9AMPA 0.27 nS - 2.95 nS 'TNMDA(on, of f) 0.5 ms, 50 ms 9NMDA 0.027 nS - 0.295 nS Esyn OmV Figure 2: Table 1. Simulation Parameters. choosing one of its friends and placing it at the next available dendritic site, and so on, until until either all of the cell's friends had already been deployed, in which case a new cell was chosen at random to restart the sequence, or all cells had been chosen, meaning that all of the 8192 (= 64 x 64 x 2) LGN synapses had been successfully mapped onto the dendritic tree. In previous modeling work it was shown that this type of clustering of correlated inputs on dendrites is the natural outcome of a balance between activity-independent synapse formation, and activity dependent synapse stabilization [6]. This method guaranteed that an optimally oriented bar stimulus activated a larger number of friend-pairs on average than did bars at non-optimal orientations. This led in turn to relatively clustery distributions of activated synapses in the dendrites in response to optimal bar orientations, in comparison to non-optimal orientations. In previous work, it was shown that synapses activated in clusters about a dendritic arbor could produce significantly larger cell responses than the same number of synapses activated diffusely about the dendritic tree [7, 8]. 3 Results Results for two series of runs are shown in fig. 5. For each bar stimulus, average spike rate was measured over a 250 ms period, beginning with the first spike initiated after stimulus onset (if any). This measure de-emphasized the initial transient climb off the resting potential, and provided a rough steady-state measure of stimulus effectiveness. Spike rates for 30 runs were averaged for each input condition. Orientation tuning curves for a thin bar (7 x 45 pixels) are shown in fig. 5. The orientation tuning peaks sharply within about 10° of vertical, and then decays slowly for larger angles. Tuning is apparent both for dark and light bars, and remains independent of location within the receptive field. 4 Discussion The results of fig. 5 indicate that a pyramidal cell driven exclusively by excitatory inputs from ON- and OFF-center LGN cells, is at a biophysical level capable of producing the hallmark nonlinear response property of visual complex cells. Furthermore, the cell's translation-invariant preference for light or dark vertical bars was established by manipulating only the spatial arrangement of connections from LGN cells onto the pyramidal cell dendrites. Since exactly 100 synapses were activated in every tested condition, the significantly larger responses to optimal bar orientations could not be explained by a simple elevation in the total synaptic activity impinging on the neuron in that condition. The origin of the cell's orientation-selective response resulted from nonlinear pooling of a large number of minimally-oriented subunits, i.e. consisting of pairs of ON and OFF cells that were co-consistent with an optimally oriented bar. We have achieved similar results in other experiments with a variety of different friend-neighborhood structures including ones both simpler and more complex than were used here, for LGN arrays with substantially different degrees of receptive field overlap, with random subsampling of the LGN array, with graded LGN activity levels, and for dendritic trees containing active sodium channels in addition to NMDA channels. Thus far we have not attempted to relate physiologically-measured orientation and width tuning curves, and other detailed aspects of complex cell physiology, to our model cell, as we have been principally interested in establishing whether the most salient nonlinear features of complex cell physiology were biophysically feasible at the single cell level. Detailed comparisons between our results and empirical tuning curves, etc., must be made with caution, since our model cell has been "explanted" from the network in which it normally exists, and is therefore absent the normal recurrent excitatory and inhibitory influences the cortical network provides. 88 B. W Me~ D. L. Ruderman and K. A. Archie w I eo oe w w I eo oe w Figure 4: Layout of friends for an ON-center LGN cell for vertically oriented thin bars (top). The linear friendship linkage for a given ideal vertical bar of width w was determined as follows. Suppose an LGN cell is chosen at random, e.g. an ONcenter cell at location (i,j) within the cell array. When a vertical bar is presented, LGN cells along the two vertical edges of the bar become active. The ON-center cell at position (i,j) is active to a light bar when it is in a column of cells just inside either edge of the bar. Those cells which are co-active under this circumstance are: (a) other on-center cells in the same vertical column, (b) on-center cells in vertical columns a distance w - 1 to the right and left (depending on the bar position), (c) off-center cells in columns a distance ±1 away (due to the negative-going edge adjacent), and (d) off-center cells a distance w to the right and left (due to the opposite edge). As "friend-pairs" we take only those LGN cells a distance ±(w -1) and ±w away. Those in the same and neighboring columns are not included. The friends of an off-center cell are shown in the bottom figure. It and its friends are optimally stimulated by bars of width w placed as shown. The width selected for our friend-pairs was w = 7, the same width as all bars presented as stimuli. Experimental validation of these simulation results would imply a significant change in our conception of the role of the single neuron in neocortical processing. Acknowledgments This work was funded by grants from the National Science Foundation and the Office of Naval Research. References [1] G.M. Ghose, R.D. Freeman, and I. Ohzawa. Local intracortical connections in the cats visual-cortex - postnatal-development and plasticity. J. Neurophysiol. , 72:1290- 1303,1994. [2] D.J. Heeger. Normalization of cell responses in cat striate cortex. Visual Neurosci., 9:181-197, 1992. [3] K.P. Hoffman and J. Stone. Conduction velocity of afferents to cat visual cortex: a correlation with cortical receptive field properties. Brain Res., 32:460466, 1971. [4] D.H. Hubel and T.N. Wiesel. Receptive fields, binocular interaction and functional architecture in the cat's visual cortex. J. Physiol., 160:106- 154, 1962. Complex-cell Responses Derived from Center-surround Inputs 70 60 50 30 " 20 . . ........ -£. '0 °0~--~,0--~~--~3~0---~L---~50--~60L---mL---~60--~~ Orientation (degrees) 89 Figure 5: Orientation tuning curves for the model neurons. 'X': light bars centered in the receptive field; diamonds: light bars displaced by 6 pixels horizontally; squares: dark bars centered in the receptive field; '+': dark bars displaced by 6 pixels. Standard errors on the data are about 5 spikes/sec. [5] J.G. Malpeli, C. Lee, H.D. Schwark, and T.G. Weyand. Cat area 17. I. Pattern of thalamic control of cortical layers. J. Neurophyiol., 46:1102-1119, 1981. [6] B.W. Mel. The clusteron: Toward a simple abstraction for a complex neuron. In J. Moody, S. Hanson, and R. Lippmann, editors, Advances in Neural Information Processing Systems, vol. 4, pages 35-42. Morgan Kaufmann, San Mateo, CA, 1992. [7] B.W. Mel. NMDA-based pattern discrimination in a modeled cortical neuron. Neural Computation, 4:502-516, 1992. [8] B.W. Mel. Synaptic integration in an excitable dendritic tree. J. Neurophysiol., 70(3):1086-1101, 1993. [9] B.W. Mel. Information processing in dendritic trees. Neural Computation, 6:1031- 1085, 1994. [10] J .A. Movshon. The velocity tuning of single units in cat striate cortex. J. Physiol. (Lond), 249:445-468, 1975. [11] I. Ohzawa, G.C. DeAngelis, and R.D Freeman. Stereoscopic depth discrimination in the visual cortex: Neurons ideally suited as disparity detectors. Science, 279:1037-1041, 1990. [12] D. Pollen and S. Ronner. Visual cortical neurons as localized spatial frequency filters. IEEE Trans. Sys. Man Cybero., 13:907- 916, 1983. [13] H.R. Wilson, D. Levi, L. Maffei, J. Rovamo, and R. DeValois. The perception of form: retina to striate cortex. In L. Spillman and J.s. Werner, editors, Visual perception: the neurophysiological foundations, pages 231-272. Academic Press, San Diego, 1990.
1996
74
1,274
Training Algorithms for Hidden Markov Models Using Entropy Based Distance Functions Yoram Singer AT&T Laboratories 600 Mountain Avenue Murray Hill, NJ 07974 singer@research.att.com Manfred K. Warmuth Computer Science Department University of California Santa Cruz, CA 95064 manfred@cse.ucsc.edu Abstract We present new algorithms for parameter estimation of HMMs. By adapting a framework used for supervised learning, we construct iterative algorithms that maximize the likelihood of the observations while also attempting to stay "close" to the current estimated parameters. We use a bound on the relative entropy between the two HMMs as a distance measure between them. The result is new iterative training algorithms which are similar to the EM (Baum-Welch) algorithm for training HMMs. The proposed algorithms are composed of a step similar to the expectation step of Baum-Welch and a new update of the parameters which replaces the maximization (re-estimation) step. The algorithm takes only negligibly more time per iteration and an approximated version uses the same expectation step as Baum-Welch. We evaluate experimentally the new algorithms on synthetic and natural speech pronunciation data. For sparse models, i.e. models with relatively small number of non-zero parameters, the proposed algorithms require significantly fewer iterations. 1 Preliminaries We use the numbers from 0 to N to name the states of an HMM. State 0 is a special initial state and state N is a special final state. Any state sequence, denoted by s, starts with the initial state but never returns to it and ends in the final state. Observations symbols are also numbers in {I, ... , M} and observation sequences are denoted by x. A discrete output hidden Markov model (HMM) is parameterized by two matrices A and B. The first matrix is of dimension [N, N] and ai,j (0:5: i :5: N - 1,1 :5: j :5: N) denotes the probability of moving from state i to state j. The second matrix is of dimension [N + 1, M] and bi ,k is the probability of outputting symbol k at state i. The set of parameters of an HMM is denoted by 0 = (A, B). (The initial state distribution vector is represented by the first row of A.) An HMM is a probabilistic generator of sequences. It starts in the initial state O. It then iteratively does the following until the final state is reached. If i is the current state then a next state j is chosen according to the transition probabilities out of the current state (row i of matrix A). After arriving at state j a symbol is output according to the output probabilities of that state (row j of matrix B). Let P(x, slO) denote the probability (likelihood) that an HMM 0 generates the observation sequence x on the path s starting at state 0 and ending at state N: P(x, sllsl = Ixl + 1, So = 0, slSI = N, 0) ~ I1~~ll as._t,s.bs.,x •. For the sake of brevity we omit the conditions on s and x. Throughout the paper we assume that the HMMs are absorbing, that is from every state there is a path to the final state with a 642 Y. Singer and M. K. Warmuth non-zero probability. Similar parameter estimation algorithms can be derived for ergodic HMMs. Absorbing HMMs induce a probability over all state-observation sequences, i.e. Ex,s P(x, s18) = 1. The likelihood of an observation sequence x is obtained by summing over all possible hidden paths (state sequences), P(xI8) = Es P(x, sI8). To obtain the likelihood for a set X of observations we simply mUltiply the likelihood values for the individual sequences. We seek an HMM 8 that maximizes the likelihood for a given set of observations X, or equivalently, maximizes the log-likelihood, LL(XI8) = r:h EXEX In P(xI8). To simplify our notation we denote the generic parameter in 8 by Oi, where i ranges from 1 to the total number of parameters in A and B (There might be less if some are clamped to zero). We denote the total number of parameters of 8 by I and leave the (fixed) correspondence between the Oi and the entries of A and B unspecified. The indices are naturally partitioned into classes corresponding to the rows of the matrices. We denote by [i] the class of parameters to which Oi belongs and by O[i) the vector of all OJ S.t. j E [i]. If j E [i] then both Oi and OJ are parameters from the same row of one of the two matrices. Whenever it is clear from the context, we will use [i] to denote both a class of parameters and the row number (i.e. state) associated with the class. We now can rewrite P(x, s18) as nf=l O~'(X,S), where ni(x, s) is the number of times parameter i is used along the path s with observation sequence x. (Note that this value does not depend on the actual parameters 8.) We next compute partial derivatives ofthe likelihood and the log-likelihood using this notation. o OOi P(x, s18) oLL(XI8) OOi lInl(X,S) lIn._I(X,S) ( ) lIn,(X,S)-l lInl(X,S) u 1 ... U i-I ni x, SUi ... U 1 Here 11i(xI8) ~ Es ni(x, s)P(slx, 8) is the expected number of occurrences of the transition/output that corresponds to Oi over all paths that produce x in 8. These values are calculated in the expectation step of the Expectation-Maximization (EM) training algorithm for HMMs [7], also known as the Baum-Welch [2] or the ForwardBackward algorithm. In the next sections we use the additional following expectations, 11i(8) ~ Ex,s ni(X, s)P(x, s18) and 11[i) (8) ~ EjE[i) 11j(8). Note that the summation here is over all legal x and s of arbitrary length and 11[i) (8) is the expected number of times the state [i] was visited. 2 Entropic distance functions for HMMs Our training algorithms are based on the following framework of Kivinen and Wannuth for motivating iterative updates [6]. Assume we have already done a number of iterations and our current parameters are 8. Assume further that X is the set of observations to be processed in the current iteration. In the batch case this set never changes and in the on-line case X is typically a single observation. The new parameters 8 should stay close to 8, which incorporates all the knowledge obtained in past iterations, but it should also maximize the log-likelihood on the current date set X. Thus, instead of maximizing the loglikelihood we maximize, U(8) = 7JLL(XI8) - d(8, 8) (see [6, 5] for further motivation). Training Algorithms/or Hidden Markov Models 643 Here d measures the dis!ance between the old and new parameters and 1] > 0 is a trade-off factor. Maximizing U~B) is usually difficult since both the distance function and the loglikelihood depend on B. As in [6, 5], we approximate the log-likelihood by a first order Taylor expansion around 9 = B and add Lagrange multipliers for the constraints that the parameters of each class must sum to one: U(8) :::::: 1] (LL(XIB) + (8 - B)\7 BLL(XIB») - d(8, B) + L A[i] L OJ. (3) [i] JEri] A commonly used distance function is the relative entropy. To calculate the relative entropy between two HMMs we need to sum over all possible hidden state sequence which leads to the following definition, d (8 B) ~f ~ P( 18) 1 P(xI8) = ~ (~ P( 18») 1 Ls P(x, s19) RE, ~ x n P(xIB) ~ '7 x, s n Ls P(x, siB) However, the above divergence is very difficult to calculate and is not a convex function in B. To avoid the computational difficulties and the non-convexity of dRE we upper bound the relative entropy using the log sum inequality [3]: ~ def ~ P(x, s19) dRE (8, B) :s; dRE(B,8) = L.t P(x, s18) In P( IB) X,s x,s _ (nIl On.(x,s») _ I O. L P(x, siB) In 't (J~'(X , S) = L P(x, siB) ?= ni(x, s) In (J: x,s n,=1 , x,s ,=1 ~ O· ~ ~ O· L.t In (J~ L.t P(x, s18) ni(x, s) = L.t ni(8) In (J~ i=1 ' X,S i=I' Note that for the distance function ~E(9, 8) an HMM is viewed as a joint distribution between observation sequences and hidden state sequences. We can further simplify the bound on the relative entropy using the following lemma (proof omitted). Lemma 1 ForanyabsorbingHMM, 8, and any parameter(Jj E 8, ni(8) = (Jin[i](B). This gives the following new formula, dRE (9, 8) = L7= 1 n[j] (9) [Oi In ~ ] , which can ~ {j be rewritten as, dRE(8, B) = L[i] n[i](8) dRE(8[iJ> B[i]) = L[i] n[i](8) LjE[i] (Jj In ~ . Equation (3) is still difficult to solve since the variables n[i] (9) depend on the new set of parameters (which~ are not known). We therefore further approximate ~E(8, 8) by the distance function, dRE(9, B) = L[i] n[i](B) LjE[i] OJ In~. 3 New Parameter Updates We now would like to use the distance functions discussed in previous section in U (9). We first derive ou~ main update using this distance function. This is done by replacing d( 8,8) in U (9) with ~E (9, 8) and setting the derivatives of the resul ting U (9) w.r.t OJ to O. This gives the following set of equations (i E {I, ... , I}), LXEX ni(xIB) A Oi _ 1] IXI(Ji n[i](B) (In (Ji - 1) + A[i] - 0 , which are equivalent to 644 Y. Singer and M. K. Warmuth We now can solve for Oi and replace A[i] by a nonnalization factor which ensures that the sum of the parameters in [i] is 1: (~ 2:XEX n.(XI8)) _ OJ exp n).)(8) IXI9. Oi = (2:x nJ(XI8)) 2:jE[i] OJ exp n[)6) '~I 9J (4) The above re-estimation rule is the entropic update for HMMs. l We now derive an alternate to the updateof(4). The mixture weights n[i](8) (whichapproximate the original mixture weights n[i] (0) in ~E (0, 8) lead to a state dependent learning rate of ~ for the parameters of class [i]. If computation time is limited (see discussion n[.)(H) below) then the expectations n[i] (8) can be approximated by values that are readily available. One possible choice is to use the sample based expectations 2:jE[i]2:xEX nj(xI8)/IXI as an approximation for n[i] (8). These weights are needed for calculating the gradient and are evaluated in the expectation step of Baum-Welch. Let, n[i](xI8) ~ 2:jE[i] nj(xI8), then this approximation leads to the following distance function "'" 2:xEX n[i](xI8) d (0. 8.) = "'" 2:xEx n[j)(xI8) "'" O· In OJ (5) LIXI RE [~l> [a) LIXI LJ 0.' [i] [i] JEri] J which results in an update which we call the approximated entropic update for HMMs: ( 1) 2:xEx n.(XI8)) Oi exp ~ . (XI8) 9, DXEX n['1 Oi = ( ~ Ll ) ~. 1) DXEX nJ(Xlo) DjE[i) OJ exp 2:xEx nbl(XI8) 9J (6) given a current set of parameters 8 and a learning rate 11 we obtain a new set of parameters 8 by iteratively evaluating the right-hand-side of the entropic update or the approximated entropic update. We calculate the expectations ni(xI8) as done in the expectation step of Baum-Welch. The weights n[i](xI8) are obtained by averaging nj(xI8) for j E [i]. This lets us evaluate the right-hand-side of the approximated entropic update. The en tropic update is slightly more involved and requires an additional calculation of n[i) (8). (Recall that n[i] (8) is the expected number oftimes state [i] is visited, unconditioned on the data). To compute these expectations we need to sum over all possible sequences of state-observation pairs. Since the probability of outputting the possible symbols at a given state sum to one, calculating n[i] (8) reduces to evaluating the probability of reaching a state for each possible time and sequence length. For absorbing HMMs n[i] (8) can be approximated efficiently using dynamic programming; we compute n[i] (8) by summing the probabilities of all legal state sequences S of up to length eN (typically C = 3 proved to be sufficient to obtain very accurate approximations of n[i] (8). Therefore, the time complexity of calculating n[i] (8) depends only on the number of states, regardless of the dimension of the output vector M and the training data X. 1 A subtle improvement is possible over the update (4) by treating the transition probabilities and output probabilities differently. First the transition probabilities are updated based on (4). Then the state probabilities n[i)(O) = n[i)(A) are recomputed based on th; new parameters A. This is possible since the state probabilities depend only on the transition probabilities and not on the output probabilities. Finally the output probabilities are updated with (4) where the n[.)(O) are used in place of the n[i](8). Training Algorithms/or Hidden Markov Models 645 4 The relation to EM and convergence properties We first show that the EM algorithm for HMMs can be derived using our framework. To do so, we approximate the relative entropy by the X2 distance (see [3]), dRE(p, p) ~ dx2(p, p) ~ ~ L:i (P.;~./, and use this distance to approximate dRE(9, 8): dRE(9, 8) ~ ~2(9, 8) 1;£ 2: 1l[i) (9) dx2(9[i),8[i)) [i) ~ A ~ L:XEX 1l[i) (xI8) ~ L-n[i)(8) dx2(8[i]>8[i)) ~ LIXI [i) [i) Here dx2(9[i), 8[i)) = ~ L:j E[i) (9'~,8.)2 . By minimizing U (9) with the last version of the X2 distance function and following the same derivation steps as for the approximated entropic update we arrive at what we call the approximated X2 update for HMMs: Oi = (1 - 7J)Oi + 7J 2: 1li(xI8) /2: 1l[j)(xI8) . (7) XEX XEX Setting TJ = 1 results in the update, Oi = L:xEX 1li(xI8)/L:xEX 1l[i) (xI8), which is the maximization (re-estimation) step of the EM algorithm. Although omitted from this paper due to the lack of space, it is can be shown that for 7J E (0,1] the en tropic updates and the X2 update improve the likelihood on each iteration. Therefore, these updates belong to the family of Generalized EM (GEM) algorithms which are guaranteed to converge to a local maximum given some additional conditions [4]. Furthennore, using infinitesimal analysis and second order approximation of the likelihood function at the (local) maximum similar to [10]. it can be shown that the approximated X2 update is a contraction mapping and close to the local maximum there exists a learning rate 7J > 1 which results in a faster rate of convergence than when using TJ = 1. 5 Experiments with Artificial and Natural Data In order to test the actual convergence rate of the algorithms and to compare them to Baum-Welch we created synthetic data using HMMs. In our experiments we mainly used sparse models, that is, models with many parameters clamped to zero. Previous work (e.g., [5, 6]) might suggest that the entropic updates will perfonn better on sparse models. (Indeed, when we used dense models to generate the data, the algorithms showed almost the same perfonnance). The training algorithms, however, were started from a randomly chosen dense model. When comparing the algorithms we used the same initial model. Due to different trajectories in parameter space, each algorithm may converge to a different (local) maximum. For the clarity of presentation we show here results for cases where all updates converged to the same maximum, which often occur when the HMM generating the data is sparse and there are enough examples (typically tens of observations per non-zero parameter). We tested both the entropic updates and the X2 updates. Learning rates greater than one speed up convergence. The two entropic updates converge almost equally fast on synthetic data generated by an HMM. For natural data the entropic update converges slightly faster than the approximated version. The X2 update also benefits from learning rates larger than one. However, the x2-update need to be used carefully since it does not necessarily ensure non-negativeness of the new parameters for 7J > 1. This problems is exaggerated when the data is not generated by an HMM. We therefore used the entropic updates in our experiments with natural data. In order to have a fair comparison, we did not tune the learning rate 7J and set it to 1.5. In Figure 1 we give a comparison of the entropic update, the approximated entropic update, and Baum-Welch (left figure), using an HMM to generate the random observation sequences, where N = M = 40 but only 25% (10 parameters on the average for each transition/observation vector) of the parameters of the 646 Y. Singer and M. K. Warmuth HMM are non-zero. The perfonnance of the entropic update and the approximated entropic update are practically the same and both updates clearly outperfonn Baum-Welch. One reason the perfonnance of the two entropic updates is the same is that the observations were indeed generated by an HMM. In this case, approximating the expectations n(il (8) by the sample based expectations seems reasonable. These results suggest a valuable alternative to using Baum-Welch with a predetermined sparse, potentially biased, HMM where a large number of parameters is clamped to zero. Instead, we suggest starting with a full model and let one of the en tropic updates find the relevant parameters. This approach is demonstrated on the right part of Figure 1. In this example the data was generated by a sparse HMM with 100 states and 100 possible output symbols. Only 10% ofthe HMM's parameters were nonzero. Three log-likelihood curves are given in the figure. One is the log-likelihood achieved by Baum-Welch when only those parameters that are non-zero in the HMM generating the data are initialized to random non-zero values. The other two are the log-likelihood of the entropic update and Baum-Welch when all the parameters are initialized randomly. The curves show that the en tropic update compensates for its inferior initialization in less than 10 iterations (see horizontal line in Figure 1) and from this point on it requires only 23 more iterations to converge compared to Baum-Welch which is given prior knowledge of the non-zero parameters. In contrast, when Baum-Welch is started with a full model then its convergence is much slower than the entropic update. -0 . 4 r---"---"'--~----r--~---, 'g -0 .6 ~ .. -0,8 ~ 7 - 1 go ol 1:1 - 1. 2 ~ ~ - 1. 4 ~ -1. 6 Entrop ic Upda t e -Entr opi c Update ....... -~ EM (Daum- we l ch) · 41' " -1. 8 '-_"'--_...L..-_-'-_-'--_--'-_--' o 10 15 20 25 30 Iter a t ion" EM (Daum-welc h ) . Random I ni t. Ent r opic Update, Random I ni t . EM (Baum-we l c h). s pa r se I ni t . . .g.. -2. 4 "-_...L..-_--'--_-'-_--'--_--'-_--' 20 40 60 80 100 120 Iter a t ion " Figure 1: Comparison of the entropic updates and Baum-Welch. We next tested the updates on speech pronunciation data. In natural speech, a word might be pronounced differently by different speakers. A common practice is to construct a set of stochastic models in order to capture the variability of the possible pronunciations. alternative pronunciations of a given word. This problem was studied previously in [9] using a state merging algorithm for HMMs and in [8] using a subclass of probabilistic finite automata. The purpose of the experiments discussed here is not to compare the above algorithms to the en tropic updates but rather compare the entropic updates to Baum-Welch. Nevertheless, the resulting HMM pronunciation models are usually sparse. Typically, only two or three phonemes have a non zero output probability at a given state and the average number of states that in practice can follow a states is about 2. Therefore, the entropic updates may provide a good alternative to the algorithms presented in [8, 9]. We used the TIMIT (Texas Instruments-MIT) database as in [8, 9]. This database contains the acoustic wavefonns of continuous speech with phone labels from an alphabet of 62 phones which constitute a temporally aligned phonetic transcription to the uttered words. For the purpose of building pronunciation models, the acoustic data was ignored and we partitioned the phonetic labels according to the words that appeared in the data. The data was filtered and partitioned so that words occurring between 20 and 100 times in the dataset were used for training and evaluation according to the following partition. 75% of the occurrences of each word were used as training data for the learning algorithm and the remaining 25% were used for evaluation. We then built for each word three pronunciation models by training a fully connected HMM whose number of states was set to 1, 1.5 and 1.75 times the longest sample (denoted by N m). The models were evaluated by calculating Training Algorithmsfor Hidden Markov Models 647 the log-likelihood (averaged over 10 different random parameter initializations) of each HMM on the phonetic transcription of each word in the test set. In Table 1 we give the negative log-likelihood achieved on the test data together with the average number of iterations needed for training. Overall the differences in the log-likelihood are small which means that the results should be interpreted with some caution. Nevertheless, the entropic update obtained the highest likelihood on the test data while needing the least number of iterations. The approximated en tropic update and Baum-Welch achieve similar results on the test data but the latter requires more iterations. Checking the resulting models reveals one reason why the en tropic update achieves higher likelihood values, namely, it does a better job in setting the irrelevant parameters to zero (and it does it faster). Negative Log-Likelihood # Iterations # States 1.0Nm 1.5Nm 1.75Nm 1.0Nm 1.5Nm 1.75Nm Baum-Welch 2448 2388 2425 27.4 36.1 41.1 Approx. EU 2440 2389 2426 25.5 35.0 37.0 Entropic Update 2418 2352 2405 23.1 30.9 32.6 Table 1: Comparison of the entropic updates and Baum-Welch on speech pronunciation data. 6 Conclusions and future research In this paper we have showed how the framework of Kivinen and Warmuth [6] can be used to derive parameter updates algorithms for HMMs. We view an HMM as a joint distribution between the observation sequences and hidden state sequences and use a bound on relative entropy as a distance between the new and old parameter settings. If we approximate of the relative entropy by the X2 distance, replace the exact state expectations by a sample based approximation, and fix the learning rate to one then the framework yields an alternative derivation of the EM algorithm for HMMs. Since the EM update uses sample based estimates of the state expectations it is hard to use it in an on-line setting. In contrast, the on-line versions of our updates can be easily derived using only one observation sequence at a time. Also, there are alternative gradient descent based methods for estimating the parameters of HMMs. Such methods usually employ an exponential parameterization (such as soft-max) of the parameters (see [1 D. For the case of learning one set of mixture coefficients an exponential parameterization led to an algorithm with a slower convergence rate compared to algorithms derived using entropic distances [5]. However, it is not clear whether this is still the case for HMMs. Our future goals is to perform a comparative study of the different updates with emphasis on the on-line versions. Acknowledgments We thank Anders Krogh for showing us the simple derivative calculations used in this paper and thank Fernando Pereira and Yasubumi Sakakibara for interesting discussions. References [1] P. Baldi and Y. Chauvin. Smooth on-line learning algorithms for Hidden Markov Models. Neural Computation. 6(2), 1994. [2] L.E. Baum and T. Petrie. Statistical inference for probabilistic functions of finite state markov chains. Annals of Mathematic a I Statisitics, 37, 1966. [3] T. Cover and J. Thomas. Elements of Information Theory. Wiley, 1991. [4] A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum-likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, B39: 1-38,1977. [5] D. P. Helmbold, R. E. Schapire, Y. Singer, and M. K. Warmuth. A comparison of new and old algorithms for a mixture estimation problem. In Proceedingsofthe Eighth Annual Workshop on Computational Learning Theory, pages 69-78,1995. [6] J. Kivinen and M. K. Warmuth. Exponentiated gradient versus gradient descent for linear predictors. Informationa and Computation, 1997. To appear. [7] LR Rabiner and B. H. Juang. An introduction to hidden markov models. IEEE ASSP Magazine, 3(1 ):4-16, 1986. [8] D. Ron, Y. Singer, and N. Tishby. On the learnability and usage of acyclic probabilistic finite automata. In Proc. of the Eighth Annual Workshop on Computational Learning Theory, 1995. [9] A. Stolcke and S. Omohundro. Hidden Markov model induction by Bayesian model merging. In Advances in Neural Information Processing Systems, volume 5. Morgan Kaufmann, 1993. [10] L. Xu and M.I. Jordan. On convergence properties of the EM algorithm for Gaussian mixtures. Neurol Computation, 8:129-151 , 1996.
1996
75
1,275
A Constructive Learning Algorithm for Discriminant Tangent Models Diego Sona Alessandro Sperduti Antonina Starita Dipartimento di Informatica, U niversita di Pisa Corso Italia, 40, 56125 Pisa, Italy email: {sona.perso.starita}di.unipi.it Abstract To reduce the computational complexity of classification systems using tangent distance, Hastie et al. (HSS) developed an algorithm to devise rich models for representing large subsets of the data which computes automatically the "best" associated tangent subspace. Schwenk & Milgram proposed a discriminant modular classification system (Diabolo) based on several autoassociative multilayer perceptrons which use tangent distance as error reconstruction measure. We propose a gradient based constructive learning algorithm for building a tangent subspace model with discriminant capabilities which combines several of the the advantages of both HSS and Diabolo: devised tangent models hold discriminant capabilities, space requirements are improved with respect to HSS since our algorithm is discriminant and thus it needs fewer prototype models, dimension of the tangent subspace is determined automatically by the constructive algorithm, and our algorithm is able to learn new transformations. 1 Introduction Tangent distance is a well known technique used for transformation invariant pattern recognition. State-of-the-art accuracy can be achieved on an isolated handwritten character task using tangent distance as the classification metric within a nearest neighbor algorithm [SCD93]. However, this approach has a quite high computational complexity, owing to the inefficient search and large number of Euclidean and tangent distances that need to be calculated. Different researchers have shown how such time complexity can be reduced [Sim94, SS95] at the cost of increased space complexity. A Constructive Learning Algorithm for Discriminant Tangent Models 787 A different approach to the problem was used by Hastie et al. [HSS95) and Schwenk & Milgram [SM95b, SM95a). Both of them used learning algorithms for reducing the classification time and space requirements, while trying to preserve the same accuracy. Hastie et al. [HSS95) developed rich models for representing large subsets of the prototypes. These models are learned from a training set through a Singular Value Decomposition based algorithm which minimizes the average 2-sided tangent distance from a subset of the training images. A nice feature of this algorithm is that it computes automatically the "best" tangent subspace associated with the prototypes. Schwenk & Milgram [SM95b) proposed a modular classification system (Diabolo) based on several autoassociative multilayer perceptrons which use tangent distance as the error reconstruction measure. This original model was then improved by adding discriminant capabilities to the system [SM95a). Comparing Hastie et al. algorithm (HSS) versus the discriminant version of Diabolo, we observe that: Diabolo seems to require less memory than HSS, however, learning is faster in HSS; Diabolo is discriminant while HSS is not; the number of hidden units to be used in Diabolo's autoassociators must be decided heuristically through a trial and error procedure, while the dimension of the tangent subspaces in HSS can be controlled more easily; Diabolo uses predefined transformations, while HSS is able to learn new transformations (like style transformations). In this paper, we introduce the tangent distance neuron (TO-neuron), which implements the I-sided version of the tangent distance, and we devise a gradient based constructive learning algorithm for building a tangent subspace model with discriminant capabilities. In this way, we are able to combine the advantages of both HSS and Diabolo: the model holds discriminant capabilities, learning is just a bit slower than HSS, space requirements are improved with respect to HSS since the TO-neuron is discriminant and thus it needs fewer prototype models, the dimension of the tangent subspace is determined automatically by the constructive algorithm, and TO-neuron is able to learn new transformations. 2 Tangent Distance In several pattern recognition problems Euclidean distance fails to give a satisfactory solution since it is unable to account for invariant transformations of the patterns. Simard et al. [SCD93) suggested dealing with this problem by generating a parameterized 7-dimensional manifold for each image, where each parameter accounts for one such invariance. The underlying idea consists in approximating the considered transformations locally through a linear model. For the sake of exposition, consider rotation. Given a digitalized image Xi of a pattern i, the rotation operation can be approximated by Xi(O) = Xi + Tx,O, where 0 is the rotation angle, and T x, is the tangent vector to the rotation curve generated by the rotation operator for Xi. The tangent vector T x, can easily be computed by finite difference. Now, instead of measuring the distance between two images as D( Xi, X j) = IIX i-X j II for any norm 11·11. Simard et al. proposed using the tangent distance DT(Xi,Xj ) = min9.,9, IIXi(Oi) - Xj(Oj)lI. If k types of transformations are considered, there will be k different tangent vectors per pattern. If II . II is the Euclidean norm, computing the tangent distance is a simple least-squares problem. A solution for this probleml can be found in Simard et al. [SCD93], where the authors used DT to drive a I-NN classification rule. 1 A special case of tangent distance, i.e., the one sided tangent distance D~-·;ded(x" X J ) = mine; IIX,((J,) - Xjll, can be computed more efficiently [SS95]. 788 D. Sona, A. Sperduti and A. Starita Figure 1: Geometric interpretation of equation 1. Note that net = (D~-8ided ):l. Unfortunately, 1-NN is expensive. To reduce the complexity ofthe above approach, Hastie et al. [HSS95] proposed an algorithm for the generation of rich models representing large subsets of patterns. This algorithm computes for each class a prototype (the centroid), and an associated subspace (described by the tangent vectors), such that the total tangent distance of the centroid with respect to the prototypes in the training set is minimised. Note that the associated subspace is not predefined as in the case of standard tangent distance, but is computed on the basis of the training set. 3 Tangent Distance Neuron In this section we define the Tangent ~istance neuron (TO-neuron), which is the computational model studied in this paper. A TO-neuron is characterized by a set of n + 1 vectors, of the same dimension as the input vectors (in our case, images). One of these vectors, W is used as reference vector (centroid), while the remaining vectors, Ti (i = 1, ••• , n), are used as tangent vectors. Moreover, the set of tangent vectors constitutes an ortho-normal basis. Given an input vector I the input net of the TO-neuron is computed as the square of the I-sided tangent distance between I and the tangent model {W, T 1 , ••• , Tn} (see Figure 1) n n where we have used the fact that the tangent vectors constitute an ortho-normal basis. For the sake of notation, d denotes the difference between the input pattern and the centroid, and the projection of d over the i-th tangent vector is denoted by "'fi. Note that, by definition, net is non-negative. The output 0 of the TO-neuron is then computed by transforming the net through a nonlinear monotone function f. In our experiments, we have used the following function 1 0= f(o.,net) = ---1 + 0. net (2) where 0. controls the steepness of the function. Note that 0 is positive since net is always positive and within the range (0, 1]. A Constructive Learning Algorithm for Discriminant Tangent Models 789 4 Learning The TD-neuron can be trained to discriminate between patterns belonging to two different classes through a gradient descent technique. Thus, given a training set {(I"t,), ... ,(IN,tN)}, where ti E {O,1} is the i-th desired output, and N is the total number of patterns in the training set, we can define the error function as N E = = 2:)tk Okr~ 2 k=l (3) where Ok is the output of the TD-neuron for the k-th input pattern. Using equations (1-2), it is trivial to compute the changes for the tangent vectors, the centroid and 0.: (4) (5) ..do. :=: -'1101 (6E) = - t netk 'I101(tk - Ok) o~ 60. k=l (6) where '11 and '1101 are learning parameters. The learning algorithm initializes the centroid W to the average of the patterns with target 1, i.e., W = /., Lf~, Ik , where N, is the number of patterns with target equal to 1, and the tangent vectors to random vectors with small modulus. Then 0., the centroid Wand the tangent vectors Ti are changed according to equations (4-6). Moreover, since the tangent vectors must constitute an ortho-normal basis, after each epoch of training the vectors Ti are ortho-normalized. 5 The Constructive Algorithm Before training the TD-neuron using equations (4-6), we have to set the tangent subspace dimension. The same problem is present in HSS and Diabolo (i.e., number of hidden units). To solve this problem we have developed a constructive algorithm which adds tangent vectors one by one according to the computational needs. The key idea is based on the observation that a typical run of the learning algorithm described in Section 4 leads to the sequential convergence of the vectors according to their relative importance. This means that the tangent vectors all remain random vectors while the centroid converges first. Then one of the tangent vectors converges to the most relevant transformation (while the remaining tangent vectors are still immature), and so on till all the tangent vectors converge, one by one, to less and less relevant transformations . This behavior suggests starting the training using only the centroid (i.e., without tangent vectors) and allow it to converge. Then, as in other constructive algorithms, the centroid is frozen and one random tangent vector T 1 is added. Learning is resumed till changes in T 1 become irrelevant. During learning, however, T, is normalized after each epoch. At convergence, T 1 is frozen, a new random tangent vector T 2 is added, and learning is resumed. New tangent vectors are iteratively added till changes in the classification accuracy becomes irrelevant. 790 D. Sona, A. Sperduti and A. Starita HSS TD-neuron # Tang. % Cor % Err % Cor % Rej % Err 0 73.78 7.24 18.98 1 78. 74 21.26 72.06 10 .• 8 17.46 2 79.10 20.90 77.99 8 .05 13.96 3 79 .94 20.06 81.14 7.17 11.69 4 81.47 18.53 82.68 6 .8. 10.48 5 76.87 23.13 84.25 5.63 10.12 6 71 .29 28.71 85.21 5 .14 9 .65 7 86.16 4.76 9 .08 8 86.37 4.89 8 .74 Table 1; The results obtained by the HSS algorithm and the TO-neuron. 6 Results We have tested our constructive algorithm versus the HSS algorithm (which uses the 2-sided tangent distance) on 10587 binary digits from the NIST-3 dataset. The binary I28xI28 digits were transformed into a 64-grey level I6xI6 format by a simple local counting procedure. No other pre-processing transformation was performed. The training set consisted of 3000 randomly chosen digits, while the remaining digits where used in the test set. A single tangent model for each class of digit was computed using both algorithms. The classification of the test digits was performed using the label of the closest model for HSS and the output of the TO-neurons for our system. The TO-neurons used a rejection criterion with parameters adapted during training. In Table 1 we have reported the performances on the test set of both HSS and our system. Oifferent numbers of tangent vectors were tested for both of them. From the results it is clear that the models generated by HSS reach a peak in performance with 4 tangent vectors and then a sharp degradation of the generalization is observed by adding more tangent vectors. On the contrary, the TO-neurons are able to steadly increase the performance with an increasing number of tangent vectors. The improvement in the performance, however, seems to saturate when using many tangent vectors. Table 2 presents the confusion matrix obtained by the TO-neurons with 8 tangent vectors. For comparison, we display some of the tangent models computed by HSS and by our algorithm in Figure 2. Note how tangent models developed by the HSS algorithm tend to be more blurred than the ones developed by our algorithm. This is due to the lake of discriminant capabilities by the HSS algorithm and it is the main cause of the degradation in performance observed when using more than 4 tangent vectors. It must be pointed out that, for a fixed number of tangent vectors, the HSS algorithm is faster than ours, because it needs only a fraction of the training examples (only one class). However, our algorithm is remarkably more efficient when a family of tangent models with an increasing number of tangent vectors must be generated2 • Moreover, since a TO-neuron uses the one sided tangent distance, it is faster in computing the output. 7 Conclusion We introduced the tangent distance neuron (TO-neuron), which implements the I-sided version of the tangent distance and gave a constructive learning algorithm for building a tangent subspace with discriminant capabilities. As stated in the in:>The tangent model computed by HSS depends on the number of ta.ngent vectors. A Constructive Learning Algorithm/or Discriminant Tangent Models 791 Figure 2: The tangent models obtained for digits '1' and '3' by the HSS algorithm (row 1 and 3, respectively) and our TD-neuron (row 2 and 4, respectively). The centroids are shown in the first column. troduction , there are many advantages of using the proposed computational model versus other techniques like HSS and Diabolo. Specifically, we believe that the proposed approach is particularly useful in those applications where it is very important to have a classification system which is both discriminant and semantically transparent, in the sense that it is very easy to understand how it works. One among these applications is the classification of ancient book scripts. In fact, the description, the comparison, and the classification of forms are the main tasks of paleographers. Until now, however, these tasks have been generally performed without the aid of a universally accepted and quantitatively based method or technique. Consequently, very often it is impossible to reach a definitive date attribution of a document to within 50 years. In this field, it is very important to have a system which is both discriminant and explanatory, so that paleographers can learn from it which are the relevant features of the script of a given epoch. These requirements rule out systems like Diabolo, which is not easily interpretable, and also tangent models developed by HSS, which are not discriminant. In Figure 3 we have reported some preliminary results we obtained within this field. Perhaps most importantly, our work suggests a number of research avenues. We used just a single TD-neuron; presumably having several neurons arranged as an adaptive pre-processing layer within a standard feed-forward neural network can yield a remarkable increase in the transformation invariant features of the network. 00 08 OQ ReJ II % Cor % Rej % Err Co 661 { 2 5 27 8 2 1 9 0 {2 86.86 5 .52 1.62 C t 4 M2 0 1 1 1 1 11 8 0 9 95.90 1.03 3.08 C~ { 1 650 2 13 2 10 6 9 0 69 M.86 9.01 6.14 C.,. 1 3 22 656 0 26 1 11 18 { 28 85.19 3.6{ 11.17 C 0 0 2 0 633 3 4 5 1 48 32 86.2{ {.36 9 .{0 . Co. 1 1 3 39 2 535 7 1 7 3 H 83.20 6.M 9.95 Cil 0 4 1 2 6 11 680 0 4 0 33 91. 77 4.45 3.18 C,7 1 1 3 0 4 3 0 127 12 24 27 90.65 3 .37 5 .99 CR 1 4 1 18 14 12 0 7 607 11 62 81.70 8 .34 9.96 C',q 0 0 0 12 43 1 0 70 36 562 25 15.03 3.34 21.63 Total Correct: 86.37% Re.Jected 4 .89% Errors 8 .74% Table 2: The confusion matrix for the TD-neurons with 8 tangent vectors.
1996
76
1,276
Effective Training of a Neural Network Character Classifier for Word Recognition Larry Yaeger Apple Computer 5540 Bittersweet Rd. Morgantown, IN 46160 larryy@apple.com Richard Lyon Apple Computer 1 Infinite Loop, MS301-3M Cupertino, CA 95014 lyon@apple.com Abstract Brandyn Webb The Future 4578 Fieldgate Rd. Oceanside, CA 92056 brandyn@brainstorm.com We have combined an artificial neural network (ANN) character classifier with context-driven search over character segmentation, word segmentation, and word recognition hypotheses to provide robust recognition of hand-printed English text in new models of Apple Computer's Newton MessagePad. We present some innovations in the training and use of ANNs al; character classifiers for word recognition, including normalized output error, frequency balancing, error emphasis, negative training, and stroke warping. A recurring theme of reducing a priori biases emerges and is discussed. 1 INTRODUCTION We have been conducting research on bottom-up classification techniques ba<;ed on trainable artificial neural networks (ANNs), in combination with comprehensive but weakly-applied language models. To focus our work on a subproblem that is tractable enough to le.:'ld to usable products in a reasonable time, we have restricted the domain to hand-printing, so that strokes are clearly delineated by pen lifts. In the process of optimizing overall performance of the recognizer, we have discovered some useful techniques for architecting and training ANNs that must participate in a larger recognition process. Some of these techniques-especially the normalization of output error, frequency balanCing, and error emphal;is-suggest a common theme of significant value derived by reducing the effect of a priori biases in training data to better represent low frequency, low probability smnples, including second and third choice probabilities. There is mnple prior work in combining low-level classifiers with various search strategies to provide integrated segmentation and recognition for writing (Tappert et al 1990) and speech (Renals et aI1992). And there is a rich background in the use of ANNs a-; classifiers, including their use as a low-level, character classifier in a higher-level word recognition system (Bengio et aI1995). But many questions remain regarding optimal strategies for deploying and combining these methods to achieve acceptable (to a real user) levels of performance. In this paper, we survey some of our experiences in exploring refinements and improvements to these techniques. 2 SYSTEM OVERVIEW Our recognition system, the Apple-Newton Print Recognizer (ANPR), consists of three conceptual stages: Tentative Segmentation, Classification, and Context-Driven Search. The primary data upon which we operate are simple sequences of (x,y) coordinate pairs, 808 L. Yaeger, R. Lyon and B. Webb plus pen-up/down infonnation, thus defining stroke primitives. The Segmentation stage decides which strokes will be combined to produce segments-the tentative groupings of strokes that will be treated as possible characters-and produces a sequence of these segments together with legal transitions between them. This process builds an implicit graph which is then scored in the Classification stage and examined for a maximum likelihood interpretation in the Search stage. (x,y) Points & Pen-Lifl'i Neural Network ~------~~ ~------~~ Classifier Character Segmentation Hypotheses Character Class Hypotheses Words Figure 1: A Simplified Block Diagram of Our Hand-Print Recognizer. 3 TRAINING THE NEURAL NETWORK CLASSIFIER Except for an integrated multiple-representations architecture (Yaeger et a11996) and the training specitics detailed here, a fairly standard multi-layer perceptron trained with BP provides the ANN character cla<;sitler at the heart of ANPR. Training an ANN character cla<;sifier for use in a word recognition system, however, has different constraints than would training such a system for stand-alone character recognition. All of the techniques below, except for the annealing schedule, at least modestly reduce individual character recognition accuracy, yet dramatically increase word recognition accuracy. A large body of prior work exist<; to indicate the general applicability of ANN technology as classifiers providing good estimates of a posteriori probabilities of each class given the input (Gish 1990, Richard and Lippmann 1991, Renals and Morgan 1992, Lippmann 1994, Morgan and Bourlard 1995, and others cited herein). 3.1 NORMALIZING OUTPUT ERROR Despite their ability to provide good first choice a posteriori probabilities, we have found that ANN cla<;sifiers do a poor job of representing second and third choice probabilities when trained in the classic way-minimizing mean squared error for target vectors that are all O's, except for a single 1 corresponding to the target class. This result<; in erratic word recognition failures as the net fails to accurately represent the legitimate ambiguity between characters. We speculated that reducing the "pressure towards 0" relative to the "pressure towards 1" as seen at the output unit", and thus reducing the large bias towards o in target vectors, might pennit the net to better model these inherent ambiguities. We implemented a technique for "nonnalizing output error" (NormOutErr) by reducing the BP error for non-target classes relative to the target class by a factor that nonnalizes the total non-target error seen at a given output unit relative to the total target error seen at that unit. Assuming a training set with equal representation of classes, this nonnalization should then be based on the number of non-target versus target classes in a typical training vector, or, simply, the number of output units (minus one). Hence for non-target output units, we scale the error at each unit by a constant: e'=Ae where eis the emlr at an output unit, and A is detlned to be: A = 1/[d(NnuIJlUrs -1)] where N"uIPuIs is the number of output units, and d is a user-adjusted tuning parameter, typically ranging from 0.1 to 0.2. Error at the target output unit is unchanged. Overall, this raises the activation values at the output units, due to the reduced pressure towards zero, particularly for low-probability samples. Thus the learning algorithm no longer Effective Training of a NN Character Classifier for Word Recognition 809 converges to a minimum mean-squared error (MMSE) estimate of P(classlinput), but to an MMSE estimate of a nonlinear function !(P(classlinput),A) depending on the factor A by which we reduced the error pressure toward zero. Using a simple version of the technique of Bourlard and Wellekens (1990), we worked out what that resulting nonlinear function is. The net will attempt to converge to minimize the modified quadratic error function (i;2) = p(l- y)2 + A(l- p)y2 by setting its output y for a particular class to y= p/(A-Ap+ p) where p = P(classlinput), and A is as defined above. The inverse function is p= yA/(yA+1- y) We verified the fit of this function by looking at histograms of character-level empirical percentage-correct versus y, a'\ in Figure 2. 1.-~~r-~~--~~--~-'--~~ 0.9 0.8 Pceel) 0.7 0.6 0.5 net output y 0.4 0.3 0.2 0.1 0 0 Figure 2: Empiric.:'ll p vs. y Histogram for a Net Trained with A=D.ll (d=.l), with the Corresponding Theoretical Curve. Note that the lower-probability samples have their output activations raised significantly, relative to the 45° line that A = 1 yields. The primary benefit derived from this technique is that the net does a much better job of representing second and third choice probabilities, and low probabilities in general. Despite a small drop in top choice character accuracy when using NormOutErr, we obtain a very significant increase in word accuracy by this technique. Figure 3 shows an exaggerated example of this effect, for an atypically large value of d (0.8), which overly penalizes character accuracy; however, the 30% decrea'\e in word error rate is normal for this technique. (Note: These data are from a multi-year-old experiment, and are not necessarily representative of current levels of performance on any absolute scale.) % E 40 30 r r 20 o 10 r o NonnOutErr = • 0.0 III 0.8 Character En-or Word Error Figure 3: Character and Word Error Rates for Two Different Values of NormOutErr (d). A Value 01'0.0 Disables NormOutErr, Yielding Normal BP. The Unusually High Value of 0.8 (A=O.013) Produces Nearly Equal Pressures Towards 0 and 1. 810 L. Yaeger, R. Lyon and B. Webb 3.2 FREQUENCY BALANCING Training data from natural English words and phrases exhibit very non-uniform priors for the various character classes, and ANNs readily model these priors. However, as with NonnOutErr, we tind that reducing the effect of these priors on the net, in a controlled way, and thus forcing the net to allocate more of its resources to low-frequency, lowprobability classes is of significant benefit to the overall word recognition process. To this end, we explicitly (partially) balance the frequencies of the classes during training. We do this by probabilistically skipping and repeating patterns, ba<;ed on a precomputed repetition factor. Each presentation of a repeated pattern is "warped" uniquely, as discussed later. To compute the repetition factor for a class i, we tirst compute a normalized frequency of that cla'ls: F; = S; IS where S; is the number of s~unples in class i, and S is the average number of srunples over all cla<;ses, computed in the obvious way: _ 1 c S =(- LS;) c ;=1 with C being the number of classes. Our repetition factor is then defined to be: R; = (a/F;t with a and b being user controls over the ~ount of skipping vs. repeating and the degree of prior normalization, respectively. Typical values of a range from 0.2 to 0.8, while b ranges from 0.5 to 0.9. The factor a < 1 lets us do more skipping than repeating; e.g. for a = 0.5, cla<;ses with relative frequency equal to half the average will neither skip nor repeat; more frequent classes will skip, and less frequent classes will repeat. A value of 0.0 for b would do nothing, giving R; = 1.0 for all classes, while a value of 1.0 would provide "full" normalization. A value of b somewhat less than one seems to be the best choice, letting the net keep some bia<; in favor of cla<;ses with higher prior probabilities. This explicit prior-bias reduction is related to Lippmann's (1994) and Morgan and Bourlard's (1995) recommended method for converting from the net's estimate of posterior probability, p(classlinput), to the value needed in an HMM or Viterbi search, p(inputlclass), which is to divide by p(class) priors. Besides eliminating potentially noisy estimates of low probability cla<;se.<; and a possible need for renormalization, our approach forces the net to actually 1e:'lfI1 a better model of these lower frequency classes. 3.3 ERROR EMPHASIS While frequency balancing corrects for under-represented classes, it cannot account for under-represented writing styles. We utilize a conceptually related probabilistic skipping of patterns, but this time for just those patterns that the net correctly classifies in it.<; forward/recognition pa<;s, as a form of "error empha<;is", to address this problem. We define a correct-train probability (0.1 to 1.0) that is used a<; a bia'led coin to determine whether a particular pattern, having been correctly classified, will also be used for the backward/training pass or not. This only applies to correctly segmented, or "positive" patterns, and miscla<;sified patterns are never skipped. Especially during early stages of training, we set this parruneter fairly low (around 0.1), thus concentrating most of the training time and the net's learning capability on patterns that are more difticult to correctly classify. This is the only way we were able to get the net to learn to correctly classify unusual character variants, such a') a 3-stroke "5" as written by only one training writer. Effective Training of a NN Character Classifier for Word Recognition 811 Variants of this scheme are possible in which mLliclalisified patterns would be repeated, or different learning rates would apply to correctly and incorrectly classified patterns. It is also related to techniques that use a training subset, from which ealiily-classified patterns are replaced by randomly selected patterns from the full training set (Guyon et aI1992). 3.4 NEGATIVE TRAINING Our recognizer's tentative segmentation stage necessarily produces a large number of invalid segments, due to inherent ambiguities in the character segmentation process. During recognition, the network must clalisify these invalid segment<; just ali it would any valid segment, with no knowledge of which are valid or invalid. A significant increase in word-level recognition accuracy wali obtained by performing negative training with these invalid segments. This consists of presenting invalid segments to the net during training, with all-zero target vectors. We retain control over the degree of negative training in two ways. First is a negative-training Jactor (0.2 to 0.5) that modulates the learning rate (equivalently by modulating the error at the output layer) for these negative patterns. This reduces the impact of negative training on positive training, and modulates the impact on characters that specifically look like element Ii of multi-stroke characters (e.g., I, 1, I, 0, 0, 0). Secondly, we control a negative-training probability (0.05 to 0.3), which determines the probability that a particular negative sample will actually be trained on (for a given presentation). This both reduces the overall impact of negative training, and significcUltly reduces training time, since invalid segment<; are more numerous than valid segments. As with NormOutErr, this modification hurtli character-level accuracy a little bit, but helps word-level accuracy a lot. 3.5 STROKE WARPING During training (but not during recognition), we produce random variations in stroke data, consisting of small changes in skew, rotation, and x and y linear and quadratic scalings. This produces alternate character forms that are consistent with stylistic variations within and between writers, and induces an explicit alipect ratio and rotation in variance within the framework of standard back-propagation. The amounts of each distortion to apply were chosen through cross-validation experimentli, as just the amount needed to yield optimum generalization. We also examined a number of such samples by eye to verify that they represent a natural range of variation. A small set of such variations is shown in Figure 4. Figure 4: A Few Random Stroke Warpings of the Same Original "m" Data. Our stroke warping scheme is somewhat related to the ideas of Tangent Dist and Tangent Prop (Simard et (II 1992, 1993), in terms of the use of predetermined families of transformations, but we believe it is much easier to implement. It is also somewhat distinct in applying transformations on the original coordinate data, as opposed to using distortions of images. The voice transformation scheme of Chang and Lippmann (1995) is also related, but they use a static replication of the training set through a small number of tr<Ulsformations, rather than dymunic random transformations of infinite variety. 3.6 ANNEALING & SCHEDULING Many discussions of back-propagation seem to assume the use of a single fixed learning rate. We view the stochastic back-propagation process ali more of a simulated annealing, with a learning rate starting very high and decreasing only slowly to a very low value. We typically start with a rate near l.0 and reduce the rate by a decay Jactor of 0.9 until it gets down to about 0.001. The rate decay factor is applied following any epoch for which the total squared error increased on the training set. Repeated tests indicate that this 812 L. Yaeger, R. Lyon and B. Webb approach yields better results than low (or even moderate) initial learning rates, which we speculate to be related to a better ability to escape local minima. In addition, we find that we obtain best overall results when we also allow some of our many training parameters to change over the course of a training run. In particular, the correct train probability needs to start out very low to give the net a chance to learn unusual character styles, but it should finish up at 1.0 in order to not introduce a general posterior probability bias in favor of classes with lots of ambiguous examples. We typic.:'llly train a net in t{)Ur "phases" according to parameters such a\; in Figure 5. Phase Epochs 1 25 2 25 3 50 4 30 Learning Rate 1.0 - 0.5 0.5 - 0.1 0.1 - 0.01 0.01 - 0.001 Correct Negative Train Train Prob Prob 0.1 0.05 0.25 0.1 0.5 0.18 1.0 0.3 Figure 5: A Typical Multi-Pha-;e Schedule of Learning Rates and Other Parameters for Training a Character-Classifier Net. 4 DISCUSSION AND FUTURE DIRECTIONS The normalization of output error, frequency balancing, and error emphasis networktraining methods discussed previously share a unifying theme: Reducing the effect of a priori biases in the training data on network learning significantly improves the network's performance in an integrated recognition system, despite a modest reduction in the network's accuracy for individual characters. Normalization of output error prevents over-represented non-target classes from bia\;ing the net against under-represented target cla\;ses. Frequency balancing prevent\; over-represented target classes from biasing the net against under-represented target classes. And error emphasis prevent\; overrepresented writing styles from bia-;ing the net against under-represented writing styles. One could even argue that negative training eliminates an absolute bias towards properly segmented characters, and that stroke warping reduces the bias towards those writing styles found in the training data, although these techniques provide wholly new information to the system as well. Though we've offered arguments for why each of these techniques, individually, helps the overall recognition process, it is unclear why prior-bias reduction, in general, should be so consistently valuable. The general effect may be related to the technique of dividing out priors, as is sometimes done to convert from p(clllsslinput) to p(inputlclass). But we also believe that forcing the net, during learning, to allocate resources to represent less frequent sample types may be directly beneficial. In any event, it is clear that paying attention to such biases and taking steps to modulate them is a vital component of effective training of a neural network serving as a classifier in a maximum likelihood recognition system. We speculate that a method of modulating the leaming rate at each output unit-based on a measure of its accuracy relative to the other output units-may be possible, and that such a method might yield the combined benefits of several of these techniques, with fewer user-controllable parameters. Effective Training of a NN Character Classifier for Word Recognition 813 Acknowledgement .. This work was done in collaboration with Bill Stafford, Apple Computer, and Les Vogel, Angel Island Technologies. We are also indebted to our many colleagues in the connectionist community and at Apple Computer. Some techniques in this paper have pending U.S. and foreign patent applications. References Y. Bengio, Y. LeCun, C. Nohl, and C. Burges, "LeRec: A NNIHMM Hybrid for On-Line Handwriting Recognition," Neural Computation, Vol. 7, pp. 1289-1303, 1995. H. Bourlard and C. .T. Wellekens, "Links hetween Markov Models and Multilayer Perceptrons," IEEE Trans. PAMI, Vol. 12, pp. 1167-1178, 1990. E. I. Chang and R. P. Lippmann, "Using Voice Transformations to Create Additional Training Talkers for Word Spotting," in Advances in Neural Information Processing Systems 7, Tesauro et al. (eds.), pp. 875-882, MIT Press, 1995. H. Gish, "A Prohahilistic Appmach to Understanding and Training of Neural Network Classifiers," Proc. IEEE Inri. Conf on Acoustic.\·, Speech, and Signal Processing (Albuquerque, NM), pp. 13611364,1990. I. Guyon, D. Henderson, P. Alhrecht, Y. LeCun, and P. Denker, "Writer independent and writer adaptive neural network for on-line character recognition," in From pixels to features III, S. Impedovo (ed.), pp. 493-506, Elsevier, Amsterdam, 1992. R. A. .Tacohs, M. I. .Tordan, S . .T. NOWlan, and G. E. Hinton, "Adaptive Mixtures of Local Experts," Neural CompUTaTion, Vol. 3, pp. 79-87, 1991. R. P. Lippmann, "Neural Networks, Bayesian a posteriori Probahilities, and Pattern Classification," pp. 83-104 in: From STaTisTics to Neural Networks-Theory and Pattern Recognition Applications, V. Cherkas sky, .T. H. Friedman, and H. Wechsler (eds.), Springer-Verlag, Berlin, 1994. N. Morgan and H. Bourlard, "Continuous Speech Recognition-An introduction to the hyhrid HMM/connectionist approach," IEEE Signal Processing Mag., Vol. 13, no. 3, pp. 24--42, May 1995. S. Renals and N. Morgan, "Connectionist Prohahility Estimation in HMM Speech Recognition," TR-92-081, International Computer Science Institute, 1992. S. Renals and N. Morgan, M. Cohen, and H. Franco "Connectionist Prohahility Estimation in the Decipher Speech Recognition System," Proc. IEEE Inti. Conf on Acoustics, Speech, and Signal Processing (San Francisco), pp. 1-601-1-604, 1992. M. D. Richard and R. P. Lippmann, "Neural Network Classifiers Estimate Bayesian a Posteriori Prohahilities," Neural ComputaTion, Vol. 3, pp. 461-483, 1991. P. Simard, B. Victorri, Y. LeCun and .T. Denker, "Tangent Prop-A Formalism for Specifying Selected Invariances in an Adaptive Network," in Advances in Neural Information Processing SYSTems 4, Moody et al. (eds.), pp. 895-903, Morgan Kaufmann, 1992. P. Simard, Y. LeCun and .T. Denker, "Efficient Pattern Recognition Using a New Transformation Distance," in Advances in Neural Information Processing Systems 5, Hanson et al. (eds.), pp. 5058, Morgan Kaufmann, 1993. C . C. Tappert, C. Y. Suen, and T. Wakahara, "The State of the Art in On-Line Handwriting Recognition," IEEE Trans. PAM/, Vol. 12, pp. 787-808, 1990. L. Yaeger, B. Wehh, and R. Lyon, "Comhining Neural Networks and Context-Driven Search for On-Line, Printed Handwriting Recognition", (unpuhlished). PART VII VISUAL PROCESSING
1996
77
1,277
Second-order Learning Algorithm with Squared Penalty Term Kazumi Saito Ryohei Nakano NTT Communication Science Laboratories 2 Hikaridai, Seika-cho, Soraku-gun, Kyoto 619-02 Japan {saito,nakano }@cslab.kecl.ntt.jp Abstract This paper compares three penalty terms with respect to the efficiency of supervised learning, by using first- and second-order learning algorithms. Our experiments showed that for a reasonably adequate penalty factor, the combination of the squared penalty term and the second-order learning algorithm drastically improves the convergence performance more than 20 times over the other combinations, at the same time bringing about a better generalization performance. 1 INTRODUCTION It has been found empirically that adding some penalty term to an objective function in the learning of neural networks can lead to significant improvements in network generalization. Such terms have been proposed on the basis of several viewpoints such as weight-decay (Hinton, 1987), regularization (Poggio & Girosi, 1990), function-smoothing (Bishop, 1995), weight-pruning (Hanson & Pratt, 1989; Ishikawa, 1990), and Bayesian priors (MacKay, 1992; Williams, 1995). Some are calculated by using simple arithmetic operations, while others utilize higher-order derivatives. The most important evaluation criterion for these terms is how the generalization performance improves, but the learning efficiency is also an important criterion in large-scale practical problems; i.e., computationally demanding terms are hardly applicable to such problems. Here, it is naturally conceivable that the effects of penalty terms depend on learning algorithms; thus, we need comparative evaluations. This paper evaluates the efficiency of first- and second-order learning algorithms 628 K. Saito and R. Nakano with three penalty terms. Section 2 explains the framework of the present learning and shows a second-order algorithm with the penalty terms. Section 3 shows experimental results for a regression problem, a graphical evaluation, and a penalty factor determination llsing cross-validation. 2 LEARNING WITH PENALTY TERM 2.1 Framework Let {(Xl, Y1),"', (xm, Ym)} be a set of examples, where Xt denotes an n-dimensional input vector and Yt a target value corresponding to Xt. In a three-layer neural network, let h be the number of hidden units, Wj (j = 1,"', h) be the weight vector between all the input units and the hidden unit j, and Wo = (WOO,"" WOh)T be the weight vector between all the hidden units and the output unit; WjO means a bias term and XtO is set to 1. Note that aT denotes the transposed vector of a. Hereafter, a vector consisting of all parameters, (w5,"" wDT , is simply expressed as ~ = (<PI, ... , <P N ) T, where N (= nh + 2h + 1) denotes the dimension of ~. Then, the training error in the three-layer neural network can be defined as follows: where O'(u) represents a sigmoidal function, O'(u) = 1/(1 + e-U ). In this paper, we consider the following three penalty terms: N 02(~) = L l<Pkl, k=l (1) Hereafter, 01, O2, and 0 3 are referred to as the squared (Hinton, 1987; MacKay, 1992), absolute (Ishikawa, 1990; Williams, 1995), and normalized (Hanson & Pratt, 1989) penalty terms, respectively. Then, learning with one of these terms can be defined as the problem of minimizing the following objective function (3) where J.L is a penalty factor. 2.2 Second-order Algorithm with Penalty Term In order to minimize the objective function, we employ a newly invented secondorder learning algorithm based on a quasi-Newton method, called BPQ (Saito & Nakano, 1997), where the descent direction, ~~, is calculated on the basis of a partial BFGS update and a reasonably accurate step-length, .x, is efficiently calculated as the minimal point of a second-order approximation. Here, the partial BFGS update can be directly applied, while the step-length .x is evaluated as follows: .x = -VFi(~)~~T ~~T'V2 Fi(~)~~ (4) Second-order Learning Algorithm with Squared Penalty Term 629 The quadratic form for the training error term, .1.~T\72 f(~).1.~, can be calculated efficiently with the computational complexity of Nm + O(hm) by using the procedure of BPQ, while those for penalty terms are calculated as follows: N .1.~T\72rh(~).1.~ = L.1.¢>~, .1.~T\72n2(~).1.~ = 0, k=1 .1.~T\72n (~).1.~ = t (1 3¢>~).1.¢>~ 3 k=1 (1 + ¢>~)3 . (5) Note that, in the step-length calculation, .1.~T\72Fi(~).1.~ is basically assumed to be positive. The three terms have a different effect on it, Le., the squared penalty term always adds a non-negative value; the absolute penalty term has no effect; the normalized penalty term may add a negative value if many weight values are larger than ...jf73. This indicates that the squared penalty term has a desirable feature. Incidentally, we can employ other second-order learning algorithms such as SCG (M(ljller, 1993) or ass (Battiti, 1992), but BPQ worked the most efficiently among them in our own experience (Saito & Nakano, 1997). 3 EVALUATION BY EXPERIMENTS 3.1 Regression Problem By using a regression problem for a function y = (1- x + 2x2)e-O.5X2 , the learning performance of adding a penalty term was evaluated. In the experiment, a value of x was randomly generated in the range of [-4,4], and the corresponding value of y was calculated from x; each value of y was corrupted by adding Gaussian noise with a mean of 0 and a standard deviation of 0.2. The total number of training examples was set to 30. The number of hidden units was set to 5, where the initial values for the weights between the input and hidden units were independently generated according to a normal distribution with a mean of 0 and a standard deviation of 1; the initial values for the weights between the hidden and output units were set to 0, but the bias value at the output unit was initially set to the average output value of all training examples. The iteration was terminated when the gradient vector was sufficiently small (Le., 11\7 Fi ( ~) 112/ N < 10-12) or the total processing time exceeded 100 seconds. The penalty factor J.t was changed from tJ to 2-19 by multiplying by 2- 1; trials were performed 20 times for each penalty factor. Figure 1 shows the training examples, the true function, and a function obtained after learning without a penalty term. We can see that such a learning over-fitted the training examples to some degree. 3.2 Evaluation using Second-order Algorithm By using BPQ, an evaluation was made after adding each penalty term. Figure 2(a) compares the generalization performance, which was evaluated by using the average RMSE (root mean squared error) for a set of 5,000 test examples. The best possible RMSE level is 0.2 because each test example includes the same amount of Gaussian noise given to each training example. For each penalty term, the generalization performance was improved when J.t was set adequately, but the normalized 630 3 2 1 0 a '0 . -4 -2 a K. Saito and R. Nakano true function leaming result o 2 4 Figure 1: Learning problem average RMSE 0.8 0.6 II II '1:3" I I \ I , , , . . . ' 0.2 -+-r-T"T"T'T'"T"""""""I"'T""r-T"T"T'''T'T'''''''' CPU time (sec.) 100 , '....•. ~ ~ , , '1:3 ' ...... ~ " ... " " 10 .... <~" : ~-----~-'~--u-t--~'~~"~#'-~~~ penalty 1 ~ 2-5 2.10 2-15 2·20~ 2-5 2-10 2.15 2-20 ~ (a) Generalization performance (b) CPU time until convergence Figure 2: Comparison using second-order algorithm BPQ penalty term was the most unstable among the three, because it frequently got stuck in undesirable local minima. Figure 2(b) compares the processing time! until convergence. In comparison to the learning without a penalty term, the squared penalty term drastically decreased the processing time especially when f.1 was large, while the absolute penalty term did not converge when f.1 was large; the normalized penalty term generally required a larger processing time. Thus, only the squared penalty term improved the convergence performance more than 2 rv 100 times, keeping a better generalization performance for an adequate penalty factor. 3.3 Evaluation using First-order Algorithm By using BP, a similar evaluation was made after adding each penalty term. Here, we adopted Silva and Almeida's learning rate adaptation rule (Silva & Almeida, 1990), i.e., learning rate "'k for each weight <Pk is adjusted by the signs of two successive gradient values2. Figure 3(a) compares the generalization performance and Figure 3(b) compares the processing time until convergence, where the average processing time for the trials without a penalty term is not displayed because all trials did not converge within 100 seconds. For each penalty term, the generalization lOur experiments were done on SUN 8-4/20 computers. 2The increasing and decreasing parameters were set to 1.1 and 1/1.1, respectively, as recommended by (Silva & Almeida, 1990); if the value of the objective function increases, all learning rates are halved until the value decreases. Second-order Learning Algorithm with Squared Penalty Term average RMSE 0.8 -n1 ....... ~ 0.6 ~ without penalty 0.2 -+-rT""r-T"T"T"T"T".,....,.."T"T"T"I'''T'''I~1''''I'''\ CPU time (sec.) 100 ~----------10 -n1 ....... ~ - - ~ O. 1 """'''''''r-T''1I'''T''"'''''-'r-r"r-T"T''T"'r"T"'T''T"T'1 211 2-5 2-10 2-15 2-20~ 2!J 2'5 2-10 2-15 2-20 ~ (a) Generalization performance (b) CPU time until convergence Figure 3: Comparison using first-order algorithm BP 631 performance was improved when f.t was set adequately. Note that BP with the squared penalty term 01 required more processing time than BPQ with 0 1. As for the normalized penalty term 03, BP with 0 3 worked more stably than BPQ with 0 3, Incidentally, the generalization performance of BP without a penalty term was better than that of BPQ without it; we predict that this is because the effect of early stopping (Bishop, 1995) worked for BP. Actually, for the training examples, the average RMSE of BP without a penalty term was 0.138, while that of BPQ without it was 0.133. 3.4 Graphical Evaluation In order to graphically examine the reasons why the effect of the addition of each penalty term differed, we designed a simple problem; that is, learning a function Y = U(WIX) + U(W2X), where only two weights, WI and W2, are adjustable. In the three-layer network, the input and output layers consist of only one unit, while the hidden layer consists of two units. Note that the weights between the hidden units and the output unit are fixed at 1, there is no bias, and the activation function of hidden units is assumed to be u(x) = 1/(1 + exp( -x)). Each target value Yt was calculated from the corresponding input value Xt E {-0.2, -0.1, 0,0.1, 0.2} by setting (WI,W2) = (1,3). Figure 4 shows the learning trajectories on error contour maps with respect to WI and W2 during 100 iterations starting at (Wt,W2) = (-1,-3), where the penalty factor f.t was set to 0.1 or 0.01. Here, BPQ was used as a learning algorithm. The contours for the squared penalty term form ovals, making BPQ learn easily. When f.t = 0.1, the contours for the absolute penalty term form an almost square-like shape, and the learning trajectories oscillate near the origin (WI = W2 = 0), due to the discontinuity of the gradient function. The contours for the normalized penalty term form a valley, making BPQ's learning more difficult. 3.5 Determining Penalty Factor In general, for a given problem, we cannot know an adequate penalty factor in advance. Given a limited number of examples, we must find a reasonably adequate 632 K. Saito and R. Nakano w2 ~ ( 1.1=0.1 ) w2 5 5 0 0 0 0.5 -5 -5 -5 -5 5 Wl -5 5w1 -5 5 w1 w w2 w2 5 5 5 0 0 0 0.5 0.5 -5 -5 -5 -5 0 5 W1 -5 0 5 w1 -5 0 5 w1 Figure 4: Graphical evaluation penalty factor. The procedure of cross-validation (Stone, 1978) is adopted for this purpose. Since we knew the combination of the squared penalty term 0 1 and the second-order algorithm BPQ works very efficiently, we performed experiments using the above regression problem with exactly the same experimental conditions. Figure 5 shows the experimental results, where the procedure of cross-validation was implemented as a leave-one-out method, and the initial weight values for evaluating the cross-validation error were set as the learning results of the entire training examples. Figure 5(a) compares the average generalization error and the average cross-validation error. Although the cross-validation error was a pessimistic estimator of the generalization error, it showed the same tendency and was minimized at almost the same penalty factor. Figure 5(b) shows the average processing time and its standard deviation; although the processing time includes the cross-validation evaluation, we can see that the learning was performed quite efficiently. 4 CONCLUSION This paper investigated the efficiency of supervised learning with each of three penalty terms, by using first- and second-order learning algorithms, BP and BPQ. Our experiments showed that for a reasonably adequate penalty factor, the combination of the squared penalty term and the second-order algorithm drastically improves the convergence performance about 20 times over the other combinations, together with an improvement in the generalization performance. In the case of other second-order learning algorithms such as SCG or OSS, similar results are possible because the main difference between BPQ and those other algorithms involves only the learning efficiency. In the future, we plan to do further evaluations using larger-scale problems. Second-order Learning Algorithm with Squared Penalty Term average AMSE 0.8 0.6 0.4 cross-validation error generalization error . CPU time (sec.) 100 average ....... S.d. 10 ;' ... .. .,.# ... # ............ 1 0.2 ,. ... . .. 0.1 -++. r-T'"T'"T""""'T'"'I""T""I''''''''''''''''-r"'lr-T'"Ir""'I""I 20 2.5 2.10 2.15 2"201.1 2!J 2.5 2.10 2.15 2"201.1 (a) Generalization performance (b) CPU time until convergence Figure 5: Learning result References 633 Battiti, R. (1992) First- and second-order methods for learning between steepest descent and Newton's method. Neural Computation 4(2):141-166. Bishop, C.M. (1995) Neural networks for pattern recognition. Clarendon Press. Hanson, S.J. & Pratt, L. Y. (1989) Comparing biases for minimal network construction with back-propagation. In D. S. Touretzky (ed.), Advances in Neural Processing Systems, Volume 1, pp. 177-185. San Mateo, CA: Morgan Kaufmann. Hinton, G.E. (1987) Learning translation invariant recognition in massively parallel networks. In J. W. de Bakker, A. J. Nijman and P. C. Treleaven (eds.), Proceedings PARLE Conference on Parallel Architectures and Languages Europe, pp. 1-13. Berlin: Springer-Verlag. Ishikawa, M. (1990) A structural learning algorithm with forgetting of link weight. Tech. Rep. TR-90-7, Electrotechnical Lab. Tsukuba-City, Japan. MacKay, D.J.C. (1992) Bayesian interpolation. Neural Computation 4(3):415-447. M¢ller, M.F. (1993) A scaled conjugate gradient algorithm for fast supervised learning. Neural Networks 6(4):525-533. Poggio, T. & Girosi, F. (1990) Regularization algorithms for learning that are equivalent to multilayer networks. Science 247:978-982. Saito, K. & Nakano, R. (1997) Partial BFGS update and efficient step-length calculation for three-layer neural networks. Neural Computation 9(1):239-257 (in press). Silva, F.M. & Almeida, L.B. (1990) Speeding up backpropagation. In R. Eckmiller (ed.), Advanced Neural Computers, pp. 151-160. Amsterdam: North-Holland. Stone, M. (1978) Cross-validation: A review. Operationsforsch. Statist. Ser. Statistics B 9(1):111-147. Williams, P.M. (1995) Bayesian regularization and pruning using a Laplace prior. Neural Computation 7(1):117-143.
1996
78
1,278
Multi-effect Decompositions for Financial Data Modeling Lizhong Wu & John Moody Oregon Graduate Institute, Computer Science Dept., PO Box 91000, Portland, OR 97291 also at: Nonlinear Prediction Systems, PO Box 681, University Station, Portland, OR 97207 Abstract High frequency foreign exchange data can be decomposed into three components: the inventory effect component, the surprise infonnation (news) component and the regular infonnation component. The presence of the inventory effect and news can make analysis of trends due to the diffusion of infonnation (regular information component) difficult. We propose a neural-net-based, independent component analysis to separate high frequency foreign exchange data into these three components. Our empirical results show that our proposed multi-effect decomposition can reveal the intrinsic price behavior. 1 Introduction Tick-by-tick, high frequency foreign exchange rates are extremely noisy and volatile, but they are not simply pure random walks (Moody & Wu 1996). The price movements are characterized by a number of "stylized facts" ,including the following two properties: (1) short tenn, weak oscillations on a time scale of several ticks and (2) erratic occurrence of turbulence lasting from minutes to tens of minutes. Property (1) is most likely caused by the market makers' inventory effect (O'Hara 1995), and property (2) is due to surprise information, such as news, rumors, or major economic announcements. The price changes due to property (1) are referred to as the inventory effect component, and the changes due to property (2) are referred to as the surprise infonnation component. The price changes due to other infonnation is referred to as the regular infonnation component. IThis terminology is borrowed from the financial economics literature. For additional properties of high frequency foreign exchange price series. see (Guilaumet, Dacorogna. Dave. Muller. Olsen & Pictet 1994). 996 L. Wu and 1. E. Moody Due to the inventory effect, price changes show strong negative correlations on short time scales (Moody & Wu 1995). Because of the surprise infonnation effect, distributions of price changes are non-nonnal (Mandelbrot 1963). Since both the inventory effect and the surprise information effect are short tenn and temporary, their corresponding price components are independent of the fundamental price changes. However, their existence will seriously affect data analysis and modeling (Moody & Wu 1995). Furthennore, the most reliable component of price changes, for forecasting purposes, is the long tenn trend. The presence of high frequency oscillations and short periods ofturbulence make it difficult to identify and predict the changes in such trends, if they occur. In this paper, we propose a novel approach with the following price model: q(t) = CIPI(t) + C2P2(t) + C3P3(t) + e(t). (1) In this model, q(t) is the observed price series and PI (t), P2(t) and P3(t) correspond respectively to the regular infonnation component, the surprise infonnation component and the inventory effect component. PI (t), P2 (t) and P3 (t) are mutually independent and may individually be either iid or correlated. e(t) is process noise, and c}, C2 and C3 are scale constants. Our goal is to find PI (t), p2(t) and P3( t) given q(t). The outline of the paper is as follows. We describe our approach for multi-effect decomposition in Section 2. In Section 3, we analyze the decomposed price components obtained for the high frequency foreign exchange rates and characterize their stochastic properties. We conclude and discuss the potential applications of our multi-effect decomposition in Section 4. 2 Multi-effect Decomposition 2.1 Independent Source Separation The task of decomposing the observed price quotes into a regular infonnation component, a surprise infonnation component and an inventory effect component can be exactly fitted into the framework of independent source separation. Independent source separation can be described as follows: Assume that X = {Xi, i = 1, 2, ... , n} are the sensor outputs which are some superposition of unknown independent sources S = {Si' i = 1, 2, ... , m }. The task of independent source separation is to find a mapping Y = f (X), so that Y ~ AS, where A is an m x m matrix in which each row and column contains only one non-zero element. Approaches to separate statistically-independent components in the inputs include • Blind source separation (Jutten & Herault 1991), • Infonnation maximization (Linsker 1989), (Bell & Sejnowski 1995), • Independent component analysis, (Comon 1994), (Amari, Cichocki & Yang 1996), • Factorial coding (Barlow 1961). All of these approaches can be implemented by artificial neural networks. The network architectures can be linear or nonlinear, multi-layer perceptrons, recurrent networks or other context sensitive networks (pearlmutter & Parra 1997). We can choose a training criterion to minimize the energy in the output units, to maximize the infonnation transferred in the network, to reduce the redundancies between the outputs, or to use the Edgeworth expansion or Gram-Charlier expansion of a probability distribution, which leads to an analytic expression of the entropy in tenns of measurable higher order cumulants. Multi-effect Decompositions/or Financial Data Modeling 997 r1(t) Orthogonal Independent q(t) Multi-scale Smoothing e r2(t) Component ~(t) Decomposition r3(t) Analysis ~(t) Figure 1: System diagram of multi-effect decomposition for high frequency foreign exchange rates. q(t) are original price quotes, ri(t) are the reference inputs, and Pie t) are the decomposed components. For our price decomposition problem, the non-Gaussian nature of price series requires that the transfer function of the decomposition system be nonlinear. In general, the nonlinearities in the transfer function are able to pick up higher order moments of the input distributions and perfonn higher order statistical redundancy reduction between outputs. 2.2 Reference input selection In traditional approaches to blind source separation, nothing is assumed to be known about the inputs, and the systems adapt on-line and without a supervisor. This works only if the number of sensors is not less than the number of independent sources. If the number of sensors is less than that of sources, the sources can, in theory, be separated into disjoint groups (Cao & Liu 1996). However, the problem is ill-conditioned for most of the above practical approaches which only consider the case where the number of sensors is equal to the number of sources. In our task to decompose the multiple components of price quotes, the problem can be divided into two cases. If the prices are sampled at regular intervals, we can use price quotes observed in different markets, and have the number of sensors be equal to the number of price components. However, in the high frequency markets, the price quotes are not regularly spaced in time. Price quotes from different markets will not appear at the same time, so we cannot apply the price quotes from different markets to the system. In this case, other reference inputs are needed. Motivated by the use of reference inputs for noise canceling (Widrow, Glover, McCool, Kaunitz, Williams, Heam, Zeidler, Dong & Goodlin 1975), we generate three reference inputs from original price quotes. They are the estimates of the three desired components. In the following, we briefly describe our procedure for generating the reference inputs. By modeling the price quotes using a "True Price" state space model (Moody & Wu 1996) q(t) = rl (t) + r3(t) , (2) where rl (t) is an estimate of the infonnation component (True Price) and r3( t) is an estimate of the inventory effect component (additive noise), and by assuming that the True Price rl (t) is a fractional Brownian motion (Mandelbrot & Van Ness 1968), we can estimate rl (t) and r3 (t) with given q( t), (Moody & Wu 1996), as (3) m,n (4) 998 L. Wu and J. E. Moody Figure 2: Multi-effect decompositions for two segments of the DEMIUSD (log prices) extracted from September 1995. The three panels in each segment display the observed prices (the dotted curve in upper panel), the regular information component (solid curve in upper panel), the surprise information component (mid panel) and the inventory effect component (lower panel). where t/!::'(t) is an orthogonal wavelet function, Q::' is the coefficient of the wavelet transform of q(t), m is the index of the scales and n is the time index of the components in the wavelet transfer, S(m,B) is a smoothing function, and its parameters can be estimated using the EM algorithm (Womell & Oppenheim 1992). We then estimate the surprise information component as the residual between the information component and its moving average: r2(t) = r}(t) - set) set) is an exponential moving average of rl(t) and set) = (1 + a)rl(t) - as(t - 1) (5) (6) where a is a factor. Although it can be optimized based on the training data, we set a = -0.9 in our current work. Our system diagram for multi-effect decomposition is shown in Figure 1. Using multiscale decomposition Eqn(3) and smoothing techniques Eqn(6), we obtain three reference inputs. We can then separate the reference inputs into three independent components via independent component analysis using an artificial neural network. Figure 2 presents multieffect decompositions for two segments of the DEMIUSD rates. The first segment contains some impulses, and the corresponding surprise information component is able to catch such volatile movements. The second segment is basically down trending, so its surprise information component is comparatively flat. 3 Empirical Analysis 3.1 Mutually Independent Analysis Mutual independence of the variables is satisfied if the joint probability density function equals the product of the marginal densities, or equivalently, the characteristic function splits into the sum of marginal characteristic functions: g(X) = E7=1 gi(Xi). Taking the Taylor expansion of both sides of the above equation, products between different variables Xi in the left-hand side must be zero since there are no such terms in the right-hand side. Multi-effect Decompositions for Financial Data Modeling 999 Table 1: Comparisons between the correlation coefficients p (nonnalized) and the crosscumulants r (unnonnalized) of order 4 before and after independent component analysis (lCA). The DEMIUSD quotes for September 1995 is divided into 147 sub-sets of 1024 ticks. The results presented here are the median values. The last column is the absolute ratio of before ICA and after ICA. We note that all ratios are greater than I, indicating that after ICA, the components become more independent. Components CrossBefore After pairs Cumulants ICA ICA P12 0.56 0.14 r13 2.7e-14 7.8e-17 Pl(t) "" P2(t) r 22 -5.6e-15 9.2e-16 r31 2.0e-11 1.3e-13 P13 0.15 0.03 r13 2.le-15 1.6e-17 pl(t) "" P3(t) r22 -2.0e-15 -4.5e-16 r31 5.ge-12 6.ge-14 P23 0.17 0.04 r13 9.le-16 -5.0e-19 P2(t) "" P3(t) r22 1.2e-15 4.ge-17 r 31 3.6e-15 3.0e-17 We observe the cross-cumulants of order 4: r13 M13 - 3M2oMll r 22 = M22 - M20M02 - 2M?1 r 31 = M31 - 3M02Mll Absolute ratio 4.1 342.2 6.0 148.5 4.7 128.9 4.5 84.5 4.3 1806.0 24.3 119.6 (7) (8) (9) where M lei = E { x~ x~} denote the moments of order k + I. If x i and x j are independent, then their cross-cumulants must be zero (Comon 1994). Table I compares the crosscumulants before and after independent component analysis (ICA) for the DEMIUSD in September 1995. For reference, the correlation coefficients before and after ICA are also listed in the table. We see that after ICA, the components have become less correlated and thus more independent. 3.2 Autocorrelation Analysis Figure 3 depicts the autocorrelation functions of the changes in individual components and compares them to the original returns. We compute the short-run autocorrelations for the lags up to 50. Figure 3 gives the means and standard deviations for September 1995. From the figure, we can see that both the inventory effect component and the original returns show very similar autocorrelation functions, which are dominated by the significant negative, first-order autocorrelations. The mean values for the other orders are basically equal to zero. The autocorrelations of the regular information component and the surprise information component show positive correlations except at first order. These non-zero autocorrelations are hidden by noise in the original series. The autocorrelation function of the surprise information component decays faster than that of the regular information component. On average, it is below the 95% confidence band for lags larger than 20 ticks. The above autocorrelation analysis suggests the following. (I) Price changes due to the information effects are slightly trending on tick-by-tick time scales. The trend in the surprise information component is shorter term than that in the regular information component. 1000 0.6 0.4 a 0.2 ~ 0 mm.mmmmHHim!lH!ml!l)mtmm « -0.2 -0.4 0.6 0.4 ~ 0.2 ~ 0 " « -0.2 -0.4 0 10 20 30 40 0.6 0.4 ~ 0.2 ~ 0 " « -0.2 -0.4 0.6 0.4 ~ 0.2 ~ 0 " « -0.2 -0.4 0 L. Wu and J. E. Moody 10 20 30 40 Figure 3: Comparison of autocorrelation functions of the changes in the original observed prices (the upper-left panel), the inventory effect component (the lower-left panel), the regular information component (the upper-right panel) and the surprise infonnation component (the lower-right panel). The results presented are means and standard deviations, and the horizontal dotted lines represent the 95% confidence band. The DEMIUSD in September 1995 is divided into 293 sub-sets of 1024 ticks with overlapping of 512 ticks. (2) The autocorrelation function of original returns reflects only the price changes due to the inventory effect. This further confirms that the existence of short tenn memory can mislead the analysis of dependence on longer time scales. Subsequently, we can see the usefulness of the multi-effect decomposition. Our empirical results should be viewed as preliminary, since they may depend upon the choice of True Price model. Additional studies are ongoing. 4 Conclusion and Discussion We have developed a neural-net-based independent component analysis (ICA) for the multi-effect decomposition of high frequency financial data. Empirical results with foreign exchange rates have demonstrated that the decomposed components are mutually independent. The obtained regular infonnation component has recovered the trending behavior of the intrinsic price movements. Potential applications for multi-effect decompositions include: (1) outlier detection and filtering: Filtering techniques for removing various noisy effects and identifYing long tenn trends have been widely studied (see for example Assimakopoulos (1995». MuIti-effect decompositions provide us with an alternative approach. As demonstrated in Section 3, the regular infonnation component can, in most cases, catch relatively stable and longer tenn trends originally embedded in the price quotes. (2) devolatilization: Price series are heteroscedastic (Boilers lev, Chou & Kroner 1992). Devolatilization has been widely studied (see, for example, Zhou (1995». The regular infonnation component obtained from our multi-effect decomposition appears less volatile, and furthennore, its volatility changes more smoothly compared to the original prices. (3) mixture of local experts modeling: In most cases, one might be interested in only stable, long tenn trends of price movements. However, the surprise infonnation and inventory effect components are not totally useless. By decomposing the price series into three mutually Multi-effect Decompositionsfor Financial Data Modeling 1001 independent components, the prices can be modeled by a mixture of local experts (Jacobs, Jordan & Barto 1990), and better modeling perfonnances can be expected. References Amari, S., Cichocki, A. & Yang. H. (1996), A new learning algorithm for blind signal separation, in D. Touretzky. M. Mozer & M. Hasselmo, eds. 'Advances in Neural Infonnation Processing Systems 8', MIT Press: Cambridge, MA. Assimakopoulos. V. (1995), 'A successive filtering technique for identifying long-tenn trends'. Journal of Forecasting 14, 35-43. Barlow, H. (1961), Possible principles underlying the transfonnation of sensory messages, in W. Rosenblith, ed., 'Sensory Communication', MIT Press: Cambridge, MA, pp. 217-234. Bell, A. & Sejnowski, T. (1995), 'An infonnation-maximization approach to blind separation and blind deconvolution', Neural Computation 7(6), 1129-1159. Bollerslev, T., Chou, R. & Kroner. K. (1992). 'ARCH modelling in finance: A review of the theory and empirical evidence', Journal of Econometrics 8, 5-59. Cao, X. & Liu, R. (1996), 'General approach to blind source separation' • IEEE Transactions on Signal Processing 44(3), 562-569. Comon, P. (1994), 'Independent component analysis, a new concept?', Signal Process 36,287-314. Guilaumet, D., Dacorogna, M., Dave, R., Muller, U., Olsen, R. & Pictet, O. (1994), From the bird's eye to the microscope, a survey of new stylized facts of the intra-daily foreign exchange markets, Technical Report DMG.1994-04-06, Olsen & Associates, Zurich, Switzerland. Jacobs. R., Jordan, M. & Barto, A. (1990), Task decomposition through competition in a modular connectionist architecture: The what and where vision tasks, Technical Report COINS 90-27, Department of Brain & Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA. Jutten, C. & Herault, 1. (1991). 'Blind separation of sources. part I: An adaptive algorithm based on neuromimetic architecture' , Signal Process 24( 1), 1-10. Linsker, R. (1989), An application of the principle of maximum infonnation preservation to linear systems, in D. Touretzky, ed., 'Advances in Neural Infonnation Processing Systems 1', Morgan Kaufmann Publishers, San Francisco, CA. Mandelbrot, B. (1963), 'The variation of certain specUlative prices' , Journal of Business 36, 394--419. Mandelbrot, B. & Van Ness, J. (1968), 'Fractional Brownian motion, fractional noise. and applications', SIAM Review 10. MoOdy, 1. & Wu. L. (1995), Statistical analysis and forecasting of high frequency foreign exchange rates, in 'The First International Conference on High Frequency Data in Finance', Zurich, Switzerland. Moody, J. & Wu, L. (1996), What is the 'True Price '? - state space models for high frequency financial data, in S. Amari, L. Xu, L. Chan, I. King & K. Leung, eds, 'Progress in Neural Infonnation Processing (Proceedings of ICONIPS*96, Hong Kong)', Springer-Verlag, Singapore. pp. 697704, Vo1.2. O'Hara, M. (1995), Market Microstructure Theory, Blackwell Business. Pearl mutter, B. A. & Parra, L. (1997), Maximum likelihood blind source separation: a contextsensitive generalization ofICA, in M. Mozer, M. Jordan & T. Petsche, eds, 'Advances in Neural Infonnation Processing Systems 9' , MIT Press: Cambridge, MA. Widrow, B., Glover, J., McCool, 1., Kaunitz, J., Williams, C., Hearn, R., Zeidler, J., Dong, E. & Goodlin, R. (1975), 'Adaptive noise cancelling: principles and applications', Proceedings of IEEE 63(12), 1692-1716. Wornell, G. & Oppenheim, A. (1992), 'Estimation of fractal signals from noisy measurements using wavelets', IEEE Transactions on Signal Processing 40(3), 611-623. Zhou, B. (1995), Forecasting foreign exchange rates series subject to de-volatilization, in 'The First International Conference on High Frequency Data in Finance', Zurich, Switzerland. PART IX CONTROL, NAVIGATION AND PLANNING
1996
79
1,279
A Micropower Analog VLSI HMM State Decoder for Wordspotting John Lazzaro and John Wawrzynek CS Division, UC Berkeley Berkeley, CA 94720-1776 lazzaroGcs.berkeley.edu. johnwGcs.berkeley.edu Richard Lippmann MIT Lincoln Laboratory Room S4-121, 244 Wood Street Lexington, MA 02173-0073 rplGsst.ll.mit.edu Abstract We describe the implementation of a hidden Markov model state decoding system, a component for a wordspotting speech recognition system. The key specification for this state decoder design is microwatt power dissipation; this requirement led to a continuoustime, analog circuit implementation. We characterize the operation of a 10-word (81 state) state decoder test chip. 1. INTRODUCTION In this paper, we describe an analog implementation of a common signal processing block in pattern recognition systems: a hidden Markov model (HMM) state decoder. The design is intended for applications such as voice interfaces for portable devices that require micropower operation. In this section, we review HMM state decoding in speech recognition systems. An HMM speech recognition system consists of a probabilistic state machine, and a method for tracing the state transitions of the machine for an input speech waveform. Figure 1 shows a state machine for a simple recognition problem: detecting the presence of keywords ("Yes," "No") in conversational speech (non-keyword speech is captured by the "Filler" state). This type of recognition where keywords are detected in unconstrained speech is called wordspotting (Lippmann et al., 1994). 728 1. Lazzaro. 1. Wawrzynek and R. Lippmann Filler Figure 1. A two-keyword ("Yes," states 1-10, "No," states 11-20) HMM. Our goal during speech recognition is to trace out the most likely path through this state machine that could have produced the input speech waveform. This problem can be partially solved in a local fashion, by examining short (80 ms. window) overlapping (15 ms. frame spacing) segments of the speech waveform. We estimate the probability bi(n) that the signal in frame n was produced by state i, using static pattern recognition techniques. To improve the accuracy of these local estimates, we need to integrate information over the entire word. We do this by creating a set of state variables for the machine, called likelihoods, that are incrementally updated at every frame. Each state i has a real-valued likelihood <Pi(n) associated with it. Most states in Figure 1 have a stereotypical form: a state i that has a self-loop input, an input from state i-I, and an output to state i + 1, with the self-loop and exit transitions being equally probable. For states in this topology, the update rule log{<Pi{n)) = log(bi(n)) + log(<pi{n - 1) + <Pi-dn - 1)) (1) lets us estimate the "log likelihood" value log{<pi(n)) for the state i; a log encoding is used to cope with the large operating range of <pj(n) values. Log likelihoods are negative numbers, whose magnitudes increase with each frame. We limit the range of log likelihood values by using a renormalization technique: if any log likelihood in the system falls below a minimum value, a positive constant is added to all log likelihoods in the machine. Figure 2 shows a complete system which uses HMM state decoding to perform wordspotting. The "Feature Generation" and "Probability Generation" blocks comprise the static pattern recognition system, producing the probabilities bi(n) at each frame. The "State Decode" block updates the log likelihood variables log( <Pi (n)). The "Word Detect" block uses a simple online algorithm to flag the occurrence of a word. Keyword end-state log likelihoods are subtracted by the filler log likelihood, and when this difference exceeds a fixed threshold a keyword is detected. Feature Generation Speech Input 1 Probability Generation 21 State 10 20 Decode 21 Word Detect Figure 2. Block diagram for the two-keyword spotting system. A Micropower Analog VLSI HMM State Decoder for Wordspotting 729 2. ANALOG CIRCUITS FOR STATE DECODING Figure 3a shows an analog discrete-time implementation of Equation 1. The delay element (labeled Z-I) acts as a edge-triggered sampled analog delay, with full-scale voltage input and output. The delay element is clocked at the frame rate of the state decoder (15 ms. clock period). The "combinatorial" analog circuits must settle within the clock period. A clock period of 15 ms. allows a relatively long settling time, which enables us to make extensive use of submicroampere currents in the circuit design. The microwatt power consumption design specification drives us to use such small currents. As a result of submicroampere circuit operation, the MOS transistors in Figure 3a are operating in the weak-inversion regime. (1) gi[n) gi[n - I] (1) (2) (3) L...J (4) L...J (4) (a) Vm log[if>i[n]) (b) Vm log[if>i[nlJ (c) Vi(t) Figure 3. (a) Analog discrete-time single-state decoder. (b) Enhanced version of (a), includes the renormalization system. (c) Continuous-time extension of (b). 730 1. Lazzaro. 1. Wawrzynek and R. Lippmann Equation 1 uses two types of variables: probabilities and log likelihoods. In the implementation shown in Figure 3, we choose unidirectional current as the signal type for probability, and large-signal voltage as the signal type for log likelihood. We can understand the dimensional scaling of these signal types by analyzing the floating-well transistor labeled (4) in Figure 3a. The equation Ih Vm log(¢i(n)) = Vm log(bi(n)) + gi(n -1) + Vm log( T) o (2) describes the behavior of this transistor, where Vm = (VoIKp) In(10), gi(n - 1) is the output of the delay element, and 10, K and Vo are MOS parameters. Both 10 and K in Equation 2 are functions of V.b. However, the floating-well topology of the transistor ensures V. b = 0 for this device. The input probability bj(n) is scaled by the unidirectional current h, defining the current flowing through the transistor. The current h is the largest current that keeps the transistor in the weak-inversion regime. We define I, to be the smallest value for hbi(n) that allows the circuit to settle within the clock period. The ratio Ihl I, sets the supported range of bi(n). In the test-chip fabrication process, Ihl I, ~ 10,000 is feasible, which is sufficient for accurate wordspotting. Likewise, the unitless log(¢i(n)) is scaled by the voltage Vm to form a large-signal voltage encoding of log likelihood. A nominal value for Vm is 85m V in the test-chip process. To support a log likelihood range of 35 (the necessary range for accurate wordspotting) a large-signal voltage range of 3 volts (i.e. 35Vm ) is required. The term gi(n - 1) in Equation 2 is shown as the output of the circuit labeled (1) in Figure 3a. This circuit computes a function that approximates the desired expression Vmlog(¢i(n - 1) + ¢i-l(n -1)), if the transistors in the circuit operate in the weak-inversion regime. The computed log likelihood log( ¢i (n)) in Equation 1 decreases every frame. The circuit shown in Figure 3a does not behave in this way: the voltage Vmlog(¢j(n)) increases every frame. This difference in behavior is attributable to the constant term Vm log(IhIIo) in Equation 2, which is not present in Equation 1, and is always larger than the negative contribution from Vm log(bj(n)). Figure 3b adds a new circuit (labeled (2)) to Figure 3a, that allows the constant term in Equation 2 to be altered under control of the binary input V. If V is Vdd , the circuit in Figure 3b is described by Vm log(¢j(n)) = Vm log(bj(n)) + gi(n - 1) + Vm log( 1;;0), (3a) v where the term Vmlog((hIo)II;) should be less than or equal to zero. If V is grounded, the circuit is described by Ih Vm log(¢j(n)) = Vm log(bj(n)) + gi(n - 1) + Vm log( T)' (3b) v where the term Vm log(Ihl Iv) should have a positive value of at least several hundred millivolts. The goal of this design is to create two different operational modes for the system. One mode, described by Equation 3a, corresponds to the normal state decoder operation described in Equation 1. The other mode, described by Equation A Micropower Analog VLSI HMM State Decoder for Wordspotting 731 3b, corresponds to the renormalization procedure, where a positive constant is added to all likelihoods in the system. During operation, a control system alternates between these two modes, to manage the dynamic range of the system. Section 1 formulated HMMs as discrete-time systems. However, there are significant advantages in replacing the z-t element in Figure 3b with a continuous-time delay circuit. The switching noise of a sampled delay is eliminated. The power consumption and cell area specifications also benefit from continuous-time implementation. Fundamentally, a change from discrete-time to continuous-time is not only an implementation change, but also an algorithmic change. Figure 3c shows a continuoustime state decoder whose observed behavior is qualitatively similar to a discrete-time decoder. The delay circuit, labeled (3), uses a linear transconductance amplifier in a follower-integrator configuration. The time constant of this delay circuit should be set to the frame rate of the corresponding discrete-time state decoder. For correct decoder behavior over the full range of input probability values, the transconductance amplifer in the delay circuit must have a wide differential-inputvoltage linear range. In the test chip presented in this paper, an amplifier with a small linear range was used. To work around the problem, we restricted the input probability currents in our experiments to a small multiple of II. Figure 4 shows a state decoding system that corresponds to the grammar shown in Figure 1. Each numbered circle corresponds to the circuit shown in Figure 3c. The signal flows of this architecture support a dense layout: a rectangular array of single-state decoding circuits, with input current signal entering from the top edge of the array, and end-state log likelihood outputs exiting from the right edge of the array. States connect to their neighbors via the Vi-l(t) and Vi(t) signals shown in Figure 3c. For notational convenience, in this figure we define the unidirectional current Pi(t) to be Ihbi{t). In addition to the single-state decoder circuit, several other circuits are required. The "Recurrent Connection" block in Figure 4 implements the loopback connecting the filled circles in Figure 1. We implement this block using a 3-input version of the voltage follower circuit labeled (1) in Figure 3c. A simple arithmetic circuit implements the "Word Detect" block. To complete the system, a high fan-in/fanout control circuit implements the renormalization algorithm. The circuit takes as input the log likelihood signals from all states in the system, and returns the binary signal V to the control input of all states. This control signal determines whether the single-state decoding circuits exhibit normal behavior (Equation 3a) or renormalization behavior (Equation 3b). Pl Pll P2 Pl2 P3 Pl3 P4 Pl4 Ps PlS P6 Pl6 P7 Pl7 P8 P18 P9 Pl9 PlO P20P21 Figure 4. State decoder system for grammar shown in Figure 1. 732 1. Lazzaro, 1. Wawrzynek and R. Lippmann 3. STATE DECODER TEST CHIP We fabricated a state decoder test chip in the 21lm, n-well process of Orbit Semiconductor, via MOSIS. The chip has been fully tested and is functional. The chip decodes a grammar consisting of eight ten-state word models and a filler state. The state decoding and word detection sections of the chip contain 2000 transistors, and measure 586 x 28071lm (586x2807.X, ). = 1.0Ilm). In this section, we show test results from the chip, in which we apply a temporal pattern of probability currents to the ten states of one word in the model (numbered 1 through 10) and observe the log likelihood voltage of the final state of the word (state 10). Figure 5 contains simulated results, allowing us to show internal signals in the system. Figure 5a shows the temporal pattern of input probability currents PI ... PIO that correspond to a simulated input word. Figure 5b shows the log likelihood voltage waveform for the end-state of the word (state 10). The waveform plateaus at L h , the limit of the operating range of the state decoder system. During this plateau this state has the largest log likelihood in the system. Figure 5c is an expanded version of Figure 5b, showing in detail the renormalization cycles. Figure 5d shows the output computed by the "Word Detect" block in Figure 4. Note the smoothness of the waveform, unlike Figure 5c. By subtracting the filler-state log likelihood from the end-state log likelihood, the Word Detect block cancels the common-mode renormalization waveform. Figure 6 shows a series of four experiments that confirm the qualitative behavior of the state decoder system. This figure shows experimental data recorded from the fabricated test chip. Each experiment consists of playing a particular pattern of input probability currents PI ... PIO to the state decoder many times; for each repetition, a certain aspect of the playback is systematically varied. We measure the peak value of the end state log likelihood during each repetition, and plot this value as a function of the varied input parameter. For each experiment shown in Figure 6, the left plot describes the input pattern, while the right plot is the measured endstate log likelihood data. The experiment shown in Figure 6a involves presenting complete word patterns of varying durations to the decoder. As expected, words with unrealistically short durations have end-state responses below L h , and would not produce successful word detection. --> Pi....../'-G '-' 0.2 -A.."0 "0 3.5 P3~ 0 0 :::---A--0 0 P5 ----"'-:S :S I -0.4 ~ v v -;1 P7---"...:.= ...:.= ~ ;.:3 ;.:3 3.0 P9----"-bO -1.0 ~ j 0 400 200 350 100 400 700 (ms) (ms) (ms) (ms) (a) (b) (c) (d) Figure 5. Simulation of state decoder: (a) Inputs patterns, (b), (c) End-state response, (d) Word-detection response. A Micropower Analog VLSI HMM State Decoder for Wordspotting Pl~ ~L, I Pl~ P3~ "0 3.4 P3~ P5~ 0 (a) P5~ 0 P7~ :5 P7--" P9~ 'ii 3. P9 ~ ~ 1 400 700 t Word Length (rns) Pl~ >" Lh Pl -P3~ "0 3.4 / P3.....,. P5~ 0 P5~ 0 (b) P7 =:f:. :5 P7~ P9~ 'ii 3. P9~ ~ :3 t 733 F:L[(C) ~ 2.8 :3 1 10 "0 o o :S ] :3 Last State (d) First State Figure 6. Measured chip data for end-state likelihoods for long, short, and incomplete pattern sequences. The experiment shown in Figure 6b also involves presenting patterns of varying durations to the decoder, but the word patterns are presented "backwards," with input current PIO peaking first, and input current PI peaking last. The end-state response never reaches L h • even at long word durations, and (correctly) would not trigger a word detection. The experiments shown in Figure 6c and 6d involve presenting partially complete word patterns to the decoder. In both experiments, the duration of the complete word pattern is 250 ms. Figure 6c shows words with truncated endings, while Figure 6d shows words with truncated beginnings. In Figure 6c, end-state log likelihood is plotted as a function of the last excited state in the pattern; in Figure 6d, end-state log likelihood is plotted as a function of the first excited state in the pattern. In both plots the end-state log likelihood falls below Lh as significant information is removed from the word pattern. While performing the experiments shown in Figure 6, the state-decoder and worddetection sections of the chip had a measured average power consumption of 141 nW (Vdd = 5v). More generally, however, the power consumption, input probability range, and the number of states are related parameters in the state decoder system. Acknowledgments We thank Herve Bourlard, Dan Hammerstrom, Brian Kingsbury, Alan Kramer, Nelson Morgan, Stylianos Perissakis, Su-lin Wu, and the anonymous reviewers for comments on this work. Sponsored by the Office of Naval Research (URI-NOOOI492-J-1672) and the Department of Defense Advanced Research Projects Agency. Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the United States Air Force. Reference Lippmann, R. P., Chang, E. I., and Jankowski, C. R. (1994). "Wordspotter training using figure-of-merit back-propagation," Proceedings International Conference on Acoustics, Speech, and Signal Processing, Vol. 1, pp. 389-392.
1996
8
1,280
Microscopic Equations in Rough Energy Landscape for Neural Networks K. Y. Michael Wong Department of Physics, The Hong Kong University of Science and Technology, Clear Water Bay, Kowloon, Hong Kong. E-mail: phkywong@usthk.ust.hk Abstract We consider the microscopic equations for learning problems in neural networks. The aligning fields of an example are obtained from the cavity fields, which are the fields if that example were absent in the learning process. In a rough energy landscape, we assume that the density of the local minima obey an exponential distribution, yielding macroscopic properties agreeing with the first step replica symmetry breaking solution. Iterating the microscopic equations provide a learning algorithm, which results in a higher stability than conventional algorithms. 1 INTRODUCTION Most neural networks learn iteratively by gradient descent. As a result, closed expressions for the final network state after learning are rarely known. This precludes further analysis of their properties, and insights into the design of learning algorithms. To complicate the situation, metastable states (i.e. local minima) are often present in the energy landscape of the learning space so that, depending on the initial configuration, each one is likely to be the final state. However, large neural networks are mean field systems since the examples and weights strongly interact with each other during the learning process. This means that when one example or weight is considered, the influence of the rest of the system can be regarded as a background satisfying Some averaged properties. The situation is similar to a number of disordered systems such as spin glasses, in which mean field theories are applicable (Mezard, Parisi & Virasoro, 1987). This explains the success of statistical mechanical techniques such as the replica method in deriving the macroscopic properties of neural networks, e.g. the storage capacity (Gardner & Derrida 1988), generalization ability (Watkin, Rau & Biehl 1993). The replica Microscopic Equations in Rough Energy Landscape for Neural Networks 303 method, though, provides much less information on the microscopic conditions of the individual dynamical variables. An alternative mean field approach is the cavity method. It is a generalization of the Thouless-Anderson-Palmer approach to spin glasses, which started from microscopic equations of the system elements (Thouless, Anderson & Palmer, 1977). Mezard applied the method to neural network learning (Mezard, 1989). Subsequent extensions were made to the teacher-student perceptron (Bouten, Schietse & Van den Broeck 1995), the AND machine (Griniasty, 1993) and the multiclass perceptron (Gerl & Krey, 1995). They yielded macroscopic properties identical to the replica approach, but the microscopic equations were not discussed, and the existence of local minima was neglected. Recently, the cavity method was applied to general classes of single and multilayer networks with smooth energy landscapes, i.e. without the local minima (Wong, 1995a). The aligning fields of the examples satisfy a set of microscopic equations. Solving these equations iteratively provides a learning algoirthm, as confirmed by simulations in the maximally stable perceptron and the committee tree. The method is also useful in solving the dynamics of feedforward networks which were unsolvable previously (Wong, 1995b). Despite its success, the theory is so far applicable only to the regime of smooth energy landscapes. Beyond this regime, a stability condition is violated, and local minima begin to appear (Wong, 1995a). In this paper I present a mean field theory for the regime of rough energy landscapes. The complete analysis will be published elsewhere and here I sketch the derivations, emphasizing the underlying physical picture. As shown below, a similar set of microscopic equations hold in this case, as confirmed by simulations in the committee tree. In fact, we find that the solutions to these equations have a higher stability than other conventional learning algorithms. 2 MICROSCOPIC EQUATIONS FOR SMOOTH ENERGY LANDSCAPES We proceed by reviewing the cavity method for the case of smooth energy landscapes. For illustration we consider the single layer neural network (for two layer networks see Wong, 1995a). There are N » 1 input nodes {Sj} connecting to a single output node by the synaptic weights {Jj}. The output state is determined by the sign of the local field at the output node, i.e. Sout = sgn(Lj JjSj ). Learning a set of p examples means to find the weights {Jj} such that the network gives the correct input-to-output mapping for the examples. If example J.l maps the inputs Sf to the output 01-', then a successful learning process should find a weight vector Jj such that sgn(Lj Jj~j) = 1, where ~j = 01-' Sf. Thus the usual approach to learning is to first define an energy function (or error function) E = Ll-'g(AI-')' where AI-' == Lj Jj~f /VN are the aligning fields, i.e. the local fields in the direction of the correct output, normalized by the factor VN. For example, the Adatron algorithm uses the energy function g(A) = (I\: - A)6(1\: - A) where I\: is the stability parameter and 6 is the step function (Anlauf & Biehl, 1989). Next, one should minimize E by gradient descent dynamics. To avoid ambiguity, the weights are normalized to '" . S~ = '" . J~ = N L...J J L...J J . The cavity method uses a self-consistency argument to consider what happens when a set of p examples is expanded to p + 1 examples. The central quantity in this method is the cavity field. For an added example labelled 0, the cavity field is the aligning field when it is fed to a network which learns examples 1 to p (but 304 K. Y. M. Wong never learns example 0), i.e. to == Ej JjeJ 1v'N. Since the original network has no information about example 0, Jj and eJ are uncorrelated. Thus the cavity field obeys a Gaussian distribution for random example inputs. After the network has learned examples 0 to p, the weights adjust from {Jj} to {Jj}, and the cavity field to adjusts to the generic aligning field Ao. As shown schematically in Fig. l(a), we assume that the adjustments of the aligning fields of the original examples are small, typically of the order O(N-l/2). Perturbative analysis concludes that the aligning field is a well defined function of the cavity field, i.e. Ao = A(to) where A(t) is the inverse function of t = A + ,9' (A), (1) and, is called the local susceptibility. The cavity fields satisfy a set of self-consistent equations t JJ = I)A(tv) - tv]QVJJ + aXA(t JJ ) (2) vtJJ where QVJJ = Lj e;ej IN . X is called nonlocal susceptibility, and a == piN. The weights Jj are given by (3) Noting the Gaussian distribution of the cavity fields, the macroscopic properties of the neural network, such as the storage capacity, can be derived, and the results are identical to those obtained by the replica method (Gardner & Derrida 1988). However, the real advantage of the cavity method lies in the microscopic information it provides. The above equations can be iterated sequentially, resulting in a general learning algorithm. Simulations confirm that the equations are satisfied in the single layer percept ron , and their generalized version holds in the committee tree at low loading (Wong, 1995a). E E J a. J J Figure 1: Schematic drawing of the change in the energy landscape in the weight space when example 0 is added, for the regime of (a) smooth energy landscape, (b) rough energy landscape. Microscopic Equations in Rough Energy Landscape for Neural Networks 305 3 MICROSCOPIC EQUATIONS FOR ROUGH ENERGY LANDSCAPES However, the above argument holds under the assumption that the adjustment due to the addition of a new example is controllable. We can derive a stability condition for this assumption, and we find that it is equivalent to the AlmeidaThouless condition in the replica method (Mezard, Parisi & Virasoro, 1987). An example for such instability occurs in the committee tree, which consists of hidden nodes a = 1, ... , K with binary outputs, each fed by K nonoverlapping groups of N / K input nodes. The output of the committee tree is the majority state of the K hidden nodes. The solution in the cavity method minimizes the change from the cavity fields {tal to the aligning fields {Aa }, as measured by La(Aa -ta)2 in the space of correct outputs. Thus for a stability parameter K, Aa = K when ta < K and the value of ta is above median among the K hidden nodes, otherwise Aa = tao Note that a discontinuity exists in the aligning field function. Now suppose ta < K is the median, but the next highest value tb happens to be slightly less than ta. Then the addition of example ° may induce a change from tb < ta to tbO > taO· Hence AbO changes from tb to K whereas Aao changes from K to taO. The adjustment of the system is no longer small, and the previous perturbative analysis is not valid. In fact, it has been shown that all networks having a gap in the aligning field function are not stable against the addition of examples (Wong, 1995a). To consider what happens beyond the stability regime, one has to take into account the rough energy landscape of the learning space. Suppose that the original global minimum for examples 1 to p is a. After adding example 0, a nonvanishing change to the system is induced, and the global minimum shifts to the neighborhood of the local minimum 13, as schematically shown in Fig. 1 (b). Hence the resultant aligning fields Ag are no longer well-defined functions of the cavity fields tg. Instead they are well-defined functions of the cavity fields tg. Nevertheless, one may expect that correlations exist between the states a and 13. Let ViiO be the correlation between the network states, i.e. (Jj J1) = ViiO. Since both states a and 13 are determined in the absence of the added example 0, the correlation (tgtg) = ViiO as well. Knowing that both tg and tg obey Gaussian distributions, the cavity field distribution can be determined if we know the prior distribution of the local minima. At this point we introduce the central assumption in the cavity method for rough energy landscapes: we assume that the number of local minima at energy E obey an exponential distribution d~( E) = C exp( -wE)dE. Similar assumptions have been used in specifying the density of states in disordered systems (Mezard, Parisi & Virasoro 1987). Thus for single layer networks (and for two layer networks with appropriate generalizations), the cavity field ditribution is given by P(ti3jt<~) = G(tgltg)exp[-w~E(-\(tg))] (4) o 0 J dtgG(tg Itg) exp[-w~E(-\(tg))]' where G(tg Itg) is a Gaussian distribution. w is a parameter describing the distribution, and -\(tg) is the aligning field function. The weights J1 are given by J1 = (1 - ax)-l ~ 2)-\(t~) - t~]~f. (5) I' Noting the Gaussian distribution of the cavity fields, self-consistent equations for both qo and the local susceptibility 'Y can be derived. 306 K. Y. M. Wong To determine the distribution of local minima, namely the parameters C and w, we introduce a "free energy" F(p, N) for p examples and N input nodes, given by d~(E) = exp[w(F(p, N) - E)]dE. This "free energy" determines the averaged energy of the local minima and should be an extensive quantity, i.e. it should scale as the system size. Cavity arguments enable us to find an expression F (p + 1, N) F(p, N). Similarly, we may consider a cavity argument for the addition of one input node, expanding the network size from N to N + l. This yields an expression for F(p, N + 1) - F(p, N). Since F is an extensive quantity, F(p, N) should scale as N for a given ratio 0' = p/ N. This implies F N = O'(F(p + 1, N) - F(p, N)) + (F(p, N + 1) - F(p, N)). (6) We have thus obtained an expression for the averaged energy of the local minima. Minimizing the free energy with respect to the parameter w gives a self-consistent equation. The three equations for qo, 'Y and w completely determines the model. The macroscopic properties of the neural network, such as the storage capacity, can be derived, and the results are identical to the first step replica symmetry breaking solution in the replica method. It remains to check whether the microscopic equations have been modified due to the roughening of the energy landscape. It turns out that while the cavity fields in the initial state 0' do not satisfy the microscopic equations (2), those at the final metastable state {3 do, except that the nonlocal susceptibility X has to be replaced by its average over the distribution of the local minima. In fact, the nonlocal susceptibility describes the reactive effects due to the background examples, which adjust on the addition of the new example. (Technically, this is called the Onsager reaction.) The adjustments due to hopping between valleys in a rough energy landscape have thus been taken into account. 4 SIMULATION RESULTS To verify the theory, I simulate a committee tree learning random examples. Learning can be done by the more conventional Least Action algorithm (Nilsson 1965), or by iterating the microscopic equations. We verify that the Least Action algorithm yields an aligning field function ..\(t) consistent with the cavity theory. Suppose the weights from input j to hidden node a is given by Jaj = 2:1' xal'~j /..IN. Comparing with Jaj = (1- O'X)-l 2:1'(Aal' tal')~j /..IN, we estimate the nonlocal susceptibility X by requiring the distribution of tal' == Aal' - (1 - O'X)xal' to have a zero first moment. tal' is then an estimate of tal" Fig. 2 shows the resultant relation between Aal' and tal" It agrees with the predictions of the cavity theory. Fig. 3 shows the values of the stability parameter K, measured from the Least Action algorithm and the microscopic equations. They have better agreement with the predictions of the rough energy landscape (first step replica symmetry breaking solution) rather than the smooth energy landscape (replica symmetric solution). Note that the microscopic equations yield a higher stability than the Least Action algorithm. Microscopic Equations in Rough Energy Landscape for Neural Networks 307 5 '" '" '" 3 .", "C Q) ~ 1 - - - -0) _. c c -1 .2' « ,. -3 , .' '" '" -5 -5 -3 -1 1 3 5 Cavity field Figure 2: The aligning fields versus the cavity fields for a branch of the committee tree with K = 3, a = 0.8 and N = 600. The dashed line is the prediction of the cavity theory for the regime of rough energy landscape. 2.0 1.5 ~ 1.0 0.5 0.0 0.0 0.5 1.0 1.5 2.0 a. Figure 3: The stability parameter K, versus the storage level a in the committee tree with K = 3 for the cavity theory of: (a) smooth energy landscape (dashed line), (b) rough energy landscape (solid line), and the simulation of: (c) iterating the microscopic equations (circles), (d) the Least Action algorithm (squares). Error bars are smaller than the size of the symbols. 5 CONCLUSION In summary, we have derived the microscopic equations for neural network learning in the regime of rough energy landscapes. They turn out to have the same form as in the case of smooth energy landscape, except that the parameters are averaged over the distribution of local minima. Iterating the equations result in a learning algorithm, which yields a higher stability than more conventional algorithms in the committee tree. However, for high loading, the iterations may not converge. 308 K. Y. M. Wong The success of the present scheme lies its ability to take into account the underlying physical picture of many local minima of comparable energy. It correctly describes the experience that slightly different training sets may lead to vastly different neural networks. The stability parameter predicted by the rough landscape ansatz has a better agreement with simulations than the smooth one. It provides a physical interpretation of the replica symmetry breaking solution in the replica method. It is possible to generalize the theory to the physical picture with hierarchies of clusters of local minima, which corresponds to the infinite step replica symmetry breaking solution, though the mathematics is much more involved. Acknowledgements This work is supported by the Hong Kong Telecom Institute of Information Techology, HKUST. References Anlauf, J .K, & Biehl, M. (1989) The AdaTron: an adaptive perceptron algorithm. Europhysics Letters 10(7) :687-692. Bouten, M., Schietse, J . & Van den Broeck, C. (1995) Gradient descent learning in perceptrons: A review of its possibilities. Physical Review E 52(2):1958-1967. Gardner, E. & Derrida, B. (1988) Optimal storage properties of neural network models. Journal of Physics A : Mathematical and General 21(1) :271-284. Gerl, F. & Krey, U. (1995) A Kuhn-Tucker cavity method for generalization with applications to perceptrons with Ising and Potts neurons. Journal of Physics A: Mathematical and General 28(23):6501-6516. Griniasty, M. (1993) "Cavity-approach" analysis of the neural-network learning problem. Physical Review E 47(6):4496-4513. Mezard, M. (1989) The space of interactions in neural networks: Gardner's computation with the cavity method. Journal of Physics A: Mathematical and General 22(12):2181-2190. Mezard, M., Parisi, G. & Virasoro, M. (1987) Spin Glass Theory and Beyond. Singapore: World Scientific. Nilsson, N.J. (1965) Learning Machines. New York: McGraw-Hill. Thouless, D.J., Anderson, P.W. & Palmer, R.G. (1977) Solution of 'solvable model of a spin glass'. Philosophical Magazin,e 35(3) :593-601. Watkin, T.L.H., Rau, A. & Biehl, M. (1993) The statistical mechanics of learning a rule. Review of Modern Physics 65(2) :499-556. Wong, KY.M. (1995a) Microscopic equations and stability conditions in optimal neural networks. Europhysics Letters 30(4):245-250. Wong, KY.M. (1995b) The cavity method: Applications to learning and retrieval in neural networks. In J.-H. Oh, C. Kwon and S. Cho (eds.), Neural Networks: The Statistical Mechanics Perspective, pp. 175-190. Singapore: World Scientific.
1996
80
1,281
Dynamic features for visual speechreading: A systematic comparison Michael S. Grayl,a, Javier R. Movellanl , Terrence J. Sejnowski2,3 Departments of Cognitive Sciencel and Biology2 University of California, San Diego La Jolla, CA 92093 and Howard Hughes Medical Institute3 Computational Neurobiology Lab The Salk Institute, P. O. Box 85800 San Diego, CA 92186-5800 Email: mgray, jmovellan, tsejnowski@ucsd.edu Abstract Humans use visual as well as auditory speech signals to recognize spoken words. A variety of systems have been investigated for performing this task. The main purpose of this research was to systematically compare the performance of a range of dynamic visual features on a speechreading task. We have found that normalization of images to eliminate variation due to translation, scale, and planar rotation yielded substantial improvements in generalization performance regardless of the visual representation used. In addition, the dynamic information in the difference between successive frames yielded better performance than optical-flow based approaches, and compression by local low-pass filtering worked surprisingly better than global principal components analysis (PCA). These results are examined and possible explanations are explored. 1 INTRODUCTION Visual speech recognition is a challenging task in sensory integration. Psychophysical work by McGurk and MacDonald [5] first showed the powerful influence of visual information on speech perception that has led to increased interest in this 752 M. S. Gray, J. R. MoveLlan and T. J. Sejnowski area. A wide variety of techniques have been used to model speech-reading. Yuhas, Goldstein, Sejnowski, and Jenkins [8] used feedforward networks to combine gray scale images with acoustic representations of vowels. Wolff, Prasad, Stork, and Hennecke [7] explicitly computed information about the position of the lips, the shape of the mouth, and motion. This approach has the advantage of dramatically reducing the dimensionality of the input, but critical information may be lost. The visual information (mouth shape, position, and motion) was the input to a time-delay neural network (TDNN) that was trained to distinguish among consonant-vowel pairs. A separate TDNN was trained on the acoustic signal. The output probabilities for the visual and acoustic signals were then combined mUltiplicatively_ Bregler and Konig [1] also utilized a TDNN architecture. In this work, the visual information was captured by the first 10 principal components of a contour model fit to the lips. This was enough to specify the full range of lip shapes ("eigenlips"). Bregler and Konig [1] combined the acoustic and visual information in the input representation, which gave improved performance in noisy environments, compared with acoustic information alone. Surprisingly, the visual signal alone carries a substantial amount of information about spoken words. Garcia, Goldschen, and Petajan [2] used a variety of visual features from the mouth region of a speaker's face to recognize test sentences using hidden Markov models (HMMs). Those features that were found to give the best discrimination tended to be dynamic in nature, rather than static. Mase and Pentland [4] also explored the dynamic information present in lip images through the use of optical flow. They found that a template matching approach on the optical flow of 4 windows around the edges of the mouth yielded results similar to humans on a digit recognition task. Movellan [6] investigated the recognition of spoken digits using only visual information. The input representation for the hidden Markov model consisted of low-pass filtered pixel intensity information at each time step, as well as a delta image that showed the pixel by pixel difference between subsequent time steps. The motivation for the current work was succinctly stated by Bregler and Konig [1]: "The real information in lipreading lies in the temporal change of lip positions, rather than the absolute lip shape." Although different kinds of dynamic visual information have been explored, there has been no careful comparison of different methods. Here we present results for four different dynamic techniques that are based on general purpose processing at the pixel level. The first approach was to combine low-pass filtered gray scale pixel values with a delta image, defined as the difference between two successive gray level images. A peA reduction of this grayscale and delta information was investigated next. The final two approaches were motivated by the kinds of visual processing that are believed to occur in higher levels of the visual cortex. We first explored optical flow, which provides us with a representation analogous to that in primate visual area MT. Optical flow output was then combined with low-pass filtered gray-scale pixel values. Each of these four representations was tested on two different datasets: (1) the raw video images, and (2) the normalized video images. Dynamic Features/or Visual Speechreading: A Systematic Comparison \1\tt!1"f ,lIIIlltIrmltltNlII . , I \ f \ , //f\t\,,_ " '\ \ I I , 753 Figure 1: Image processing techniques. Left column: Two successive video frames (frames 1 and 2) from a subject saying the digit "one". These images have been made symmetric by averaging left and right pixels relative to the vertical midline. Middle column: The top panel shows gray scale pixel intensity information of frame 2 after low-pass filtering and down-sampling to a resolution of 15 x 20 pixels. The bottom panel shows the delta image (pixel-wise subtraction of frame 1 from frame 2), after low-pass filtering and downsampling. Right column: The top panel shows the optical flow for the 2 video frames in the left column. The bottom panel shows the reconstructed optical flow representation learned by a 1-state HMM. This can be considered the canonical or prototypical representation for the digit "one" across our database of 12 individuals. 2 METHODS AND MODELS 2.1 TRAINING SAMPLE The training sample was the Thlips1 database (Movellan [6]): 96 digitized movies of 12 undergraduate students (9 males, 3 females) from the Cognitive Science Department at U C-San Diego. Video capturing was performed in a windowless room at the Center for Research in Language at UC-San Diego. Subjects were asked to talk into a video camera and to say the first four digits in English twice. Subjects could monitor the digitized images in a small display conveniently located in front of them. They were asked to position themselves so that their lips were roughly centered in the feed-back display. Gray scale video images were digitized at 30 frames per second, 100 x 75 pixels, 8 bits per pixel. The video tracks were hand segmented by selecting a few relevant frames before the beginning and after the end of activity in the acoustic track. There were an average of 9.7 frames for each movie. Two sample frames are shown in the left column of Figure 1. 2.2 IMAGE PROCESSING We compared the performance of four different visual representations for the digit recognition task: low-pass + delta, PCA of (gray-scale + delta), flow, and low-pass
1996
81
1,282
The effect of correlated input data on the dynamics of learning S~ren Halkjrer and Ole Winther CONNECT, The Niels Bohr Institute Blegdamsvej 17 2100 Copenhagen, Denmark halkjaer>winther~connect.nbi.dk Abstract The convergence properties of the gradient descent algorithm in the case of the linear perceptron may be obtained from the response function. We derive a general expression for the response function and apply it to the case of data with simple input correlations. It is found that correlations severely may slow down learning. This explains the success of PCA as a method for reducing training time. Motivated by this finding we furthermore propose to transform the input data by removing the mean across input variables as well as examples to decrease correlations. Numerical findings for a medical classification problem are in fine agreement with the theoretical results. 1 INTRODUCTION Learning and generalization are important areas of research within the field of neural networks. Although good generalization is the ultimate goal in feed-forward networks (perceptrons), it is of practical importance to understand the mechanism which control the amount of time required for learning, i. e. the dynamics of learning. This is of course particularly important in the case of a large data set. An exact analysis of this mechanism is possible for the linear perceptron and as usual it is hoped that the results to some extend may be carried over to explain the behaviour of non-linear perceptrons. We consider N dimensional input vectors x E nN and scalar output y. The linear 170 S. Halkjrer and O. Winther perceptron is parametrized by the weight vector wE nN 1 y(x) = ..JNwT x (1) Let the training set be {( xt' , yt'), J.l = 1, . .. ,p} and the training error be the usual squared error, E(w) = ~ 2:t' (yt' - y(xt'))2. We will use the well-known gradient descent aigorithmi w(k + 1) = w(k) - rl'\1 E(w(k)) to estimate the minimum points w'" of E. Here 7] denotes the learning parameter. Collecting the input examples in the N x p matrix X and the corresponding output in y, the error function is written E(w) = ~(wTRw-2qTw+c), where R == }, 2:t' xt'(xt')T, q = -;};;XY and c = yT y. As in (Le Cun et at., 1991) the convergence properties of the minimum points w'" are examined in the coordinate system where R is diagonal. Let U denote the matrix whose columns are the eigenvectors of R and ~ = diag( AI, ... , AN) the diagonal matrix containing the eigenvalues of R. The new coordinates then become v = U T (w - w"') with corresponding error function2 1 1 E(v) = '2vT ~v + Eo = '2 ~Aivl + Eo (2) • where Eo = E(w"') . Gradient descent now leads to the decoupled equations vi(k + 1) = (1 - 7]Ai)Vi(k) = (1 - 7]Adkvi(O) (3) with i = 1, ... , N. Clearly, v -+ 0 requires 11 - 7]Ai I < 1 for all i, so that 7] must be chosen in the interval 0 < 7] < 2/ Amax. In the extreme case Ai = A we will have convergence in one step for 7] = 1/ A. However, in the usual case of unequal Ai the convergence for large k will be exponential vi(k) = exp(-7]Aik)Vi(O). (7]Ai)-1 therefore defines the time constant of the i'th equation giving a slowest time constant (7]Amin)-I. A popular choice for the learning parameter is 7] = 1/ Amax resulting in a slowest time constant Amax / Amin called the learning time T in the following. The convergence properties of the gradient descent algorithm is thus characterized by T. In the case of a singular matrix R, one or more of the eigenvalues will be zero, and there will be no convergence along the corresponding eigendirections. This has however no influence on the error according to (2). Thus, Amin will in the following denote the smallest non-zero eigenvalue. We will in the article calculate the eigenvalue spectrum of R in order to obtain the learning time of the gradient descent algorithm. This may be done by introducing the response function 1 1 ( 1 ) GL == G(L, H) = N TrL 1 _ RH == L 1 _ 'RH (4) where L, H are arbitrary N x N matrices. Using a standard representation of the Dirac o-function (Krogh, 1992) we may write the eigenvalue spectrum of R as (5) IThe Newton-Raphson method, w(k + 1) = w(k) - vr E(w(k))(vr2 E(w(k)))-l is of course much more effective in the linear case since it gives convergence in one step. This method however requires an inversion of the Hessian matrix. 2Note that this analysis is valid for any part of an error surface in which a quadratic approximation is valid. In the general case R should be exchanged with the Hessian vrvr E(w"'). The Effect of Correlated Input Data on the Dynamics of Learning 171 where). has an infinitesimal imaginary part which is set equal to zero at the end of the calculation. In the 'thermodynamic limit' N -+ 00 keeping a = Il constant and finite, G (and thus the eigenvalue spectrum) is a self-averaging quantity (Sollich, 1996) i. e. G - G = O(N- 1 ), where G is defined as the response function averaged over the input distribution. Previously G has been calculated for independent input variables (Hertz et al., 1989; Sollich, 1996). In section 2 we derive an implicit equation for the averaged response function for arbitrary correlations using random matrix techniques (Brody et al., 1981). This equation is solved showing that simple input correlations may slow down learning significantly. Based on this finding we propose in section 3 data transformations for improving the learning speed and test the transformation numerically on a medical classification problem in section 4. We conclude in section 5 with a discussion of the results. 2 THE RESPONSE FUNCTION The method for deriving the averaged response function is based on the fact that the response function (4) maybe written as a geometrical series GL = L~o (L(RHY). We will assume that the input examples x/J are drawn independently from a Gaussian distribution with means mi and correlations xixi mimi = Gii , i. e. x/J(xv)T = /5/JvZ and R = aZ where Z == C + mmT . The Gaussian distribution has the property that the average of products of x's can be calculated by making all possible pair correlations, e.g. XiXjXkXI = ZiiZkl + ZikZjl + ZilZjk. To take the average of (L(RHY), we must therefore make all possible pairs of the x's and exchange each pair xixi with Zij. This combinatorial problem will be solved below in a recursive fashion leading to an implicit equation for GL. Using underbraces to indicate pairings of x 's, we get for r 2: 2 (L(RH)r) ~ 2:(L~H(RH)r-1) /J + ~2 ~ ~ (LX" Sx")TH~RH)'x",<x"JTH(RH)"-'-2) r-2 a(LZH(RHY-1) + 2:-(L-(R-H-Y--3---1) (ZH(RH)3) (6) 3=0 Resumming this we get the response function 00 00 r-2 GL = (L) + a 2: (LZH(RHY) + 2: 2: ~(L~(R-H~)-r--3 --1) (ZH(RH)3) (7) r=O Exchanging the order of summation in the last term we can write everything in terms of the response function 00 00 GL (L) + aGLzH + 2: 2: (L(RHY-3) (ZH(RH)s) (L) + aGLzH + (GL - (L))GCH 172 S. Halkjrer and O. Winther (L) + aGLZH , 1- GZH (8) Using (8) recursively setting L equal to LZH, L(ZH)2 etc. one obtains _ 00 ( ( aZH )r) ( 1 ) GL = L L = L aZH 1 - GZH 1 ....:;:.,;;,=... r=O 1-GZH (9) This is our main result for the response function. To get the response function G 1. = G(..\ -1, ..\ -1) requires two steps, first set L = ZH and solve for GZH and then ... solve for G.!.. If Z has a particularly simple form (9) may be solved analytically, but in gene:al it must be solved numerically. In the following we will calculate the eigenvalue spectrum for a correlation matrix on the form C = nnT + rI and general mean m. To ensure that C is positive semi definite r 2: 0 and /n/2 + r 2: 0 where /n/2 == n . n. The eigenvalues of Z = nnT + mmT + rI are straight forwardly shown to be a1 = r (with multiplicity N - 2), a2 = r + [/n/2 + /m/2 - v'D] /2 and a3 = r + [/n/2 + /m/2 + v'D] /2 with D = (/nI2 - Im12)2 + 4(n . m)2 . Carrying out the trace in eq. (9) we get N-2 1 1 1 1 1 Gt = ----;;-..\ _ ---E4L- + N ..\ _ ~ + N ..\ _ ~ (10) 1-GZH l-GZH 1-GZH This expression suggests that we may solve G 1. in powers of 1/ N (see e.g. (Sollich, ... 1996)). However for purposes of the discussion of learning times the only ~-term that will be of importance is the last term above. We therefore only need to solve for GZH (setting L = ZH in (9)) to leading order G ..\ + a1(1- a) - J(..\ + a1(1- a)2) - 4..\a1 ZH 2,,\ (11) Note that GZH will vanish for large..\ implying that the last term in (10) to leading order is singular for ..\ = <('W3. Inserting the result in (10) gives G.!. = -+- [..\ + a2(1 - a) - J(..\ + a2(1- a))2 - 4..\a2] + ~..\ 1 (12) ... 2Aa2 N - aa1 According to (5) the eigenvalue spectrum is determined by the imaginary part and poles of G.!.. G.!. has an imaginary part for ..\_ < ..\ < ..\+ where ..\± = al (1 ± y'a)2 and the p01es l = 0, ..\ = aa3. The poles contribute each with a 6-function such that the eigenvalue spectrum up to corrections of order ~ becomes 1 1 p(..\) = (1 - a)6(1 - a)6(..\) + N 6(..\ - aa3) + 27r ..\al J(..\+ - ..\)(..\ - ..\_) (13) where 6(x) = 1 for x > 0 and 0 otherwise. The first term expresses the trivial fact that for p < N the whole input space is not spanned and R will have a fraction of 1 a zero-eigenvalues. The continuous spectrum (the root term) only contributes for ..\_ < ..\ < ..\+ . Numerical simulations has been performed to test the validity of the spectrum (13) (Halkjrer, 1996). They are in good agreement with predicted results indicating that finite size effects are unimportant. The continuous spectrum ' ... (13) has also been calculated using the replica method (Halkjrer, 1996). The Effect of Correlated Input Data on the Dynamics of Learning 173 From the spectrum the learning time T may be read of directly ( A+ aa3) ((1 + va) 2 aa3 ) T = max -, -= max , 2 A_ A_ 1 - va adl- va) (14) To illustrate how input correlations and bias may affect learning time consider simple correlations Cij = oijv(1 - c) + vc and mi = m. With this special choice aN(m~+vc) F 2 O· £ . . I . T = ( )( r:)2. or m + cv > , I.e. or non-zero mean or posItIve corre atlOns, v I-c I-va the convergence time will blow up by a factor proportional to N. The input bias effect has previously been observed by (Le Cun et al.,Wendemuth et al.). In the next section we will consider transformations to remove the large eigenvalue and thus to speed up learning. 3 DATA TRANSFORMATIONS FOR INCREASING LEARNING SPEED In this section we consider two data transformations for minimizing the learning time T of a data set, based on the results obtained in the previous sections. The PCA transformation (Jackson, 1991) is a data transformation often used in data analysis. Let U be the matrix whose columns are the eigenvectors of the sample covariance matrix and let x mean denote the sample average vector (see below). It is easy to check that the transformed data set (15) have uncorrelated (zero-mean) variables. However, the new PCA variables will often have a large spread in variances which might result in slow convergence. A simple rescaling of the new variables will remove this problem, such that according to (14) a PCA transformed data set with rescaled variables will have optimal convergence properties. The other transformation, which we will call double centering, is based on the removal of the observation means and the variable means. However, whereas the PCA transformation doesn't care about the initial distribution, this transformation is optimal for a data set generated from the matrix Zij = Oij v( l-c )+vc+mi mj studied earlier. Define xrean = } L:~ xf (mean of the i'th variable), x~ean = ~ L:i xf (mean of the p'th example) and x~:~~ = p;'" L:~i xf (grand mean). Consider first the transformed data set - ~ _ . _ mean _ ~ + mean Xi X, Xi Xmean Xmean The new variables are readily seen to have zero mean, variance ii = v(l-c)- ~(I-c) and correlation c = N-!I. Since v(1 - c) = v(1 - c) we immediately get from (13) that the continuous eigenvalue spectrum is unchanged by this transformation. Furthermore the 'large' eigenvalue aal is equal to zero and therefore uninteresting. Thus the learning time becomes T = (1 + va? /(1 - va)2. This transformation however removes perhaps important information from the data set, namely the observation means. Motivated by these findings, we create a new data set {x~} where this information is added as an extra component -~ _ (-~ -~ ~ mean) X X I , ... , X N, Xmean Xmean (16) 174 s. Halkjrer and O. Winther Table 1: Required number of iterations and corresponding learning times for different data transformations. 'Raw' is the original data set, 'Var. cent.' indicates the variable centered (mi = 0) data set, 'Doub. cent.' denotes (16), while the two last columns concerns the peA transformed data set (15) without and with rescaled variables. Raw Var. cent. Doub. cent. peA peA (res.) Iterations 00 300 50 630 7 T 161190 3330 237 3330 1 The matrix Ii resulting from this data set is identical to the above case except that a column and a row have been added. We therefore conclude that the eigenvalue spectrum of this data set consists of a continuous spectrum equal to the above and a single eigenvalue which is found to be A = ~ v(1 - c) + cv. For c 1= 0 we will therefore have a learning time T of order one indicating fast convergence. For independent variables (c = 0) the transformation results in a learning time of order N but in this case a simple removal of the variable means will be optimal. After training, when an (N + 1 )-dim parameter set w has been obtained, it is possible to transform back to the original data set using the parameter transformation WI 1 1 "",N WI + NWN+l N L...i=l Wi · 4 NUMERICAL INVESTIGATIONS The suggested transformations for improving the convergence properties have been tested on a medical classification problem. The data set consisted of 40 regional values of cerebral glucose metabolism from 85 patients, 48 HIV -negatives and 37 HIV-positives. A simple perceptron with sigmoidal tanh output was trained using gradient descent on the entropic error function to diagnose the 85 patients correctly. The choice of an entropic error function was due to it's superior convergence properties compared to the quadratic error function considered in the analysis. The learning was stopped once the perceptron was able to diagnose all patients correctly. Table 1 shows the average number of required iterations for each of the transformed data sets (see legend) as well as the ratio T = Amax/Amin for the corresponding matrix R. The 'raw' data set could not be learned within the allowed 1000 iterations which is indicated by an 00. Overall, there's fine agreement between the order of calculated learning times and the corresponding order of required number of iterations. Note especially the superiority of the peA transformation with rescaled variables. 5 CONCLUSION For linear networks the convergence properties of the gradient descent algorithm may be derived from the eigenvalue spectrum of the covariance matrix of the input data. The convergence time is controlled by the ratio between the largest and smallest (non-zero) eigenvalue. In this paper we have calculated the eigenvalue spectrum of a covariance matrix for correlated and biased inputs. It turns out that correlation and bias give rise to an eigenvalue of order the input dimension as well as a continuous spectrum of order one. This explains why a peA transformation (with The Effect of Correlated Input Data on the Dynamics of Learning 175 a variable rescaling) may increase learning speed significantly. We have proposed to center (setting equal to zero) the empirical mean both for each variable and each observation in order to remove the large eigenvalue. We add an additional component containing the observation mean to the input vector in order have this information in the training set. At the end of training it is possible to transform the solution back to the original representation. Numerical investigations are in fine agreement with the theoretical analysis for improving the convergence properties. 6 ACKNOWLEDGMENTS We would like to thank Sara A. Solla and Lars Kai Hansen for valuable comments and discussions. Furthermore we wish to thank Ido Kanter for providing us with notes on some of his previous work. This work has been supported by the Danish National Councils for the Natural and Technical Sciences through the Danish Computational Neural Network Center CONNECT. REFERENCES Brody, T. A., Flores J ., French J. B., Mello, P. A., Pendey, A., & Wong, S. S. (1981) Random-matrix physics. Rev. Mod. Phys. 53:385. Halkjrer, S. (1996) Dynamics of learning in neural networks: application to the diagnosis of HIV and Alzheimer patients. Master's thesis, University of Copenhagen. Hertz, J. A., Krogh, A. & Thorbergsson G. 1. (1989) Phase transitions in simple learning. J. Phys. A 22:2133-2150. Jackson, J. E. (1991) A User's Guide to Principal Components. John Wiley & Sons. Krogh, A. (1992) Learning with noise in a linear perceptron J. Phys A 25:1119-1133. Le Cun, Y., Kanter, 1. & Solla, S.A. (1991) Second Order Properties of Error Surfaces: Learning Time and Generalization. NIPS, 3:918-924. Sollich, P. (1996) Learning in large linear perceptrons and why the thermodynamic limit is relevant to the real world. NIPS, 7:207-214 Wendemuth, A., Opper, M. & Kinzel W. (1993) The effect of correlations in neural networks, J. Phys. A 26:3165.
1996
82
1,283
Learning temporally persistent hierarchical representations Suzanna Becker Department of Psychology McMaster University Hamilton, Onto L8S 4K1 becker@mcmaster.ca Abstract A biologically motivated model of cortical self-organization is proposed. Context is combined with bottom-up information via a maximum likelihood cost function. Clusters of one or more units are modulated by a common contextual gating Signal; they thereby organize themselves into mutually supportive predictors of abstract contextual features. The model was tested in its ability to discover viewpoint-invariant classes on a set of real image sequences of centered, gradually rotating faces. It performed considerably better than supervised back-propagation at generalizing to novel views from a small number of training examples. 1 THE ROLE OF CONTEXT The importance of context effectsl in perception has been demonstrated in many domains. For example, letters are recognized more quickly and accurately in the context of words (see e.g. McClelland & Rumelhart, 1981), words are recognized more efficiently when preceded by related words (see e.g. Neely, 1991), individual speech utterances are more intelligible in the context of continuous speech, etc. Further, there is mounting evidence that neuronal responses are modulated by context. For example, even at the level of the LGN in the thalamus, the primary source of visual input to the cortex, Murphy & Sillito (1987) have reported cells with "endstopped" or length-tuned receptive fields which depend on top-down inputs from the cortex. The end-stopped behavior disappears when the top-down connections are removed, suggesting that the cortico-thalamic connections are providing contextual modulation to the LGN. Moving a bit higher up the visual hierarchy, von der Heydt et al. (1984) found cells which respond to "illusory contours", in the absence of a contoured stimulus within the cells' classical receptive fields. These examples demonstrate that neuronal responses can be modulated by secondary sources of information in complex ways, provided the information is consistent with their expected or preferred input. 1 We use the term context rather loosely here to mean any secondary source of input. It could be from a different sensory modality, a different input channel within the same modality, a temporal history of the input, or top-down information. Learning Temporally Persistent Hierarchical Representations 825 Figure 1: Two sequences of 48 by 48 pixel images digitized with an IndyCam and preprocessed with a Sobel edge filter. Eleven views of each of four to ten faces were used in the simulations reported here. The alternate (odd) views of two of the faces are shown above. Why would contextual modulation be such a pervasive phenomenon? One obvious reason is that if context can influence processing, it can help in disambiguating or cleaning up a noisy stimulus. A less obvious reason may be that if context can influence learning, it may lead to more compact representations, and hence a more powerful processing system. To illustrate, consider the benefits of incorporating temporal history into an unsupervised classifier. Given a continuous sensory signal as input, the classifier must try to discover important partitions in its training data. If it can discover features that are temporally persistent, and thus insensitive to transformations in the input, it should be able to represent the signal compactly with a small set offeatures. FUrther, these features are more likely to be associated with the identity of objects rather than lower-level attributes. However, most classifiers group patterns together on the basis of spatial overlap. This may be reasonable if there is very little shift or other form of distortion between one time step and the next, but is not a reasonable assumption about the sensory input to the cortex. Pre-cortical stages of sensory processing, certainly in the visual system (and probably in other modalities), tend to remove low-order correlations in space and time, e.g. with centre-surround filters. Consider the image sequences of gradually rotating faces in Figure 1. They have been preprocessed by a simple edgefilter, so that successive views of the same face have relatively little pixel overlap. In contrast, identical views of different faces may have considerable overlap. Thus, a classifier such as k-means, which groups patterns based on their Euclidean distance, would not be expected to do well at classifying these patterns. So how are people (and in fact very young children) able to learn to classify a virtually infinite number of objects based on relatively brief exposures? It is argued here that the assumption of temporal persistence is a powerful constraining factor for achieving this, and is one which may be used to advantage in artificial neural networks as well. Not only does it lead to the development of higher-order feature analyzers, but it can result in more compact codes which are important for applications like image compression. Further, as the simulations reported here show, improved generalization may be achieved by allowing high-level expectations (e.g. of class labels) to influence the development of lower-level feature detectors. 2 THE MODEL Competitive learning (for a review, see Becker & Plumbley, 1996) is considered by many to be a reasonably strong candidate model of cortical learning. It can be implemented, in its simplest form, by a Hebbian learning rule in a network 826 S. Becker with lateral inhibition. However, a major limitation of competitive learning, and the majority of unsupervised learning procedures (but see the Discussion section), is that they treat the input as a set of independent identically distributed (iid) samples. They fail to take into account context. So they are unable to take advantage of the temporal continuity in signals. In contrast, real sensory signals may be better viewed as discretely sampled, continuously varying time-series rather than iid samples. The model described here extends maximum likelihood competitive learning (MLCL) (Nowlan, 1990) in two important ways: (i) modulation by context, and (ii) the incorporation of several "canonical features" of neocortical circuitry. The result is a powerful framework for modelling cortical self-organization. MLCL retains the benefits of competitive learning mentioned above. Additionally, it is more easily extensible because it maximizes a global cost function: L = t, log [t, ~iYi(a) 1 (1) where the 7r/s are positive weighting coefficients which sum to one, and the Yi'S are the clustering unit activations: y/a ) N(fla ), Wi, ~i) (2) where j(a) is the input vector for pattern a, and NO is the probability of j(a) under a Gaussian centred on the ith unit's weight vector, Wi, with covariance matrix 2:i . For simplicity, Nowlan used a single global variance parameter for all input dimensions, and allowed it to shrink during learning. MLCL actually maximizes the log likelihood (L) of the data under a mixture of Gaussians model, with mixing proportions equal to the 7r'S. L can be maximized by online gradient ascent2 with learning rate E: D..Wij = E ()L = E "'" 7ri Yi(a) (I/ a ) - Wij) (3) ()Wij ~ L:k 7rk Yk(a) Thus, we have a Hebbian update rule with normalization of post-synaptic unit activations (which could be accomplished by shunting inhibition) and weight decay. 2.1 Contextual modulation To integrate a contextual information source into MLCL, our first extension is to replace the mixing proportions (7r/s) by the outputs of contextual gating units (see Figure 2). Now the 7r/s are computed by separate processing units receiving their own separate stream of input, the "context". The role of the gating signals here is analagous to that of the gating network in the (supervised) "competing experts" model (Jacobs et al., 1991),3 For the network shown in Figure 2, the context is simply a time-delayed version of the outputs of a module (explained in the next subsection). However, more general forms of context are possible (see Discussion). In the simulations reported here, the context units computed their outputs according to a softmax function of their weighted summed inputs Xi: (a) _ ex;(a) 7r. - ---..,.--:Z L:j eXj(a) (4) We refer to the action of the gating units (the 7r/s) as modulatory because of the 2Nowlan (1990) used a slightly different online weight update rule that more closely approximates the batch update rule of the EM algorithm (Dempster et al., 1977) 3 However , in the competing experts architecture, both the experts and gating network receive a common source of input. The competing experts model could be thought of as fitting a mixture model of the training signal. Learning Temporally Persistent Hierarchical Representations 827 ~~~~~ .... ta (f) I ....... .nta Figure 2: The architecture used in the simulations reported here. Except where indicated, the gating units received all their inputs across unit delay lines with fixed weights of 1. o. multiplicative effect they have on the activities of the clustering units (the y/s). This multiplicative interaction is built into the cost function (Equation 1), and consequently, arises in the learning rule (Equation 3). Thus, clustering units are encouraged to discover features that agree with the current context signal they receive. If their context signal is weak or if they fail to capture enough of the activation relative to the other clustering units, they will do very little learning. Only if a unit's weight vector is sufficiently close to the current input vector and it's corresponding gating unit is strongly active will it do substantial learning. 2.2 Modular, hierarchical architecture Our second modification to MLCL is required to apply it to the architecture shown in Figure 2, which is motivated by several ubiquitous features of the neocortex: a laminar structure, and a functional organization into "cortical clusters" of spatially nearby columns with similar receptive field properties (see e.g. Calvin, 1995). The cortex, when flattened out, is like a large six-layered sheet. As Calvin (1995, pp. 269) succinctly puts it, " ... the bottom layers are like a subcortical 'out' box, the middle layer like an 'in' box, and the superficial layers somewhat like an 'interoffice' box connecting the columns and different cortical areas". The middle and superficial layer cells are analagous to the first-layer clustering units and gating units respectively. Thus, we propose that the superficial cells may be providing the contextual modulation. (The bottom layers are mainly involved in motor output and are not included in the present model.) To induce a functional modularity in our model analogous to cortical clusters, clustering units within the same module receive a shared gating signal. The cost function and learning rule are now: L n [m 1 I 1 ~ log ~ 1r~a) l ~Yi/a) (5) ~ 1r(a) Yi .(a) ( ) = E L..J 2: (~) i (a) Ik(a) -Wijk a q1rq rYqr (6) Thus, units in the same module form predictions y~j) of the same contextual feature 1r~a). Fortunately, there is a disincentive to all of them discovering identical weights: they would then do poorly at modelling the input. 3 EXPERIMENTS As a simple test of this model, it was first applied to a set of image sequences of four centered, gradually rotating faces (see Figure 1), divided into training and test 828 S. Becker Training Set Test Set no context, 4 faces: Layer 1 59.2 (2.4) 65 (3.5) context, 4 faces: Layer 1 88.4 (3.9) 74.5 (4.2) Layer 2 88.8 (4.0) 72.7 (4.8) context, 10 faces: Layer 1 96.3 (1.2) 71.0 (3.0) Layer 2 91.8 (2.4) 70.2 (4.3) Table 1 : Mean percent (and standard error) correctly classified faces , across 10 runs, for unsupervised clustering networks trained for 2000 iterations with a learning rate of 0.5, with and without temporal context. Layer 1: clustering units. Layer 2: gating units. Performance was assessed as follows: Each unit was assigned to predict the face class for which it most frequently won (was the most active). Then for each pattern, the layer's activity vector was counted as correct if the winner correctly predicted the face identity. sets by taking alternating views. It was predicted that the clustering units should discover "features" such as individual views of specific faces. Further, different views of the same face should be clustered together within a module because they will be observed in the same temporal context, while the gating units should discover the identity of faces, independent of viewpoint. First, the baseline effect of the temporal context on clustering performance was assessed by comparing the network shown in Figure 2 to the same network with the input connections to the gating layer removed. The latter is equivalent to MLCL with fixed, equal 7ri'S . The results are summarized in Table 1. As predicted, the temporal context provides incentive for the clustering units to group successive instances of the same face together, and the gating layer can therefore do very well at classifying the faces with a much smaller number of units - i.e., independently of viewpoint. In contrast, the clustering units without the contextual signal are more likely to group together similar views of different people's faces. Next, to explore the scaling properties of the model, a network like the one shown in Figure 2 but with 10 modules was presented with a set of 10 faces, 11 views each. As before, the odd-numbered views were trained on and the even-numbered views were tested on. To achieve comparable performance to the smaller network, the weights on the self-pointing connections on the gating units were increased from 1.0 to 3.0, which increased the time constant of temporal agveraging. The model then had no difficulty scaling up to the larger training set size, as shown in Table 1. Based on the unexpected success of this model, it's classification performance was then compared against supervised back-propagation networks on the four face sequences. The first supervised network we tried was a simple recurrent network with essentially the same architecture: one layer of Gaussian units followed by one layer of recurrent soft max units with fixed delay lines. Over ten runs of each model, although the unsupervised classifier did worse on the training set (it averaged 88% while the supervised model always scored 100% correct), it outperformed the supervised model in its generalization ability by a considerable margin (it averaged 73% while the supervised model averaged 45% correct). Finally, a feedforward back-propagation network with sigmoid units was trained. The following constraint on the hidden layer activations, hj(t): 4 hidden state cost = ,\ l:)hj(t) - hj(t - 1»2 j 4 As Geoff Hinton pointed out, the above constraint, if normalized by the variance, maximizes the mutual information between hidden unit states at adjacent time steps. Learning Temporally Persistent Hierarchical Representations 829 Training Set Performance Test Set Performance 1000 100.0 u U BOO Ql BOO Ql L.. L.. L.. L.. 0 0 U 60.0 U 60.0 C C Ql Ql 400 400 U a u a L.. L.. Ql -4 Ql a... 20.0 ····_·· 2 a... - - -. 1 - - - . 1 O. 1000.0 2000.0 JOOO.O 4000.0 1000.0 2000.0 JOOO.O 4000.0 Learning epoch Learning epoch Figure 3: Learning curves, averaged over five runs, for feedforward supervised net with a temporal smoothness constraint, for each of four levels of the parameter >.. was added to the cost function to encourage temporal smoothness. As the results in Figure 3 show, a feedforward network with no contextual input was thereby able to perform as well as our unsupervised model when it was constrained to develop hidden layer representations that clustered temporally adjacent patterns together. 4 DISCUSSION The unsupervised model's markedly better ability to generalize stems from it's cost function; it favors hidden layer features which contribute to temporally coherent predictions at the output (gating) layer. Multiple views of a given object are therefore more likely to be detected by a given clustering unit in the unsupervised model, leading to considerably improved interpolation of novel views. The poor generalization performance of back-propagation is not just due to overtraining, as the learning curves in Figure 3 show. Even with early stopping, the network with the lowest value of >. would not have done as well as the unsupervised network. There is simply no reason why supervised back-propagation should cluster temporally adjacent views together unless it is explicitly forced to do so. A "contextual input" stream was implemented in the simplest possible way in the simulations reported here, using fixed delay lines. However, the model we have proposed provides for a completely general way of incorporating arbitrary contextual information, and could equally well integrate other sources of input. The incoming weights to the gating units could also be learned. In fact, the gating unit activities actually represent the probabilities of each clustering unit's Gaussian model fitting the data, conditioned on the temporal history; hence, the entire model could be viewed as a Hidden Markov Model (Geoff Hinton, personal communication). However, current techniques for fitting HMMs are intractable if state dependencies span arbitrarily long time intervals. The model in its present implementation is not meant to be a realistic account of the way humans learn to recognize faces. Viewpoint-invariant recognition is achieved, if at all, in a hierarchical, multi-stage system. One could easily extend our model to achieve this, by connecting together a sequence of networks like the one shown in Figure 2, each having progressively larger receptive fields. A number of other unsupervised learning rules have been proposed based on the assumption oftemporally coherent inputs (FOldiak, 1991; Becker, 1993; Stone, 1996). Phillips et al. (1995) have proposed an alternative model of cortical self-organization they call coherent Infomax which incorporates contextual modulation. In their model, the outputs from one processing stream modulate the activity in another 830 s. Becker stream, while the mutual information between the two streams is maximized. A wide range of perceptual and cognitive abilities could be modelled by a network that can learn features of its primary input in particular contexts. These include multi-sensor fusion, feature segregation in object recognition using top-down cues, and semantic disambiguation in natural language understanding. Finally, it is widely believed that memories are stored rapidly in the hippocampus and related brain structures, and gradually incorporated into the slower-learning cortex for long-term storage. The model proposed here may be able to explain how such interactions between disparate information sources are learned. Acknowledgements This work evolved out of discussions with Ron Racine and Larry Roberts. Thanks to Geoff Hinton for contributing several valuable insights, as mentioned in the paper, and to Ken Seergobin for the face images. Software was developed using the Xerion neural network simulation package from Hinton's lab, with programming assistance from Lianxiang Wang. This work was supported by a McDonnell-Pew Cognitive Neuroscience research grant and a research grant from the Natural Sciences and Engineering Research Council of Canada. References Becker, S. (1993). Learning to categorize objects using temporal coherence. In S. J. Hanson, J. D. Cowan, & C. L. Giles (Eds.), Advances in Neural Information Processing Systems 5 (pp. 361-368). San Mateo, CA: Morgan Kaufmann. Becker, S. & Plumbley, M. (1996). Unsupervised neural network learning procedures for feature extraction and classification. International Journal of Applied Intelligence, 6(3). Calvin, W. H. (1995). Cortical columns, modules, and Hebbian cell assemblies. In M. Arbib (Ed.), The handbook of brain theory and neural networks. Cambridge, MA: MIT Press. Dempster, A. P., Laird, N. M., & Rubin, D. B. (1977). Maximum likelihood from incomplete data via the EM algorithm. Proceedings of the Royal Statistical Society, B-39:1-38. Foldiak, P. (1991). Learning invariance from transformation sequences. Neural Computation, 3(2):194-200. Jacobs, R. A., Jordan, M. I., Nowlan, S. J., & Hinton, G. E. (1991). Adaptive mixtures of local experts. Neural Computation, 3(1):79-87. McClelland, J. L. & Rumelhart, D. E. (1981). An interactive activation model of context effects in letter perception, part I: An account of basic findings. Psychological Review, 88:375-407. Murphy, C. & Sillito, A. M. (1987). Corticofugal feedback influences the generation of length tuning in the visual pathway. Nature, 329:727-729. Neely, J. (1991). Semantic priming effects in visual word recognition: A selective review of current findings and theories. In D. Besner & G. W. Humphreys (Eds.), Basic processes in reading: Visual Word Recognition (pp. 264-336). Hillsdale, NJ: Lawrence Erlbaum Associates. Nowlan, S. J. (1990). Maximum likelihood competitive learning. In D. S. Touretzky (Ed.), Neural Information Processing Systems, Vol. 2 (pp. 574-582). San Mateo, CA: Morgan Kaufmann. Phillips, W. A., Kay, J., & Smyth, D. (1995). The discovery of structure by multi-stream networks of local processors with contextual guidance. Network, 6:225-246. Stone, J. (1996). Learning perceptually salient visual parameters using spatiotemporal smoothness constraints. Neural Computation, 8:1463-1492. von der Heydt, R., Peterhans, E., & Baumgartner, G. (1984). Illusory contours and cortical neural responses. Science, 224 :1260-1262.
1996
83
1,284
Exploiting Model Uncertainty Estimates for Safe Dynamic Control Learning Jeff G. Schneider The Robotics Institute Carnegie Mellon University Pittsburgh, PA 15213 schneide@cs.cmu.edu Abstract Model learning combined with dynamic programming has been shown to be effective for learning control of continuous state dynamic systems. The simplest method assumes the learned model is correct and applies dynamic programming to it, but many approximators provide uncertainty estimates on the fit. How can they be exploited? This paper addresses the case where the system must be prevented from having catastrophic failures during learning. We propose a new algorithm adapted from the dual control literature and use Bayesian locally weighted regression models with dynamic programming. A common reinforcement learning assumption is that aggressive exploration should be encouraged. This paper addresses the converse case in which the system has to reign in exploration. The algorithm is illustrated on a 4 dimensional simulated control problem. 1 Introduction Reinforcement learning and related grid-based dynamic programming techniques are increasingly being applied to dynamic systems with continuous valued state spaces. Recent results on the convergence of dynamic programming methods when using various interpolation methods to represent the value (or cost-to-go) function have given a sound theoretical basis for applying reinforcement learning to continuous valued state spaces [Gordon, 1995]. These are important steps toward the eventual application of these methods to industrial learning and control problems. It has also been reported recently that there are significant benefits in data and computational efficiency when data from running a system is used to build a model, rather than using it once for single value function updates (as Q-learning would do) and discarding it [Sutton, 1990, Moore and Atkeson, 1993, Schaal and Atkeson, 1993, Davies, 1996]. Dynamic programming sweeps can then be done on the learned model either off-line or on-line. In its vanilla form, this method assumes the model is correct and does deterministic dynamic programming using the model. This assumption is often not correct, especially in the early stages of learning. When learning simulated or software systems, there may be no harm in the fact that this 1048 J. G. Schneider assumption does not hold. However, in real, physical systems there are often states that really are catastrophic and must be avoided even during learning. Worse yet, learning may have to occur during normal operation of the system in which case its performance during learning must not be significantly degraded. The literature on adaptive and optimal linear control theory has explored this problem considerably under the names stochastic control and dual control. Overviews can be found in [Kendrick, 1981, Bar-Shalom and Tse, 1976]. The control decision is based on three components call the deterministic, cautionary, and probing terms. The deterministic term assumes the model is perfect and attempts to control for the best performance. Clearly, this may lead to disaster if the model is inaccurate. Adding a cautionary term yields a controller that considers the uncertainty in the model and chooses a control for the best expected performance. Finally, if the system learns while it is operating, there may be some benefit to choosing controls that are suboptimal and/or risky in order to obtain better data for the model and ultimately achieve better long-term performance. The addition of the probing term does this and gives a controller that yields the best long-term performance. The advantage of dual control is that its strong mathematical foundation can provide the optimal learning controller under some assumptions about the system, the model, noise, and the performance criterion. Dynamic programming methods such as reinforcement learning have the advantage that they do not make strong assumptions about the system, or the form of the performance measure. It has been suggested [Atkeson, 1995, Atkeson, 1993] that techniques used in global linear control, including caution and probing, may also be applicable in the local case. In this paper we propose an algorithm that combines grid based dynamic programming with the cautionary concept from dual control via the use of a Bayesian locally weighted regression model. Our algorithm is designed with industrial control applications in mind. A typical scenario is that a production line is being operated conservatively. There is data available from its operation, but it only covers a small region of the state space and thus can not be used to produce an accurate model over the whole potential range of operation. Management is interested in improving the line's response to changes in set points or disturbances, but can not risk much loss of production during the learning process. The goal of our algorithm is to collect new data and optimize the process while explicitly minimizing the risk. 2 The Algorithm Consider a system whose dynamics are given by xk+1 = f(xk, uk). The state, x, and control,u, are real valued vectors and k represents discrete time increments. A model of f is denoted as j. The task is to minimize a cost functional of the form J = E:=D L(xk, uk, k) subject to the system dynamics. N mayor may not be fixed depending on the problem. L is given, but f must be learned. The goal is to acquire data to learn f in order to minimize J without incurring huge penalties in J during learning. There is an implicit assumption that the cost function defines catastrophic states. If it were known that there were no disasters to avoid, then simpler, more aggressive algorithms would likely outperform the one presented here. The top level algorithm is as follows: 1. Acquire some data while operating the system from an existing controller. 2. Construct a model from the data using Bayesian locally weighted regression. 3. Perform DP with the model to compute a value function and a policy. 4. Operate the system using the new policy and record additional data. Exploiting Model Uncertainty Estimates for Safe Dynamic Control Learning 1049 5. Repeat to step 2 while there is still some improvement in performance. In the rest of this section we describe steps 2 and 3. 2.1 Bayesian locally weighted regression We use a form of locally weighted regression [Cleveland and Delvin, 1988, Atkeson, 1989, Moore, 1992] called Bayesian locally weighted regression [Moore and Schneider, 1995] to build a model from data. When a query, x q , is made, each of the stored data points receives a weight Wi = exp( -llxi - xql1 2 / K). K is the kernel width which controls the amount of localness in the regression. For Bayesian LWR we assume a wide, weak normal-gamma prior on the coefficients of the regression model and the inverse of the noise covariance. The result of a prediction is a t distribution on the output that remains well defined even in the absence of data (see [Moore and Schneider, 1995] and [DeGroot, 1970] for details) . The distribution of the prediction in regions where there is little data is crucial to the performance of the DP algorithm. As is often the case with learning through search and experimentation, it is at least as important that a function approximator predicts its own ignorance in regions of no data as it is how well it interpolates in data rich regions. 2.2 Grid based dynamic programming In dynamic programming, the optimal value function, V, represents the cost-to-go from each state to the end of the task assuming that the optimal policy is followed from that point on. The value function can be computed iteratively by identifying the best action from each state and updating it according to the expected results of the action as given by a model of the system. The update equation is: Vk+1(x) = minL(x, u) + Vk(j(x, u» (1) In our algorithm, updates to the ~que function are computed while considering the probability distribution on the results of each action. If we assume that the output of the real system at each time step is an independent random variable whose probability density function is given by the uncertainty from the model, the update equation is as follows: Vk+1(x) = minL(x, u) + E[Vk(f(x, u))lj] (2) Note that the independence as~~fhption does not hold when reasonably smooth system dynamics are modeled by a smooth function approximator. The model error at one time step along a trajectory is highly correlated with the model error at the following step assuming a small distance traveled during the time step. Our algorithm for DP with model uncertainty on a grid is as follows: 1. Discretize the state space, X, and the control space, U. 2. For each state and each control cache the cost of taking this action from this state. Also compute the probability density function on the next state from the model and cache the information. There are two cases which are shown graphically in fig. 1: • If the distribution is much narrower than the grid spacing, then the model is confident and a deterministic update will be done according to eq. 1. Multilinear interpolation is used to compute the value function at the mean of the predicted next state [Davies, 1996] . • Otherwise, a stochastic update will be done according to eq. 2. The pdf of each of the state variables is stored, discretized at the same intervals as the grid representing the value function. Output independence is 1050 J G. Schneider High Confidence Next State Low Confidence Next State .......---:: ~ V ...---:l V. ~ V v7 v8 vlO vI,.!! (.-/. 17 __ V -..;:~ Figure 1: Illustration of the two kinds of cached updates. In the high confidence scenario the transition is treated as deterministic and the value function is computed with multilinear interpolation: Vl~+l = L(x, u) + OAV; + 0.3V; + 0.2V1k1 + 0.1 vl2• In the low confidence scenario the transition is treated stochastically and the update takes a weighted sum over all the vertices of significant weight as well as the ~ ,. Ie I b b·l· ·d h ·d TTk+l _ L( ) L-f.,/ lp(.,/l><} p(x lJ ,x,u)V (x) pro a Iltymassoutsl et egn: VIO X,u +~ I' L-{.,'Ip(,,' »<} p(x If ,x ,u) assumed and later the pdf of each grid point will be computed as the product of the pdfs for each dimension and a weighted sum of all the grid points with significant weight will be computed. Also the total probability mass outside the bounds of the grid is computed and stored. 3. For each state, use the cached information to estimate the cost of choosing each action from that state. Update the value function at that state according to the cost of the best action found. 4. Repeat 3 until the value function converges, or the desired number of steps has been reached in finite step problems. 5. Record the best action (policy) for each grid point. 3 Experiments: Minimal Time Cart-Pole Maneuvers The inverted pendulum is a well studied problem. It is easy to learn to stabilize it in a small number oftrials, but not easy to learn quick maneuvers. We demonstrate our algorithm on the harder problem of moving the cart-pole stably from one position to another as quickly as possible. We assume we have a controller that can balance the pole and would like to learn to move the cart quickly to new positions, but never drop the pole during the learning process. The simulation equations and parameters are from [Barto et aI., 1983] and the task is illustrated at the top of fig. 2. The state vector is x = [ pole angle (0), pole angular velocity (8), cart position (p), cart velocity (p) ]. The control vector, u, is the one dimensional force applied to the cart. Xo is [0 0 170] and the cost function is J = E~o xT X + 0.01 uT u. N is not fixed. It is determined by the amount of time it takes for the system to reach a goal region about the target state, [0 0 0 0]. If the pole is dropped, the trial ends and an additional penalty of 106 is incurred. This problem has properties similar to familiar process control problems such as cooking, mixing, or cooling, because it is trivial to stabilize the system and it can be moved slowly to a new desired position while maintaining the stability by slowly changing positional setpoints. In each case, the goal is to learn how to respond faster without causing any disasters during, or after, the learning process. Exploiting Model Uncertainty Estimates for Safe Dynamic Control Learning 1051 3.1 Learning an LQR controller We first learn a linear quadratic regulator that balances the pole. This can be done with minimal data. The system is operated from the state, [0 0 0 0] for 10 steps of length 0.1 seconds with a controller that chooses u randomly from a zero mean gaussian with standard deviation 0.5. This is repeated to obtain a total of 20 data points. That data is used to fit a global linear model mapping x onto x'. An LQR controller is constructed from the model and the given cost function following the derivation in [Dyer and McReynolds, 1970]. The resulting linear controller easily stabilizes the pole and can even bring the system stably (although very inefficiently as it passes through the goal several times before coming to rest there) to the origin when started as far out as x = [0 0 10 0]. If the cart is started further from the origin, the controller crashes the system. 3.2 Building the initial Bayesian LWR model We use the LQR controller to generate data for an initial model. The system is started at x = [0 0 1 0] and controlled by the LQR controller with gaussian noise added as before. The resulting 50 data points are stored for an LWR model that maps [e, 0, u] -+ [0, pl. The data in each dimension of the state and control space is scaled to [0 1]. In this scaled space, the LWR kernel width is set to 1.0. Next, we consider the deterministic DP method on this model. The grid covers the ranges: [±1.0 ±4.0 ±21.0 ±20.0] and is discretized to [11 9 11 9] levels. The control is ±30.0 discretized to 15 levels. Any state outside the grid bounds is considered failure and incurs the 106 penalty. If we assume the model is correct, we can use deterministic DP on the grid to generate a policy. The computation is done with fixed size steps in time of 0.25 seconds. We observe that this policy is able to move the system safely from an initial state of [0 0 12 0], but crashes if it is started further out. Failure occurs because the best path g.enerated using the model strays far from the region of the data (in variables e and e) used to construct the model. It is disappointing that the use of LWR for nonlinear modeling didn't improve much over a globally linear model and an LQR controller. We believe this is a common situation. It is difficult to build better controllers from naive use of nonlinear modeling techniques because the available data models only a narrow region of operation and safely acquiring a wider range of data is difficult. 3.3 Cautionary dynamic programming At this point we are ready to test our algorithm. Step 3 is executed using the LWR model from the data generated by the LQR controller as before. A trace of the system's operation when started at a distance of 17 from the goal is shown at the top of fig. 2. The controller is extremely conservative with respect to the angle of the pole. The pole is never allowed to go outside ±0.13 radians. Even as the cart approaches the goal at a moderate velocity the controller chooses to overshoot the goal considerably rather than making an abrupt action to brake the system. The data from this run is added to the model and the steps are repeated. Traces of the runs from three iterations of the algorithm are shown in fig. 2. At each trial, the controller becomes more aggressive and completes the task with less cost. After the third iteration, no significant improvement is observed. The costs are summarized and compared with the LQR and deterministic DP controllers in table 1. Fig. 3 is another illustration of how the policy becomes increasingly aggressive. It plots the pole angle vs. the pole angular velocity for the original LQR data and the executions at each of the following three trials. In summary, our algorithm is able 1052 13 24 1. G. Schneider Goal Reg-ion o 12 11 10 3 2. 25 2' 2'1. 'A i •• 10 • o I I I I I I I I I I I t I I I I , • 5 01 o I I I I I I I I I I ' , Figure 2: The task is to move the cart to the origin as quickly as possible without dropping the pole. The bottom three pictures show a trace of the policy execution obtained after one, two, and three trials (shown in increments of 0.5 seconds) Controller Number of data points used Cost from initial state 17 to build the controller LQR 20 failure Deterministic D P 50 failure Stochastic DP trial 1 50 12393 Stochastic DP trial 2 221 7114 Stochastic DP trial 3 272 6270 Table 1: Summary of experimental results to start from a simple controller that can stabilize the pole and learn to move it aggressively over a long distance without ever dropping the pole during learning. 4 Discussion We have presented an algorithm that uses Bayesian locally weighted regression models with dynamic programming on a grid. The result is a cautionary adaptive control algorithm with the flexibility of a non-parametric nonlinear model instead of the more restrictive parametric models usually considered in the dual control literature. We note that this algorithm presents a viewpoint on the exploration vs exploitation issue that is different from many reinforcement learning algorithms, which are devised to encourage exploration (as in the probing concept in dual control) . However, we argue that modeling the data first with a continuous function approximator and then doing DP on the model often leads to a situation where exploration must be inhibited to prevent disasters. This is particularly true in the case of real, physical systems. Exploiting Model Uncertainty Estimatesfor Safe Dynamic Control Learning 1.5 1 0.5 Angular 0 Velocity . . -0.5 -1 " " " " " " . " " " -1.5 -0.8 -0.6 -0.4 " -' .. --. . . " " " " " " . " ",,"""" " -0.2 0 Pole Angle ",," " '. LQR data 0 1st trial _ ....... 2nd trial '3td trial --., . " .. ,,"" " 0.2 --.... : 0.4 '. 1053 0.6 Figure 3: Execution trace. At each iteration, the controller is more aggressive. References [Atkeson, 1989) C. Atkeson. Using local models to control movement. In Advances in Neural Information Processing Systems, 1989. [Atkeson, 1993] C. Atkeson. Using local trajectory optimizers to speed up global optimization in dynamic programming. In Advances in Neural Information Processing Systems (NIPS-6), 1993. [Atkeson, 1995) C. Atkeson. Local methods for active learning. Invited talk at AAAI Fall Symposium on Active Learning, 1995. [Bar-Shalom and Tse, 1976) Y. Bar-Shalom and E. Tse. Concepts and Methods in Stochastic Control. Academic Press, 1976. [Barto et al., 1983) A. Barto, R. Sutton, and C. Anderson. Neuronlike adaptive elements that can solve difficult learning control problems. IEEE Transactions on Systems, Man, and Cybernetics, 1983. [Cleveland and Delvin, 1988) W . Cleveland and S. Delvin. Locally weighted regression: An approach to regression analysis by local fitting. Journal of the American Statistical Association, pages 596-610, September 1988. [Davies, 1996] S. Davies. Applying grid-based interpolation to reinforcement learning. In Neural Information Proceuing Systems 9, 1996. [DeGroot, 1970) M. DeGroot. Optimal Statistical Decisions. McGraw-Hill, 1970. [Dyer and McReynolds, 1970) P. Dyer and S. McReynolds. The Computation and Theory of Optimal Control. Academic Press, 1970. [Gordon, 1995] G. Gordon. Stable function approximation in dynamic programming. In The 12th International Conference on Machine Learning, 1995 . [Kendrick, 1981) D. Kendrick. Stochastic Control for Economic Models. McGraw-Hill, 1981. [Moore and AtkesoD, 1993) A. Moore and C. Atkeson. Prioritized sweeping: Reinforcement learning with less data and less real time. Machine Learning, 13(1):103-130,1993. [Moore and Schneider, 1995] A. Moore and J. Schneider. Memory based stochastic optimization. In Advances in Neural Information Proceuing Systems (NIPS-B), 1995. [Moore, 1992) A. Moore. Fast, robust adaptive control by learning only forward models. In Advances in Neural Information Processing Systems 4, 1992. [Schaal and Atkeson, 1993) S. Schaal and C . Atkeson. Assessing the quality of learned local models. In Advances in Neural Information Processing Systems (NIPS-6), 1993. [Sutton, 1990) R. Sutton. First results with dyna, an intergrated architecture for learning, planning, and reacting. In AAAI Spring Symposium on Planning in Uncertain, Unpredictable, or Changing Environment", 1990.
1996
84
1,285
Spectroscopic Detection of Cervical Pre-Cancer through Radial Basis Function Networks Kagan Tumer kagan@pine.ece.utexas.edu Dept. of Electrical and Computer Engr. Nirmala Ramanujam nimmi@ccwf.cc.utexas.edu Biomedical Engineering Program The University of Texas at Austin The University of Texas at Austin, Rebecca Richards-Kortum kortum@mail.utexas.edu Biomedical Engineering Program The University of Texas at Austin Joydeep Ghosh ghosh@ece.utexas.edu Dept. of Electrical and Computer Engr. The University of Texas at Austin Abstract The mortality related to cervical cancer can be substantially reduced through early detection and treatment. However, current detection techniques, such as Pap smear and colposcopy, fail to achieve a concurrently high sensitivity and specificity. In vivo fluorescence spectroscopy is a technique which quickly, noninvasively and quantitatively probes the biochemical and morphological changes that occur in pre-cancerous tissue. RBF ensemble algorithms based on such spectra provide automated, and near realtime implementation of pre-cancer detection in the hands of nonexperts. The results are more reliable, direct and accurate than those achieved by either human experts or multivariate statistical algorithms. 1 Introduction Cervical carcinoma is the second most common cancer in women worldwide, exceeded only by breast cancer (Ramanujam et al., 1996). The mortality related to cervical cancer can be reduced if this disease is detected at the pre-cancerous state, known as squamous intraepitheliallesion (SIL). Currently, a Pap smear is used to 982 K. Turner, N. Ramanujam, R. Richards-Kortum and J. Ghosh screen for cervical cancer {Kurman et al., 1994}. In a Pap test, a large number of cells obtained by scraping the cervical epithelium are smeared onto a slide which is then fixed and stained for cytologic examination. The Pap smear is unable to achieve a concurrently high sensitivityl and high specificity2 due to both sampling and reading errors (Fahey et al., 1995). Furthermore, reading Pap smears is extremely labor intensive and requires highly trained professionals. A patient with a Pap smear interpreted as indicating the presence of SIL is followed up by a diagnostic procedure called colposcopy. Since this procedure involves biopsy, which requires histologic evaluation, diagnosis is not immediate. In vivo fluorescence spectroscopy is a technique which has the capability to quickly, non-invasively and quantitatively probe the biochemical and morphological changes that occur as tissue becomes neoplastic. The measured spectral information can be correlated to tissue histo-pathology, the current "gold standard" to develop clinically effective screening and diagnostic algorithms. These mathematical algorithms can be implemented in software thereby, enabling automated, fast, non-invasive and accurate pre-cancer screening and diagnosis in hands of non-experts. A screening and diagnostic technique for human cervical pre-cancer based on laser induced fluorescence spectroscopy has been developed recently (Ramanujam et al., 1996). Screening and diagnosis was achieved using a multivariate statistical algorithm (MSA) based on principal component analysis and logistic discrimination of tissue spectra acquired in vivo. Furthermore, we designed Radial Basis Function (RBF) network ensembles to improve the accuracy of the multivariate statistical algorithm, and to simplify the decision making process. Section 2 presents the data collection/processing techniques. In Section 3, we discuss the MSA, and describe the neural network based methods. Section 4 contains the experimental results and compares the neural network results to both the results of the MSA and to current clinical detection methods. A discussion of the results is given in Section 5. 2 Data Collection and Processing A portable fluorimeter consisting of two nitrogen pumped-dye lasers, a fiber-optic probe and a polychromator coupled to an intensified diode array controlled by an optical multi-channel analyzer was utilized to measure fluorescence spectra from the cervix in vivo at three excitation wavelengths: 337, 380 and 460 nm (Ramanujam et al., 1996). Tissue biopsies were obtained only from abnormal sites identified by colposcopy and subsequently analyzed by the probe to comply with routine patient care procedure. Hemotoxylin and eosin stained sections of each biopsy specimen were evaluated by a panel of four board certified pathologists and a consensus diagnosis was established using the Bethesda classification system. Samples were classified as normal squamous (NS), normal columnar (NC), low grade (LG) SIL and high grade (HG) SIL. Table 1 provides the number of samples in the training (calibration) and test sets. Based on this data set, a clinically useful algorithm needs to discriminate SILs from the normal tissue types. Figure 1 illustrates average fluorescence spectra per site acquired from cervical sites at 337 nm excitation from a typical patient. Evaluation of the spectra at 337 nm exlSensitivity is the correct classification percentage on the pre-cancerous tissue samples. 2Specificity is the correct classification percentage on normal tissue samples. Spectroscopic Detection of Cervical Pre-cancer through RBF Networks Table 1: Histo-pathologic classification of samples. Histo-pathology Nonnal SIL 1faining Set 107 (SN: 94; SC: 13) 58 (LG: 23; HG: 35) Test Set 108 (SN: 94; SC: 14) 59 (LG: 24; HG: 35) 983 citation highlights one of the classification difficulties, namely that the fluorescence intensity of SILs (LG and HG) is less than that of the corresponding normal squamous tissue and greater than that of the corresponding normal columnar tissue over the entire emission spectrum3 . Fluorescence spectra at all three excitation wavelengths comprise of a total of 161 excitation-emission wavelengths pairs. However, there is a significant cost penalty for using all 161 values. To alleviate this concern, a more cost-effective fluorescence imaging system was developed, using component loadings calculated from principal component analysis. Thus, the number of required fluorescence excitation-emission wavelength pairs were reduced from 161 to 13 with a minimal drop in classification accuracy (Ramanujam et al., 1996). 0.50 ,----,-----,-----,----,-----.-----,-----.----, .~ 0.40 I/) c <I> C <I> g <I> o I/) <I> .... o :::l II 0.30 0.20 0.10 NC NS HG LG 0.00 <--__ --'-=.:=------'-____ ---l... __ --L-__ -'--__ ~ __ L...-_---' 300.0 400.0 500.0 Wavelength (nm) 600.0 700.0 Figure 1: Fluorecsence spectra from a typical patient at 337 nm excitation. 3 Algorithm Development 3.1 Multivariate Statistical Algorithms The multivariate statistical algorithm development described in (Ramanujam et al., 1996) consists of the following five steps: (1) pre-processing to reduce inter-patient and intra-patient variation of spectra from a tissue type, (2) dimension reduction of the pre-processed tissue spectra using Principal Component Analysis (PCA), (3) selection of diagnostically relevant principal components, (4) development of a classification algorithm based on logistic discrimination, and finally (5) retrospective and prospective evaluation of the algorithm's accuracy on a training (calibration) and test (prediction) set, respectively. Discrimination between SILs and the two normal tissue types could not be achieved effectively using MSA. Therefore two 3Spectral features observed in Figure 1 are representative of those measured at 380 nm and 460 nm excitation (not shown here). 984 K. Tumer, N. Ramanujam, R. Richards-Kortum and 1. Ghosh constituent algorithms were developed: algorithm (1), to discriminate between SILs and normal squamous tissues, and algorithm (2), to discriminate between SILs and normal columnar tissues (Ramanujam et al., 1996). 3.2 Algorithms based on Neural Networks The second stage of algorithm development consists of evaluating the applicability of neural networks to this problem. Initially, both Multi-Layered Perceptrons (MLPs) and Radial Basis function (RBF) networks were considered. However, MLPs failed to improve upon the MSA results for both algorithms (1) and (2), and frequently converged to spurious solutions. Therefore, our study focuses on RBF networks and RBF network ensembles. Radial Basis Function Networks: The first step in applying RBF networks to this problem consisted of retracing the two-step process outlined for the multivariate statistical algorithm. For constituent algorithm (1) the kernels were initialized using a k-means clustering algorithm on the training set containing NS tissue samples and SILs. The RBF networks had 10 kernels, whose locations and spreads were adjusted during training. For constituent algorithm (2), we selected 10 kernels, half of which were fixed to patterns from the columnar normal class, while the other half were initialized using a k-means algorithm. Neither the kernel locations nor their spreads were adjusted during training. This process was adopted to rectify the large discrepancy between the samples from each category (13 for columnar normal vs. 58 for SILs). For each algorithm, the training time was estimated by maximizing the performance on one validation set. Once the stopping time was established, 20 cases were run for each algorithm4. Linear and Order statistics Combiners: There were significant variations among different runs of the RBF networks for all three algorithms. Therefore, selecting the "best" classifier was not the ideal choice. First, the definition of "best" depends on the selection of the validation set, making it difficult to ascertain whether one network will outperform all others given a different test set, as the validation sets are small. Second, selecting only one classifier discards a large amount of potentially relevant information. In order to use all the available data, and to increase both the performance and the reliability of the methods, the outputs of RBF networks were pooled before a classification decision was made. The concept of combining classifier outputs5 has been explored in a multitude of articles (Hansen and Salamon, 1990; Wolpert, 1992). In this article we use the median combiner, which belongs to the class order statistics combiners introduced in (Turner and Ghosh, 1995), and the averaging combiner, which performs an arithmetic average of the corresponding outputs. 4 Results Two-step algorithm: The ensemble results reported are based on pooling 20 different runs of RBF networks, initialized and trained as described in the previous section. This procedure was repeated 10 times to ascertain the reliability of the 4Each run has a different initialization set of kernels/spreads/weights. 5 An extensive bibliography is available in (Turner and Ghosh, 1996). Spectroscopic Detection of Cervical Pre-cancer through RBF Networks 985 result and to obtain the standard deviations. For an application such as pre-cancer detection, the cost of a misclassification varies greatly from one class to another. Erroneously labeling a healthy tissue as pre-cancerous can be corrected when further tests are performed. Labeling a pre-cancerous tissue as healthy however, can lead to disastrous consequences. Therefore, for algorithm (1), we have increased the cost of a misclassified SIL until the sensitivity6 reached a satisfactory level. The sensitivity and specificity values for constituent algorithm (1) based on both MSA and RBF ensembles are provided in Table 2. Table 3 presents sensitivity and specificity values for constituent algorithm (2) obtained from MSA and RBF ensembles 7 . For both algorithms (1) and (2), the RBF based combiners provide higher specificity than the MSA. The median combiner provides results similar to those of the average combiner, except for algorithm (2) where it provides better specificity. In order to obtain the final discrimination between normal tissue and SILs, constituent algorithms (1) and (2) are used sequentially, and the results are reported in Table 4. Table 2: Accuracy of constituent algorithm (1) for differentiating SILs and normal squamous tissues, using MSA and RBF ensembles. Algorithm Specificity Sensitivity MSA 63% 90% RBF-ave 66% ±1% 90% ±O% RBF-med 66% ±1% 90% ±1% Table 3: Accuracy of constituent algorithm (2) for differentiating SILs and normal columnar tissues, using MSA and RBF ensembles. Algorithm Specificity Sensitivity MSA 36% 97% RBF-ave 37% ±5% 97% ±O% RBF-med 44% ±7% 97% ±O% One-step algorithm: The results presented above are based on the multi-step algorithm specifically developed for the MSA, which could not consolidate algorithms (1) and (2) into one step. Since the ultimate goal of these two algorithms is to separate SILs from normal tissue samples, a given pattern has to be processed through both algorithms. In order to simplify this decision process, we designed a one step RBF network to perform this separation. Because the pre-processing for algorithms (1) and (2) is different8 , the input space is now 26-dimensional. We initialized 10 kernels using a k-means algorithm on a trimmed9 version of the training set. The kernel locations and spreads were not adjusted during training. The cost of a misclassified SIL was set at 2.5 times the cost of a misclassified normal tissue SIn this case, the cost of misclassifying a SIL was three times the cost of misclassifying a normal tissue sample. 7In this case, there was no need to increase the cost of a misclassified SIL, because of the high prominence of SILs in the training set. 8Normalization vs. normalization followed by mean scaling. 9The trimmed set has the same number of patterns from each class. Thus, it forces each class to have a similar number of kernels. This set is used only for initializing the kernels. 986 K. Tumer; N. Ranulnujam. R. Richards-Kortum and 1. Ghosh sample, in order to provide the best sensitivity/specificity pair. The average and median combiner results are obtained by pooling 20 RBF networks10. Table 4: One step RBF algorithm compared to multi-step MSA and clinical methods for differentiating SILs and normal tissue samples. Algorithm Specificity Sensitivity 2-step MSA 63% 83% 2-step RB F -ave 65% ±2% 87% ±1% 2-step RBF-med 67% ±2% 87% ±1% RBF-ave 67% ±.75% 91% ±1.5% RBF-med 65.5% ±.5% 91% ±1% Pap smear (human expert) 68% ±21% 62% ±23% Colposcopy (human expert) 48%±23 % 94% ±6% The results of both the two-step and one-step RBF algorithms and the results of the two-step MSA are compared to the accuracy of Pap smear screening and colposcopy in expert hands in Table 4. A comparison of one-step RBF algorithms to the two-step RBF algorithms indicates that the one-step algorithms have similar specificities, but a moderate improvement in sensitivity relative to the two-step algOrithms. Compared to the MSA, the one-step RBF algorithms have a slightly decreased specificity, but a substantially improved sensitivity. In addition to the improved sensitivity, the one step RBF algorithms simplify the decision making process. A comparison between the one step RBF algorithms and Pap smear screening indicates that the RBF algorithms have a nearly 30% improvement in sensitivity with no compromise in specificity; when compared to colposcopy in expert hands, the RBF ensemble algorithms maintain the sensitivity of expert colposcopists, while improving the specificity by almost 20%. Figure 2 shows the trade-off between specificity and sensitivity for clinical methods, MSA and RBF ensembles, obtained by changing the misclassification cost. The RBF ensembles provide better sensitivity and higher reliability than any other method for a given specificity value. 1.0 .z. ·S 0.8 ~ '(;; c Q) (fJ 0.6 + + + A A + + A A + A ~RBF-ave G- ~ RBF-mad ~MSA + Pap smear A Colposcopy 0.4 ~~ __ ~ ____ ~ ____ ~ ____ ~ ______ ~ ____ ~ ____ ~ ____ -J 0.0 0.2 0.4 0.6 0.8 1 - Specificity Figure 2: Trade-off between sensitivity and specifity for MSA and RBF ensembles. For reference, Pap smear and colposcopy results from the literature are included (Fahey et al., 1995). lOThis procedure is repeated 10 times, in order to determine the standard deviation. Spectroscopic Detection of Cervical Pre-cancer through RBF Networks 987 5 Discussion The classification results of both the multivariate statistical algorithms and the radial basis function network ensembles demonstrate that significant improvement in classification accuracy can be achieved over current clinical detection modalities using cervical tissue spectral data obtained from in vivo fluorescence spectroscopy. The one-step RBF algorithm has the potential to significantly reduce the number of pre-cancerous cases missed by Pap smear screening and the number of normal tissues misdiagnosed by expert colposcopists. The qualitative nature of current clinical detection modalities leads to a significant variability in classification accuracy. For example, estimates of the sensitivity and specificity of Pap smear screening have been shown to range from 11-99% and 1497%, respectively (Fahey et al., 1995). This limitation can be addressed by the RBF network ensembles which demonstrate a significantly smaller variability in classification accuracy therefore enabling more reliable classification. In addition to demonstrating a superior sensitivity, the RBF ensembles simplify the decision making process of the two-step algorithms based on MSA into a single step that discriminates between SILs and normal tissues. We note that for the given data set, both MSA and MLP were unable to provide satisfactory solutions in one step. The one-step algorithm development process can be readily implemented in software, enabling automated detection of cervical pre-cancer. It provides near real time implementation of pre-cancer detection in the hands of non-experts, and can lead to wide-scale implementation of screening and diagnosis and more effective patient management in the prevention of cervical cancer. The success of this application will represent an important step forward in both medical laser spectroscopy and gynecologic oncology. Acknowledgements: This research was supported in part by NSF grant ECS 9307632, AFOSR contract F49620-93-1-0307, and Lifespex, Inc. References Fahey, M. T., Irwig, L., and Macaskill, P. (1995). Meta-analysis of pap test accuracy. American Journal of Epidemiology, 141(7):680-689. Hansen, L. K. and Salamon, P. (1990). Neural network ensembles. IEEE 1Tansactions on Pattern Analysis and Machine Intelligence, 12(10):993-1000. Kurman, R. J., Henson, D. E., Herbst, A. L., Noller, K. L., and Schiffman, M. H. (1994). Interim guidelines of management of abnormal cervical cytology. Journal of American Medical Association, 271:1866-1869. Ramanujam, N., Mitchell, M. F., Mahadevan, A., Thomsen, S., Malpica, A., Wright, T., Atkinson, N., and Richards-Kortum, R. R. (1996). Cervical pre-cancer detecion using a multivariate statistical algorithm based on fluorescence spectra at multiple excitation wavelengths. Photochemistry and Photobiology, 64(4):720-735. Turner, K. and Ghosh, .J. (1995). Order statistics combiners for neural classifiers. In Proceedings of the World Congress on Neural Networks, pages 1:31-34, Washington D.C. INNS Press. Turner, K. and Ghosh, J. (1996). Error correlation and error reduction in ensemble classifiers. Connection Science. (to appear). Wolpert, D. H. (1992). Stacked generalization. Neural Networks, 5:241-259.
1996
85
1,286
10........,....·..... Basis Function Networks and Complexity --_.IIiIIIIIIIIlo............... JIIIL....'IIIoo4II• .,JIIIL'IIU"JIIILJIIIL in Function Learning Adam Krzyzak Department of Computer Science Concordia University Montreal, Canada krzyzak@cs.concordia.ca Tamas Linder Dept. of Math. & Comp. Sci. Technical University of Budapest Budapest, Hungary linder@inf.bme.hu Abstract In this paper we apply the method of complexity regularization to derive estimation bounds for nonlinear function estimation using a single hidden layer radial basis function network. Our approach differs from the previous complexity regularization neural network function learning schemes in that we operate with random covering numbers and 11 metric entropy, making it po~sibleto consider much broader families of activation functions, namely functions of bounded variation. Some constraints previously imposed on the network parameters are also eliminated this way. The network is trained by means of complexity regularization involving empirical risk minimization. Bounds on the expected risk in tenns of the sample size are obtained for a large class of loss functions. Rates of convergence to the optimal loss are also derived. 1 INTRODUCTION Artificial neural networks have been found effective in learninginput-outputmappings from noisy examples. In this learning problem an unknown target function is to beinferredfrom a set of independent observations drawn according to some unknown probability distribution from the input-output space JRd x JR. Using this data set the learner tries to determine a function which fits the data in the sense of minimizing some given empirical loss function. The target function mayor may not be in the class of functions which are realizable by the learner. In the case when the class of realizable functions consists of some class of artificial neural networks, the above problem has been extensively studiedfrom different viewpoints. In recent years a special class of artificial neural networks, the radial basis function (RBF) networks have received considerable attention. RBF networks have been shown to be the solution of the regularization problem in function estimation with certain standard smoothness functionals used as stabilizers (see [5], and the references therein). Universal 198 A. Krzyiak and T. Linder convergence of RBF nets in function estimation and classification has been proven by Krzyzak et at. [6]. Convergence rates of RBF approximation schemes have been shown to be comparable with those for sigmoidal nets by Girosi and Anzellotti [4]. In a recent paper Niyogi and Girosi [9] studied the tradeoff between approximation and estimation errors and provided an extensive review of the problem. In this paper we consider one hidden layer RBF networks. We look at the problem of choosing the size of the hidden layer as a function of the available training data by means of complexity regularization. Complexity regularization approach has been applied to model selection by Barron [1], [2] resulting in near optimal choice of sigmoidal network parameters. Our approach here differs from Barron's in that we are using II metric entropy instead of the supremum norm. This allows us to consider amore general class of activation function, namely the functions of bounded variation, rather than a restricted class of activation functions satisfying a Lipschitz condition. For example, activations with jump discontinuities are allowed. In our complexity regularization approach we are able to choose the network parameters more freely, and no discretization of these parameters is required. For RBF regression estimation with squared error loss, we considerably improve the convergence rate result obtained by Niyogi and Girosi [9]. In Section 2 theproblem is formulated and two results on the estimation error of complexity regularized RBF nets are presented: one for general loss functions (Theorem 1) and a sharpened version of the first one for the squared loss (Theorem 2). Approximation bounds are combined with the obtained estimation results in Section 3 yielding convergence rates for function learning with RBF nets. 2 PROBLEM FORMULATION The task is to predict the value of a real random variable Y upon the observation of an lRd valued random vector X. The accuracy of the predictor f : lRd --+ R is measured by the expected risk J(/) = EL(f(X), Y), where L : lR x lR --+ lR+ is a nonnegative loss function. It will be assumed that there exists a minimizing predictor f* such that J(!*) == inf J(/). J A good predictor f n is to be detennined based on the data (Xl, Y]), · .. , (Xn , Yn ) which are i.i.d. copies of (X, Y). The goal is to make the expected risk EJ(fn) as small as possible, while fn is chosen from among a given class :F of candidate functions. In this paper the set of candidate functions :F will be the set of single-layer feedforward neural networks with radial basis function activation units and we let:F == Uk=l:Fk, where :Fk is the family of networks with k hidden nodes whose weight parameters satisfy certain constraints. In particular, for radial basis functions characterized by a kernel K : lR+ --+ JR., :Fk is the family of networks k I(x) = LWiK([x-Ci]tAi[X-Ci)) +wo, i=l where Wo, WI ..• , Wlc are real numbers called weights, CI, .•• ,Ck E R d , Ai are nonnegative' definite d x d matrices, and x t denotes the transpose of the column vector x. _ The complexity regularization principle for the learning problem was introduced by Vapnik [10] and fully developed by Barron [1], [2] (see also Lugosi and Zeger [8]). It enables the learning algorithm to choose the candidate class :Fk automatically, fro~ which is picks Radial Basis Function Networks and t;OJ'1Wi~exlltv N,t:Jnu'ln7'J"7n1~..n1lP'll 199 the estimate function by minimizing the empirical error over the training data. Complexity regularization penalizes the large candidate classes, which are bound to have small approximation error, in favor ofthe smaller ones, thus balancing the estimation and approximation errors. Let :F be a subset of a space X of real functions over some set, and let p be a pseudometric on X. For f > 0 the covering number N (f, :F, p) is defined to be the minimal number of closed f balls whose union cover :F. In other words, N (f, :F, p) is the least integer such that there exist 11, ... ,IN with N = N(f.,:F, p) satisfying sup ~n p(/, fi) S: f. IEF l~I~N In our case, :F is a family of real functions on lRm , and for any two functions I and g, p is given by ~hereZl, ... , Zn aren givenpointsinlRm • Inthiscasewewill usethe notation N(f,:F, p) == N (f, :F, zl)' emphasizing the dependence of the metric p on zl == (Zl' ... , zn). Let us define the families of functions 1-£k, k = 1, 2, ... by 1-lIc == {L(f(·), .) : I E :FIc}. Thus each member of1-£k maps lRd+ 1 intoR. It will be assumed that for each k we aregiven a finite, almost sure uniform upper bound on the random covering numbers N(f, 1-£k, Zr), where Zj == «Xl, Yl), ... , (Xn , Yn )). We may assume without loss of generality that N(E,1ik) is monotone decreasing in E. Finally, assume that L(f(X), Y) is uniformly almost surely bounded by a constant B, Le., P{L(/(X), Y) ~ B} == 1, f E :FIc, k == 1,2, ... (1) The complexity penalty ofthe kth class for n training samples is a nonnegative number ~kn satisfying Akn ~ lZ8B21og N(Akn/8,1lk) + Ck , n (2) (3) where the nonnegative constants Ck satisfy Er=l e- Ck ' :s 1. Note that since N(f, 1-£k) is nonincreasing in f, itis possibleto choose such ~kn for all k and n. Theresulting complexity penalty optimizes the upper bound on the estimation error in the proofofTheorem 1 below. We can now define our estimate. Let that is, fkn minimizes over rk the empirical risk for n training samples. The penalized empirical risk is defined for each f E rIc as The estimate f n is then defined as the fkn minimizing the penalized empirical risk over all classes: !~ == argminJn(!kn). lkn:k?;.l We have the following theorem for the expected estimation error of the above complexity regularization scheme. 200 A. Krzyiakand T. Linder' Theorem 1 For any nand k the complexity regularization estimate (3) satisfies EJ(/n) - J(/*) ~ min (Rkn + inf J(f) - J(f*)) , k~l . JE:Fk where (4) ~kn == Assuming without loss ofgenerality that log N (f., 1-lk) ~ 1, it is easy to see that the choice 128B21ogN(B/vnl1it.:)+(:;: n satisfies (2). 2.1 SQUARED ERROR LOSS For the 'special case when L(x, y) == (x _ y)2 we can obtain a better upper bound. The estimate will be the same as before, but instead of (2), the complexity penalty Akn now has to satisfy A 0 log N(Akn/C2, :Fk) + Ck (5) o.kn 2:: 1 , n where 01 == 349904, C2 == 25603,and 0 == max{B, I}. Here N (f. , :Fk) is a uniformupper bound on the random 11 covering numbers N(f., :Fk, Xl). Assume that the class:F == Uk:Fk is convex, and let :F be the closure of :F in L2(lJ), where Jl denotes the distribution of X. Then there is a unique1E :F whose squared loss J(1) achieves infJE:F J(I). We have the following bound on the difference EJ(fn) - J(I). Theorem 2 Assume that:F == Uk:Fk is·a convex set offunctions, and consider the squared error loss. Suppose that I/(x)1 ~ B for all x E ]Rd and f E :F, and P(IYI > B) == o. Then complexity regularization estimate with complexity penalty satisfying (5) gives EJ(fn) - J(l) .::; 2min (Akn + inf J(f) - J(I)) + 2C1 . k~l fE:Fk n The proofof this result uses an idea ofBarron [1] and a Bernstein-typeuniform probability inequality recently obtained by Lee et all [7]. 3 RBF NETWORKS We will consider radial basis function (RBF) networks with one hidden layer. Such a network is characterized by a kernel K : lR+ -+ JR. An RBF net of k nodes is of the fonn k f(x) =L Wi K ([x - Ci]tA[x - Cil) +wo, (6) i=l where wo, WI, .•. , Wk are real numbers called weights, CI, ... , Ck E ]Rd, and the Ai are nonnegative definite d x d matrices. The kth candidate class :FA: for the function estimation task is defined as the class of networks with k nodes which satisfy the weight condition I:~=o tWil ~ bfor a fixed b> 0: :FII: = {t. WiK([x - Ci]1A[x - cil) +wo: ~ IWil ::; b}. (7) Radial Basis Function Networks and LOllrlDlexztv KeRru[airization Let L(x, y) == Ix - yiP, and J(/) == EI/(X) - YIP, 201 (8) where 1 ~ p < 00. Let JJ denote the probability measure induced by X. Define:F to be the closui"e in V(JJ) ofthe convex hull of the functions bK([x - c]tA[x - c]) and the constant function h(x) == 1, x E lRd , where fbi ~ b, c E lRd, and A varies over all nonnegative d x d matrices. That is, :F is the closure of :F == Uk:Fk, where:Fk is given in (7). Let 9 E :F be arbitrary. If we assume that IK Iis uniformly bounded, then by Corollary 1of Darken et ala [3], we have for 1 :::; p ~ 2 that (9) where Ilf - 911£1'(1') denotes the LP(jl) norm (f If - r IPdjl) IIp, and.1"k is givenin (7). The approximation error infJErk J(/) - J(f*) can be dealt with using this result if the optimal 1* happens to be in :F. In this case, we obtain inf J(/) - J(/*) == O(ljVk) JErk for all 1 ~ p ~ 2. Values ofp close to 1 are of great importance for robust neural network regression. When the kernel K has a bounded total variation, it can be shown that N (t:, 1ik ) ~ (AI/t:)Azk, where the constants AI, A2 depend on ~upx IK(x)1, the total variation V of K, the dimension d, and on the the constant b in the definition (7) of :Fk. Then, if 1 ~ p ~ 2, the following consequence of Theorem 1 can be proved for LP regression estimation. Theorem 3 Let the kernel K be of bounded variation and assume that IYI is bounded. Thenfor 1 ~ p :::; 2 the error (8) ofthe complexity regularized estimate satisfies EJ(fn) - J(!*) ~ ~1 [0(Jkl~g~) +0(J*)] o ( Co~n)1/4) . For p = 1, i.e., for L 1 regression estimation, this rate is known to be optimal within the logarithmic factor. For squared error loss J(f) == E(f(X) - y)2 we have 1* (x) == E(YIX == x). Iff* E :F, then by (9) we obtain inf 1(/) - J(/*) == O(ljk). JErk, (10) It is easy to check that the class Uk:Fk is convex if the:Fk are the collections of RBF nets defined in (7). The next result shows that we can get rid of the square root in Theorem 3. Theorem 4 Assume that K is of bounded variation. Suppose furthermore that IY t is a bounded random variable, and let L(x, y) = (x - y)2. Then the complexity regularization RBF squared regression estimate satisfies EJ(fn) inf J(f) < 2 min ( inf J(/) inf J(/) + 0 (k 10gn)) + 0 (~) . . JEr k~l JErk JE:F n n 202 A. Krzyiakand T. Linder If f~- E :F, this result and (10) give EJ(fn) - J(I*) ~ ~? [0 (kl:gn ) + 0 (~ )] o ( Co~ n ) 1/2) . (11) This result sharpens and extends Theorem 3.1 of Niyogi and Girosi [9] where the weaker o (Jk l:gn) + 0 (t) convergence rate was obtained (in a PAC-like formulation) for the squared loss ofGaussianRBF network regression estimation. The ratein (11) varies linearly with dimension. Our result is valid for a very large class of RBF schemes, including the Gaussian RBF networks considered in [9]. Besides having improved on the convergence rate, our result has the advantage of allowing kernels which are not continuous, such as the window kernel. The above convergence rate results hold in the case when there exists an f* minimizing the risk which is a member of the LP(JJ) closure of :F = U:Fk, where :Fk is given in (7). In other words, f* should be such that for all ( > 0 there exists a k and a member f of Fk with Itf - f* IILP(tt) < f. The precise characterization of:F seems to be difficult. However, based on the work of Girosi and Anzellotti [4] we can describe a large class of functions that is contained in :F. Let H (x, t) be a real and bounded function of two variables x E lRd and t E lRn. Suppose that Ais a signed measure on lRn with finite total variation IfAll. If g(x) is defined as g(x) = ( H(x, t)>'(dt), JRD then 9 E LP(J.l) for any probability measure J.l on lRd. One can reasonably expect that g can be approximated well by functions f (x) of the fonn k f(x) = L wiH(x, ti), i=} where t}, ... ,tk E lRn and 2::=1 {Wil ~ ItAII. The case m = d and H(x, t) == G(x - t) is investigatedin [4], where a detailed description offunction spaces arising from the different choices of the basis function G is given. Niyogi and Girosi [9] extends this approach to approximation by convex combinations of translates and dilates of a Gaussian function. In general, we can prove the following Lemmal Let g(x) = ( H(x, t)>'(dt), (12) JRD where H (x, t) and A are as above. Define for each k ~ 1 the class offunctions gk = {f(X) =t wiH(x,ti):t. 1wd ~ II>'II}. Then for any probab'ility measure JJ on lRd andforany 1 ~ p < 00, thefunctiong can be approximated in £P(J.l) arbitrarily closely by members of9 == Ugk, i.e., inf Ilf - gIILP(tt) ~ 0 as k -+ 00. lEYk Radial Basis Function Networks and l:o~rrw~~exttvKeR14~larlW~~ton 203 To prove this lemma one need only slightly adapt the proof of Theorem 8.2 in [4], or in a more elementary way following the lines of the probabilistic proof of Theorem 1 of [6]. To apply the lemma for RBF networks considered in this paper, let n == d2 + d, t == (A, c), and H(x, t) == ]( ([x - c]tA[x - c]). Then we obtain that:F contains all the functions 9 with the integral representation g(x) = f K ([x - e]tA[x - en A(dedA), JRcil+d for which 11.A1-I ~ b, where bis the constraint on the weights as in (7). Acknowledgements This work was supported in partby NSERC grant OGPOO0270, CanadianNational Networks of Centers of Excelle~cegrant 293 and OTKA grant F014174. References [1] A. R. Barron. Complexity regularization with application to artificial neural networks. In G. Roussas, editor, Nonparametric Functional EstimationandRelatedTopics, pages 561-576. NATO ASI Series, Kluwer Academic Publishers, Dordrecht, 1991. [2] A. R. Barron. Approximation and estimation bounds for artificial neural networks. Machine Leaming, 14:115-133,1994. [3] C. Darken, M. Donahue, L. Gurvits, and E. Sontag. Rate of approximation results motivated by robust neural network learning. In Proc. Sixth Annual Workshop on ComputationalLeaming Theory,·pages 303-309. Morgan Kauffman, 1993. [4] F. Girosi and G. Anzellotti. Rates of convergence for radial basis functions and neural networks. In R. J. Mammone, editor, ArtijicialNeuralNetworksfor Speech and Vision, pages 97-113. Chapman & Hall, London, 1993. [5] F. Girosi, M. Jones, and T. Poggio. Regularization theory and neural network architectures. Neural Computation, 7:219-267,1995. [6] A.Krzyzak, T. Linder, and G. Lugosi. Nonparametric estimation and classification using radial basis function nets and empirical risk minimization. IEEE Transactions on Neural Networks, 7(2):475-487, March 1996.. [7] W. S. Lee, P. L. Bartlett, and R. C. Williamson. Efficient agnostic learning of neural networks with bounded fan-in. to be published in IEEE Transactions on Information Theory, 1995. [8] G. Lugosi and K. Zeger. Concept learning using complexity regularization. IEEE Transactions on Information Theory, 42:48-54, 1996. . [9] P. Niyogi and F. Girosi. On the relationship between generalization error, hypothesis complexity, and sample complexity for radial basis functions. Neural Computation, 8:819-842,1996. [10] V. N. Vapnik. Estimation ofDependencies Based on Empirical Data. Springer-Verlag, New York, 1982.
1996
86
1,287
Adaptively Growing Hierarchical Mixtures of Experts Jiirgen Fritsch, Michael Finke, Alex Waibel {fritsch+,finkem, waibel }@cs.cmu.edu Interactive Systems Laboratories Carnegie Mellon University Pittsburgh, PA 15213 Abstract We propose a novel approach to automatically growing and pruning Hierarchical Mixtures of Experts. The constructive algorithm proposed here enables large hierarchies consisting of several hundred experts to be trained effectively. We show that HME's trained by our automatic growing procedure yield better generalization performance than traditional static and balanced hierarchies. Evaluation of the algorithm is performed (1) on vowel classification and (2) within a hybrid version of the JANUS r9] speech recognition system using a subset of the Switchboard large-vocabulary speaker-independent continuous speech recognition database. INTRODUCTION The Hierarchical Mixtures of Experts (HME) architecture [2,3,4] has proven useful for classification and regression tasks in small to medium sized applications with convergence times several orders of magnitude lower than comparable neural networks such as the multi-layer perceptron. The HME is best understood as a probabilistic decision tree, making use of soft splits of the input feature space at the internal nodes, to divide a given task into smaller, overlapping tasks that are solved by expert networks at the terminals of the tree. Training of the hierarchy is based on a generative model using the Expectation Maximisation (EM) [1,3] algorithm as a powerful and efficient tool for estimating the network parameters. In [3], the architecture of the HME is considered pre-determined and remains fixed during training. This requires choice of structural parameters such as tree depth and branching factor in advance. As with other classification and regression techniques, it may be advantageous to have some sort of data-driven model-selection mechanism to (1) overcome false initialisations (2) speed-up training time and (3) adapt model size to task complexity for optimal generalization performance. In [11], a constructive algorithm for the HME is presented and evaluated on two small classification tasks: the two spirals and the 8-bit parity problems. However, this 460 1. Fritsch, M. Finke and A. Waibel algorithm requires the evaluation of the increase in the overall log-likelihood for all potential splits (all terminal nodes) in an existing tree for each generation. This method is computationally too expensive when applied to the large HME's necessary in tasks with several million training vectors, as in speech recognition, where we can not afford to train all potential splits to eventually determine the single best split and discard all others. We have developed an alternative approach to growing HME trees which allows the fast training of even large HME's, when combined with a path pruning technique. Our algorithm monitors the performance of the hierarchy in terms of scaled log-likelihoods, assigning penalties to the expert networks, to determine the expert that performs worst in its local partition. This expert will then be expanded into a new subtree consisting of a new gating network and several new expert networks. HIERARCHICAL MIXTURES OF EXPERTS We restrict the presentation of the HME to the case of classification, although it was originally introduced in the context of regression. The architecture is a tree with gating networks at the non-terminal nodes and expert networks at the leaves. The gating networks receive the input vectors and divide the input space into a nested set of regions, that correspond to the leaves of the tree. The expert networks also receive the input vectors and produce estimates of the a-posteriori class probabilities which are then blended by the gating network outputs. All networks in the tree are linear, with a softmax non-linearity as their activation function. Such networks are known in statistics as multinomiallogit models, a special case of Generalized Linear Models (GLIM) [5] in which the probabilistic component is the multinomial density. This allows for a probabilistic interpretation of the hierarchy in terms of a generative likelihood-based model. For each input vector x, the outputs of the gating networks are interpreted as the input-dependent multinomial probabilities for the decisions about which child nodes are responsible for the generation of the actual target vector y. After a sequence of these decisions, a particular expert network is chosen as the current classifier and computes multinomial probabilities for the output classes. The overall output of the hierarchy is N N P(ylx,0) = L9i(X, Vi) L9jli(X, Vij)P(ylx, Oij) i=1 j=1 where the 9i and 9jli are the outputs of the gating networks. The HME is trained using the EM algorithm [1] (see [3] for the application of EM to the HME architecture) . The E-step requires the computation of posterior node probabilities as expected values for the unknown decision indicators: hi = 9i I:j 9jli Pij(y) I:i 9i I:j 9jliPij(y) The M-step then leads to the following independent maximum-likelihood equations """" (t) "" (t) (t) Vij = arg ~ax L.,.. L.,.. hk L.,.. hqk log911k 'J t k I AdaptiveLy Growing HierarchicaL Mixtures of Experts 461 where the (Jij are the parameters of the expert networks and the Vi and Vij are the parameters of the gating networks. In the case of a multinomiallogit model, Pij{y) = Yc, where Yc is the output of the node associated with the correct class. The above maximum likelihood equations might be solved by gradient ascent, weighted least squares or Newton methods. In our implementation, we use a variant of Jordan & Jacobs' [3] least squares approach. GROWING MIXTURES In order to grow an HME, we have to define an evaluation criterion to score the experts performance on the training data. This in turn will allow us to select and split the worst expert into a new subtree, providing additional parameters which can help to overcome the errors made by this expert. Viewing the HME as a probabilistic model of the observed data, we partition the input dependent likelihood using expert selection probabilities provided by the gating networks 1(0; X) LlogP(y(t)lx(t) ,0) = LLgk log P (y(t)lx(t) ,0) k LL log[P(y(t)lx(t) , 0)]gk = Llk(0;X) k k where the gk are the products of the gating probabilities along the path from the root node to the k-th expert. gk is the probability that expert k is responsible for generating the observed data (note, that the gk sum up to one) . The expertdependent scaled likelihoods Ik (0; X) can be used as a measure for the performance of an expert within its region of responsibility. We use this measure as the basis of our tree growing algorithm: 1. Initialize and train a simple HME consisting of only one gate and several experts. 2. Compute the expert-dependent scaled likelihoods lk(9j X) for each expert in one additional pass through the training data. 3. Find the expert k with minimum lk and expand the tree, replacing the expert by a new gate with random weights and new experts that copy the weights from the old expert with additional small random perturbations. 4. Train the architecture to a local minimum of the classification error using a crossvalidation set. 5. Continue with step (2) until desired tree size is reached. The number of tree growing phases may either be pre-determined, or based on difference in the likelihoods before and after splitting a node. In contrast to the growing algorithm in [11], our algorithm does not hypothesize all possible node splits, but determines the expansion node(s) directly, which is much faster, especially when dealing with large hierarchies. Furthermore, we implemented a path pruning technique similar to the one proposed in [11], which speeds up training and testing times significantly. During the recursive depth-first traversal of the tree (needed for forward evaluation, posterior probability computation and accumulation of node statistics) a path is pruned temporarily if the current node's probability of activation falls below a certain threshold. Additionally, we also prune subtrees permanently, if the sum of a node's activation probabilities over the whole training set falls below a certain threshold. This technique is consistent with the growing algorithm and also helps preventing instabilities and singularities in the parameter updates, since nodes that accumulate too little training information will not be considered for a parameter update because such nodes are automatically pruned by the algorithm. 462 1. Fritsch, M. Finke and A. Waibel Figure 1: Histogram trees for a standard and a grown HME VOWEL CLASSIFICATION In initial experiments, we investigated the usefulness of the proposed tree growing algorithm on Peterson and Barney's [6] vowel classification data that uses formant frequencies as features. We chose this data set since it is small, non-artificial and low-dimensional, which allows for visualization and understanding of the way the growing HME tree performs classification tasks. x x ~ 0 o 0 The vowel data set contains 1520 samples consisting of the formants FO, F1, F2 and F3 and a class label, indicating one of 10 different vowels. Experiments were carried out on the 4-dimensional feature space, however, in this paper graphical representations are restricted to the F1-F2 plane. The figure to the left shows the data set represented in this plane (The formant frequencies are normalized to the range [0,1]). In the following experiments, we use binary branching HME's exclusively, but in general the growing algorithm poses no restrictions on the tree branching factor. We compare a standard, balanced HME of depth 3 with an HME that grows from a two expert tree to a tree with the same number of experts (eight) as the standard HME. The size of the standard HME was chosen based on a number of experiments with different sized HME's to find an optimal one. Fig. 1 shows the topology of the standard and the fully grown HME together with histograms of the gating probability distributions at the internal nodes. Fig. 2 shows results on 4-dimensional feature vectors in terms of correct classification rate and log-likelihood. The growing HME achieved a slightly better (1.6% absolute) classification rate than the fixed HME. Note also, that the growing HME outperforms the fixed HME even before it reaches its full size. The growing HME was expanded every 4 iterations, which explains the bumpiness of the curves. Fig. 3 shows the impact of path pruning during training on the final classification rate of the grown HME's. The pruning factor ranges from no pruning to full pruning (e.g. only the most likely path survives). Fig. 4 shows how the gating networks partition the feature space. It contains plots Adaptively Growing Hierarchical Mixtures of Experts Avorage Claaalficlllion Ra'o for dlfforenl '0" eo'" 100 r-----r-'-r---,--,-----,r----;--.--, , i eeLS% .-- -:c'= --- .--:==- =--SiC' 80 -..... t ·-r •. , .• -.. 1 --.... ·'1: _ _ ._ H~ .--.• -+~.=-t. 40 _ . -1 __ ----L--.--.: .. ~ -=--+= -1= ...... --.10 ----·t· -+.---. -.--+--- .---o o 4 8 12 16 20 24 28 epoch g ! } -500 ·1000 ·1500 ·2500 -3000 lOl/"lkolihood for o"'ndard and growing HME ! --+--~.- r-.,. 16 20 24 28 opoch 463 Figure 2: Classification rate and log-likelihood for standard and growing HME ~ CIa •• Hlcatlon rato when pruning gro,mg HME·. '*'ring training ~r--.--~~-r~~--r--~ 91 I "j' ""190 l .O __ ll.1--_ I l-~--. ....J.-2.2------ __ _ _ _ 89 88 87 88 85 84 83 -6 -5 -4 -3 -2 -1 o prunklg factor (10'1<) Figure 3: Impact of path pruning during training of growing HME's of the activation regions of all 8 experts of the standard HME in the 2-dimensional range [-0.1, 1.1]2. Activation probabilities (product of gating probabilities from root to expert) are colored in shades of gray from black to white. Fig. 5 shows the same kind of plot for all 8 experts of the grown HME. The plots in the upper right corner illustrate the class boundaries obtained by each HME. Figure 4: Expert activations for standard HME Fig. 4 reveals a weakness of standard HME's: Gating networks at high levels in the tree can pinch off whole branches, rendering all the experts in the subtree useless. In our case, half of the experts of the standard HME do not contribute to the final decision at all (black boxes). The growing HME's are able to overcome this effect. All the experts of the grown HME (Fig. 5) have non-zero activation patterns and the overlap between experts is much higher in the growing case, which indicates a higher degree of cooperation among experts. This can also be seen in the histogram trees in Fig. 3, where gating networks in lower levels of the grown tree tend to 464 1. Fritsch, M. Finke and A. Waibel Figure 5: Expert activations for grown HME average the experts outputs. The splits formed by the gating networks also have implications on the way class boundaries are formed by the HME. There are strong de~endencies visible between the class boundaries and some of the experts activation reglOns. EXPERIMENTS ON SWITCHBOARD We recently started experiments using standard and growing HME's as estimators of .posterior phone probabilities in a hybrid version of the JANUS [9] speech recognizer. Following the work in [12], we use different HME's for each state of a phonetic HMM. The posteriors for 52 phonemes computed by the HME's are converted into scaled likelihoods by dividing by prior probabilities to account for the likelihood based training and decoding of HMM's. During training, targets for the HME's are generated by forced-alignment using a baseline mixture of Gaussian HMM system. We evaluate the system on the Switchboard spontaneous telephone speech corpus. Our best current mixture of Gaussians based context-dependent HMM system achieves a word accuracy of 61.4% on this task, which is among the best current systems [7]. We started by using phonetic context-independent (CI) HME's for 3-state HMM's. We restricted the training set to all dialogues involving speakers from one dialect region (New York City), since the whole training set contains over 140 hours of speech. Our aim here was, to reduce training time (the subset contains only about 5% of the data) to be able to compare different HME architectures. Context I ff experts Word Acc. I 64 CI 33.8~ 64 35.1 0 Figure 6: Preliminary results on Switchboard telephone data To improve performance, we then build context-dependent (CD) models consisting of a separate HME for each biphone context and state. The Cb HME's output is smoothed with the CI models based on prior conteie probabilities. Current work focuses on improving context modeling (e.g. larger contexts and decision tree based clustering) . Fig. 6 summarizes the results so far, showing consistently that growing HME's outperform equally sized standard HME's. The results are not directly comparable Adaptively Growing Hierarchical Mixtures of Experts 465 with our best Gaussian mixture system, since we restricted context modeling to biphones and used only a small subset of the Switchboard database for training. CONCLUSIONS In this paper, we presented a method for adaptively growing Hierarchical Mixtures of Experts. We showed, that the algorithm allows the HME to use the resources (experts) more efficiently than a standard pre-determined HME architecture. The tree growing algorithm leads to better classification performance compared to standard HME's with equal numbers of parameters. Using growing instead of fixed HME's as continuous density estimators in a hybrid speech recognition system also improves performance. References [1] Dempster, A.P., Laird, N.M. & Rubin, D.B. (1977) Maximum likelihood from incomplete data via the EM algorithm. J.R. Statist. Soc. B 39, 1-38. [2] Jacobs, R. A., Jordan, M. 1., Nowlan, S. J., & Hinton, G. E. (1991) Adaptive mixtures of local experts. In Neural Computation 3, pp. 79-87, MIT press. [31 Jordan, M.l. & Jacobs RA. (1994) Hierarchical Mixtures of Experts and the EM Algorithm. In Neural Computation 6, pp. 181-214. MIT press. [41 Jordan, M.l. & Jacobs, RA. (1992) Hierarchies of adaptive experts. In Advances in Neural Information Processing Systems 4, J. Moody, S. Hanson, and R Lippmann, eds., pp. 985-993. Morgan Kaufmann, San Mateo, CA. [5] McCullagh, P. & Nelder, J.A. (1983) Generalized Linear Models. Chapman and Hall, London. [6] Peterson, G. E. & Barney, H. L. (1952) Control measurements used in a study of the vowels. J oumal of the Acoustical Society of America 24, 175-184. [7] Proceedings of LVCSR Hub 5 workshop, Apr. 29 - May 1 (1996) MITAGS, Linthicum Heights, Maryland. [8] Syrdal, A. K. & Gopal, H. S. (1986) A perceptual model of vowel recognition based on the auditory representation of American English vowels. Journal of the Acoustical Society of America, 79 (4):1086-1100. [9] Zeppenfeld T., Finke M., Ries K., Westphal M. & Waibel A. (1997) Recognition of Conversational Telephone Speech using the Janus Speech Engine. Proceedings of ICASSP 97, Muenchen, Germany [10] Waterhouse, S.R, Robinson, A.J. (1994) Classification using Hierarchical Mixtures of t;xperts. In Proc. 1994 IEEE Workshop on Neural Networks for Signal Processing IV, pp. 177-186. [11] Waterhouse, S.R, Robinson, A.J. (1995) Constructive Algorithms for Hierarchical Mixtures of Experts. In Advances in Neural information Processing Systems 8. [12] Zhao, Y., Schwartz, R, Sroka, J. & Makhoul, J. (1995) Hierarchical Mixtures of Experts Methodology Applied to Continuous Speech Recognition. In ICASSP 1995, volume 5, pp. 3443-6, May 1995.
1996
87
1,288
On-line Policy Improvement using Monte-Carlo Search Gerald Tesauro IBM T. J. Watson Research Center P. O. Box 704 Yorktown Heights, NY 10598 Abstract Gregory R. Galperin MIT AI Lab 545 Technology Square Cambridge, MA 02139 We present a Monte-Carlo simulation algorithm for real-time policy improvement of an adaptive controller. In the Monte-Carlo simulation, the long-term expected reward of each possible action is statistically measured, using the initial policy to make decisions in each step of the simulation. The action maximizing the measured expected reward is then taken, resulting in an improved policy. Our algorithm is easily parallelizable and has been implemented on the IBM SP! and SP2 parallel-RISC supercomputers. We have obtained promising initial results in applying this algorithm to the domain of backgammon. Results are reported for a wide variety of initial policies, ranging from a random policy to TD-Gammon, an extremely strong multi-layer neural network. In each case, the Monte-Carlo algorithm gives a substantial reduction, by as much as a factor of 5 or more, in the error rate of the base players. The algorithm is also potentially useful in many other adaptive control applications in which it is possible to simulate the environment. 1 INTRODUCTION Policy iteration, a widely used algorithm for solving problems in adaptive control, consists of repeatedly iterating the following policy improvement computation (Bertsekas, 1995): (1) First, a value function is computed that represents the longterm expected reward that would be obtained by following an initial policy. (This may be done in several ways, such as with the standard dynamic programming algorithm.) (2) An improved policy is then defined which is greedy with respect to that value function. Policy iteration is known to have rapid and robust convergence properties, and for Markov tasks with lookup-table state-space representations, it is guaranteed to convergence to the optimal policy. On-line Policy Improvement using Monte-Carlo Search 1069 In typical uses of policy iteration, the policy improvement step is an extensive off-line procedure. For example, in dynamic programming, one performs a sweep through all states in the state space. Reinforcement learning provides another approach to policy improvement; recently, several authors have investigated using RL in conjunction with nonlinear function approximators to represent the value functions and/or policies (Tesauro, 1992; Crites and Barto, 1996; Zhang and Dietterich, 1996). These studies are based on following actual state-space trajectories rather than sweeps through the full state space, but are still too slow to compute improved policies in real time. Such function approximators typically need extensive off-line training on many trajectories before they achieve acceptable performance levels. In contrast, we propose an on-line algorithm for computing an improved policy in real time. We use Monte-Carlo search to estimate Vp(z, a), the expected value of performing action a in state z and subsequently executing policy P in all successor states. Here, P is some given arbitrary policy, as defined by a "base controller" (we do not care how P is defined or was derived; we only need access to its policy decisions). In the Monte-Carlo search, many simulated trajectories starting from (z, a) are generated following P, and the expected long-term reward is estimated by averaging the results from each of the trajectories. (N ote that Monte-Carlo sampling is needed only for non-deterministic tasks, because in a deterministic task, only one trajectory starting from (z, a) would need to be examined.) Having estimated Vp(z, a), the improved policy pI at state z is defined to be the action which produced the best estimated value in the Monte-Carlo simulation, i.e., PI(z) = argmaxa Vp(z, a). 1.1 EFFICIENT IMPLEMENTATION The proposed Monte-Carlo algorithm could be very CPU-intensive, depending on the number of initial actions that need to be simulated, the number of time steps per trial needed to obtain a meaningful long-term reward, the amount of CPU per time step needed to make a decision with the base controller, and the total number of trials needed to make a Monte-Carlo decision. The last factor depends on both the variance in expected reward per trial, and on how close the values of competing candidate actions are. We propose two methods to address the potentially large CPU requirements of this approach. First, the power of parallelism can be exploited very effectively. The algorithm is easily parallelized with high efficiency: the individual Monte-Carlo trials can be performed independently, and the combining of results from different trials is a simple averaging operation. Hence there is relatively little communication between processors required in a parallel implementation. The second technique is to continually monitor the accumulated Monte-Carlo statistics during the simulation, and to prune away both candidate actions that are sufficiently unlikely (outside some user-specified confidence bound) to be selected as the best action, as well as candidates whose values are sufficiently close to the value of the current best estimate that they are considered equivalent (i.e., choosing either would not make a significant difference). This technique requires more communication in a parallel implementation, but offers potentially large savings in the number of trials needed to make a decision. 2 APPLICATION TO BACKGAMMON We have initially applied the Monte-Carlo algorithm to making move decisions in the game of backgammon. This is an absorbing Markov process with perfect state1070 G. Tesauro and G. R. Galperin space information, and one has a perfect model of the nondeterminism in the system, as well as the mapping from actions to resulting states. In backgammon parlance, the expected value of a position is known as the "equity" of the position, and estimating the equity by Monte-Carlo sampling is known as performing a "rollout." This involves playing the position out to completion many times with different random dice sequences, using a fixed policy P to make move decisions for both sides. The sequences are terminated at the end of the game (when one side has borne off all 15 checkers), and at that time a signed outcome value (called "points") is recorded. The outcome value is positive if one side wins and negative if the other side wins, and the magnitude of the value can be either 1, 2, or 3, depending on whether the win was normal, a gammon, or a backgammon. With normal human play, games typically last on the order of 50-60 time steps. Hence if one is using the Monte-Carlo player to play out actual games, the Monte-Carlo trials will on average start out somewhere in the middle of a game, and take about 25-30 time steps to reach completion. In backgammon there are on average about 20 legal moves to consider in a typical decision. The candidate plays frequently differ in expected value by on the order of .01. Thus in order to resolve the best play by Monte-Carlo sampling, one would need on the order of 10K or more trials per candidate, or a total of hundreds of thousands of Monte-Carlo trials to make one move decision. With extensive statistical pruning as discussed previously, this can be reduced to several tens of thousands of trials. Multiplying this by 25-30 decisions per trial with the base player, we find that about a million base-player decisions have to be made in order to make one MonteCarlo decision. With typical human tournament players taking about 10 seconds per move, we need to parallelize to the point that we can achieve at least lOOK base-player decisions per second. Our Monte-Carlo simulations were performed on the IBM SP! and SP2 parallelRISC supercomputers at IBM Watson and at Argonne National Laboratories. Each SP node is equivalent to a fast RSj6000, with floating-point capability on the order of 100 Mflops. Typical runs were on configurations of 16-32 SP nodes, with parallel speedup efficiencies on the order of 90%. We have used a variety of base players in our Monte-Carlo simulations, with widely varying playing abilities and CPU requirements. The weakest (and fastest) of these is a purely random player. We have also used a few single-layer networks (i.e., no hidden units) with simple encodings of the board state, that were trained by backpropagation on an expert data set (Tesauro, 1989). These simple networks also make fast move decisions, and are much stronger than a random player, but in human terms are only at a beginner-to-intermediate level. Finally, we used some multi-layer nets with a rich input representation, encoding both the raw board state and many hand-crafted features, trained on self-play using the TD(>.) algorithm (Sutton, 1988; Tesauro, 1992). Such networks play at an advanced level, but are too slow to make Monte-Carlo decisions in real time based on full rollouts to completion. Results for all these players are presented in the following two sections. 2.1 RESULTS FOR SINGLE-LAYER NETWORKS We measured the game-playing strength of three single-layer base players, and of the corresponding Monte-Carlo players, by playing several thousand games against a common benchmark opponent. The benchmark opponent was TD-Gammon 2.1 (Tesauro, 1995), playing on its most basic playing level (I-ply search, i.e., no lookahead). Table 1 shows the results. Lin-1 is a single-layer neural net with only the raw board description (number of White and Black checkers at each location) as On-line Policy Improvement using Monte-Carlo Search 1071 Network Base player Monte-Carlo player Monte-Carlo CPU Lin-1 -0.52 ppg -0.01 ppg 5 sec/move Lin-2 -0.65 ppg -0.02 ppg 5 sec/move Lin-3 -0.32 ppg +0.04 ppg 10 sec/move Table 1: Performance of three simple linear evaluators, for both initial base players and corresponding Monte-Carlo players. Performance is measured in terms of expected points per game (ppg) vs. TO-Gammon 2.11-ply. Positive numbers indicate that the player here is better than TO-Gammon. Base player stats are the results of 30K trials (std. dev. about .005), and Monte-Carlo stats are the results of 5K trials (std. dev. about .02). CPU times are for the Monte-Carlo player running on 32 SP 1 nodes. input. Lin-2 uses the same network structure and weights as Lin-l, plus a significant amount of random noise was added to the evaluation function, in order to deliberately weaken its playing ability. These networks were highly optimized for speed, and are capable of making a move decision in about 0.2 msec on a single SP1 node. Lin-3 uses the same raw board input as the other two players, plus it has a few additional hand-crafted features related to the probability of a checker being hit; there is no noise added. This network is a significantly stronger player, but is about twice as slow in making move decisions. We can see in Table 1 that the Monte-Carlo technique produces dramatic improvement in playing ability for these weak initial players. As base players, Lin-1 should be regarded as a bad intermediate player, while Lin-2 is substantially worse and is probably about equal to a human beginner. Both of these networks get trounced by TO-Gammon, which on its 1-ply level plays at strong advanced level. Yet the resulting Monte-Carlo players from these networks appear to play about equal to TO-Gammon l-ply. Lin-3 is a significantly stronger player, and the resulting MonteCarlo player appears to be clearly better than TO-Gammon l-ply. It is estimated to be about equivalent to TO-Gammon on its 2-ply level, which plays at a strong expert level. The Monte-Carlo benchmarks reported in Table 1 involved substantial amounts of CPU time. At 10 seconds per move decision, and 25 mOve decisions per game, playing 5000 games against TO-Gammon required about 350 hours of 32-node SP machine time. We have also developed an alternative testing procedure, which is much less expensive in CPU time, but still seems to give a reasonably accurate measure of performance strength. We measure the average equity loss of the MonteCarlo player on a suite of test positions. We have a collection of about 800 test positions, in which every legal play has been extensively rolled out by TO-Gammon 2.11-ply. We then use the TO-Gammon rollout data to grade the quality of a given player's move decisions. Test set results for the three linear evaluators, and for a random evaluator, are displayed in Table 2. It is interesting to note for comparison that the TO-Gammon l-ply base player scores 0.0120 on this test set measure, comparable to the Lin-1 Monte-Carlo player, while TO-Gammon 2-ply base player scores 0.00843, comparable to the Lin-3 Monte-Carlo player. These results are exactly in line with what we measured in Table 1 using full-game benchmarking, and thus indicate that the test-set methodology is in fact reasonably accurate. We also note that in each case, there is a huge error reduction of potentially a factor of 4 or more in using the Monte-Carlo technique. In fact, the rollouts summarized in Table 2 were done using fairly aggressive statistical pruning; we expect that rolling out decisions more 1072 G. Tesauro and G. R. Galperin Evaluator Base loss Monte-Carlo loss Ratio Random 0.330 0.131 2.5 Lin-1 0.040 0.0124 3.2 Lin-2 0.0665 0.0175 3.8 Lin-3 0.0291 0.00749 3.9 Table 2: Average equity loss per move decision on an 800-position test set, for both initial base players and corresponding Monte-Carlo players. Units are ppgj smaller loss values are better. Also computed is ratio of base player loss to Monte-Carlo loss. extensively would give error reduction ratios closer to factor of 5, albeit at a cost of increased CPU time. 2.2 RESULTS FOR MULTI-LAYER NETWORKS Using large multi-layer networks to do full rollouts is not feasible for real-time move decisions, since the large networks are at least a factor of 100 slower than the linear evaluators described previously. We have therefore investigated an alternative Monte-Carlo algorithm, using so-called "truncated rollouts.» In this technique trials are not played out to completion, but instead only a few steps in the simulation are taken, and the neural net's equity estimate of the final position reached is used instead of the actual outcome. The truncated rollout algorithm requires much less CPU time, due to two factors: First, there are potentially many fewer steps per trial. Second, there is much less variance per trial, since only a few random steps are taken and a real-valued estimate is recorded, rather than many random steps and an integer final outcome. These two factors combine to give at least an order of magnitude speed-up compared to full rollouts, while still giving a large error reduction relative to the base player. Table 3 shows truncated rollout results for two multi-layer networks: TD-Gammon 2.1 1-ply, which has 80 hidden units, and a substantially smaller network with the same input features but only 10 hidden units. The first line of data for each network reflects very extensive rollouts and shows quite large error reduction ratios, although the CPU times are somewhat slower than acceptable for real-time play. (Also we should be somewhat suspicious of the 80 hidden unit result, since this was the same network that generated the data being used to grade the Monte-Carlo players.) The second line of data shows what h~ppens when the rollout trials are cut off more aggressively. This yields significantly faster run-times, at the price of only slightly worse move decisions. The quality of play of the truncated rollout players shown in Table 3 is substantially better than TD-Gammon I-ply or 2-ply, and it is also substantially better than the full-rollout Monte-Carlo players described in the previous section. In fact, we estimate that the world's best human players would score in the range of 0.005 to 0.006 on this test set, so the truncated rollout players may actually be exhibiting superhuman playing ability, in reasonable amounts of SP machine time. 3 DISCUSSION On-line search may provide a useful methodology for overcoming some of the limitations of training nonlinear function approximators on difficult control tasks. The idea of using search to improve in real time the performance of a heuristic controller On-line Policy Improvement using Monte-Carlo Search 1073 Hidden Units Base loss Truncated Monte-Carlo loss Ratio M-C CPU 10 0.0152 0.00318 \ ll-step, thoroug~) 4.8 25 sec/move 0.00433 (ll-step, optimistic) 3.5 9 sec/move 80 0.0120 0.00181 \!-step, thoroug~) 6.6 65 sec(move 0.00269 (7-step, optimistic) 4.5 18 sec/move Table 3: Truncated rollout results for two multi-layer networks, with number of hidden units and rollout steps as indicated. Average equity loss per move decision on an 800-position test set, for both initial base players and corresponding Monte-Carlo players. Again, units are ppg, and smaller loss values are better. Also computed is ratio of base player loss to Monte-Carlo loss. CPU times are for the Monte-Carlo player running on 32 SP1 nodes. is an old one, going back at least to (Shannon, 1950). Full-width search algorithms have been extensively studied since the time of Shannon, and have produced tremendous success in computer games such as chess, checkers and Othello. Their main drawback is that the re~uired CPU time increases exponentially with the depth of the search, i.e., T '" B ,where B is the effective branching factor and D is the search depth. In contrast, Monte-Carlo search provides a tractable alternative for doing very deep searches, since the CPU time for a full Monte-Carlo decision only scales as T", N· B . D, where N is the number of trials in the simulation. In the backgammon application, for a wide range of initial policies, our on-line Monte-Carlo algorithm, which basically implements a single step of policy iteration, was found to give very substantial error reductions. Potentially 80% or more of the base player's equity loss can be eliminated, depending on how extensive the MonteCarlo trials are. The magnitude of the observed improvement is surprising to us: while it is known theoretically that each step of policy iteration produces a strict improvement, there are no guarantees on how much improvement one can expect. We have also noted a rough trend in the data: as one increases the strength of the base player, the ratio of error reduction due to the Monte-Carlo technique appears to increase. This could reflect superlinear convergence properties of policy iteration. In cases where the base player employs an evaluator that is able to estimate expected outcome, the truncated rollout algorithm appears to offer favorable tradeoffs relative to doing full rollouts to completion. While the quality of Monte-Carlo decisions is not as good using truncated rollouts (presumably because the neural net's estimates are biased), the degradation in quality is fairly small in at least some cases, and is compensated by a great reduction in CPU time. This allows more sophisticated (and thus slower) base players to be used, resulting in decisions which appear to be both better and faster. The Monte-Carlo backgammon program as implemented on the SP offers the potential to achieve real-time move decision performance that exceeds human capabilities. In future work, we plan to augment the program with a similar Monte-Carlo algorithm for making doubling decisions. It is quite possible that such a program would be by far the world's best backgammon player. Beyond the backgammon application, we conjecture that on-line Monte-Carlo search may prove to be useful in many other applications of reinforcement learning and adaptive control. The main requirement is that it should be possible to simulate the environment in which the controller operates. Since basically all of the recent successful applications of reinforcement learning have been based on training in simulators, this doesn't seem to be an undue burden. Thus, for example, Monte1074 G. Tesauro and G. R. Galperin Carlo search may well improve decision-making in the domains of elevator dispatch (Crites and Barto, 1996) and job-shop scheduling (Zhang and Dietterich, 1996). We are additionally investigating two techniques for training a controller based on the Monte-Carlo estimates. First, one could train each candidate position on its computed rollout equity, yielding a procedure similar in spirit to TD(1). We expect this to converge to the same policy as other TD(..\) approaches, perhaps more efficiently due to the decreased variance in the target values as well as the easily parallelizable nature of the algorithm. Alternately, the base position - the initial position from which the candidate moves are being made - could be trained with the best equity value from among all the candidates (corresponding to the move chosen by the rollout player). In contrast, TD(..\) effectively trains the base position with the equity of the move chosen by the base controller. Because the improved choice of move achieved by the rollout player yields an expectation closer to the true (optimal) value, we expect the learned policy to differ from, and possibly be closer to optimal than, the original policy. Acknowledgments We thank Argonne National Laboratories for providing SPI machine time used to perform some of the experiments reported here. Gregory Galperin acknowledges support under Navy-ONR grant N00014-96-1-0311. References D. P. Bertsekas, Dynamic Programming and Optimal Control. Athena Scientific, Belmont, MA (1995). R. H. Crites and A. G. Barto, "Improving elevator performance using reinforcement learning." In: D. Touretzky et al., eds., Advances in Neural Information Processing Systems 8, 1017-1023, MIT Press (1996). C. E. Shannon, "Programming a computer for playing chess." Philosophical Magazine 41, 265-275 (1950). R. S. Sutton, "Learning to predict by the methods of temporal differences." Machine Learning 3, 9-44 (1988). G. Tesauro, "Connectionist learning of expert preferences by comparison training." In: D. Touretzky, ed., Advances in Neural Information Processing Systems 1, 99106, Morgan Kaufmann (1989). G. Tesauro, "Practical issues in temporal difference learning." Machine Learning 8, 257-277 (1992). G. Tesauro, "Temporal difference learning and TD-Gammon." Comm. of the ACM, 38:3, 58-67 (1995). W. Zhang and T. G. Dietterich, "High-performance job-shop scheduling with a time-delay TD("\) network." In: D. Touretzky et al., eds., Advances in Neural Information Processing Systems 8, 1024-1030, MIT Press (1996).
1996
88
1,289
Blind separation of delayed and convolved sources. Te-Won Lee Max-Planck-Society, GERMANY, AND Interactive Systems Group Carnegie Mellon University Pittsburgh, PA 15213, USA tewonOes. emu. edu Anthony J. Bell Computational Neurobiology, The Salk Institute 10010 N. Torrey Pines Road La Jolla, California 92037, USA tonyOsalk.edu Russell H. Lambert Dept of Electrical Engineering University of South California, USA rlambertOsipi.use.edu Abstract We address the difficult problem of separating multiple speakers with multiple microphones in a real room. We combine the work of Torkkola and Amari, Cichocki and Yang, to give Natural Gradient information maximisation rules for recurrent (IIR) networks, blindly adjusting delays, separating and deconvolving mixed signals. While they work well on simulated data, these rules fail in real rooms which usually involve non-minimum phase transfer functions, not-invertible using stable IIR filters. An approach that sidesteps this problem is to perform infomax on a feedforward architecture in the frequency domain (Lambert 1996). We demonstrate real-room separation of two natural signals using this approach. 1 The problem. In the linear blind signal processing problem ([3, 2] and references therein), N signals, s(t) = [S1(t) ... SN(t)V, are transmitted through a medium so that an array of N sensors picks up a set of signals x(t) = [Xl (t) ... XN(t)V, each of which Blind Separation of Delayed and Convolved Sources 759 has been mixed, delayed and filtered as follows: N M-l Xi(t) = L L aijkSj(t - Dii - k) (1) j=l k=O (Here Dij are entries in a matrix of delays and there is an M-point filter, aij, between the the jth source and the ith sensor.) The problem is to invert this mixing without knowledge of it, thus recovering the original signals, s(t). 2 Architectures. The obvious architecture for inverting eq.1 is the feed forward one: N M-l Ui(t) = L L WijkXj(t - dij - k). (2) j=l k=O which has filters, Wij, and delays, dij, which supposedly reproduce, at the Ui , the original uncorrupted source Signals, Si. This was the architecture implicitly assumed in [2]. However, it cannot solve the delay-compensation problem, since in eq.1 each delay, Dij , delays a single source, while in eq.2 each delay, dii is associated with a mixture, Xj. Torkkola [8], has addressed the problem of solving the delay-compensation problem with a feedback architecture. Such an architecture can, in principle, solve this problem, as shown earlier by Platt & Faggin [7]. Torkkola [9] also generalised the feedback architecture to remove dependencies across time, to achieve the deconvolution of mixtures which have been filtered, as in eq.1. Here we propose a slightly different architecture than Torkkola's ([9], eq.15). His architecture could fail since it is missing feedback cross-weights for t = 0, ie: WijO. A full feedback system looks like: N M-l Ui(t) = Xi - L L WijkUj(t - dij - k). (3) j=l k=O and is illustrated in Fig.I. Because terms in Ui(t) appear on both sides, we rewrite this in vector terms: u(t) = x(t) - Wou(t) - E~~l Wku(t - k), in order to solve it as follows: M-l u(t) = (I + WO)-l(X(t) - L WkU(t - k» (4) k=l In these equations, there is a feedback unmixing matrix, W k, for each time point of the filter, but the 'leading matrix', Wo has a special status in solving for u(t). The delay terms are useful since one metre of distance in air at an 8kHz sampling rate, corresponds to a whole 25 zero-taps of a filter. Reintroducing them gives us: N M-l u(t) = (I + WO)-l(X(t) - net(t», neti(t) = L L WijkU(t - dij - k» (5) j=l k=l 760 T. Lee, A. 1. Bell and R. H. Lambert Ma:tdmize Joint Entropy H(Y) / S I ---.----; I---.--XI UI S2--<----; U2 A(z) W(z) Figure 1: The feedback neural architecture of eq.9, which is used to separate and deconvolve signals. Each box represents a causal filter and each circle denotes a time delay. 3 Algorithms. Learning in this architecture is performed by maximising the joint entropy, H(y(t)), of the random vector y(t) = g(u(t)), where 9 is a bounded monotonic nonlinear function (a sigmoid function). The success of this for separating sources depends on four assumptions: (1) that the sources are statistically independent, (2) that each source is white, ie: there are no dependencies between time points, (3) that the non-linearity, g, has a derivative which has higher kurtosis than the probability density functions (pdf's) of the sources, and (4) that a stable IIR (feedback) inverse of the mixing exists; ie: that a is minimum phase (see section 5). Assumption (1) is reasonable and Assumption (3) allows some tailoring of our algorithm to fit data of different types. Assumption (2), on the other hand, is not true for natural signals. Our algorithm will whiten: it will remove dependencies across time which already existed in the original source signals, Si. However, it is possible to restore the characteristic autocorrelations (amplitude spectra) of the sources by post-processing. For the reasoning behind Assumption (3) see [2]. We will discuss Assumption 4 in section 5. In the static feedback case of eq.5, when M = 1, the learning rule for the feedback weights W 0 is just a co-ordinate transform of the rule for feedforward weights, W 0, in the equivalent architecture of u(t) = Wox(t). Since Wo = (I + WO)-l, we have Wo = WOI - I, which, due to the quotient rule for matrix differentiation, differentiates as: (6) The best way to maximise entropy in the feedforward system is not to follow the entropy gradient, as in [2J, but to follow its 'natural' gradient, as reported by Amari et al [1]: (7) This is an optimal rescaling of the entropy gradient [1, 3]. It simplifies the learning Blind Separation of Delayed and Convolved Sources 761 rule and speeds convergence considerably. Evaluated, it gives [2]: A T A AWO ex: (I + yu )WO, (8) Substituting into eq.7 gives the natural gradient rule for static feedback weights: (9) This reasoning may be extended to networks involving filters. For the feedforward filter architecture u(t) = ~:!:~1 W kX(t - k), we derive a natural gradient rule (for k> 0) of: A T A A W k ex: YUt-k W k (10) where, for convenience, time has become subscripted. Performing the same coordinate transforms as for Wo above, gives the rule: (11) (We note that learning rules similar to these have been independently derived by Cichocki et al [4]). Finally, for the delays in eq.5, we derive [2, 8]: aH( ) M-l a Ad·· ex: Y = _yA. ~ -W" kU(t - d .. - k) 13 ad. . 1 L at 13 13 13 k=l (12) This rule is different from that in [8] because it uses the collected temporal gradient information from all the taps. The algorithms of eq.9, eq.11 and eq.12 are the ones we use in our experiments on the architecture of eq.5. 4 Simulation results for the feedback architecture To test the learning rules in eq.9, eq.ll and eq.12 we used an IIR filter system to recover two sources which had been mixed and delayed as follows (in Z-transform notation): An (z) = 0.9 + 0.5z-1 + 0.3z-2 A21 (z) = -0.7z-5 - 0.3z-6 - 0.2z-7 A12 (Z) = 0.5z- 5 + 0.3z-6 + 0.2z-7 A22(Z) = 0.8 - 0.lz-1 (13) The mixing system, A(z), is a minimum-phase system with all its zeros inside the unit circle. Hence, A(z) can be inverted using a stable causal IIR system since all poles of the inverting systems are also inside the unit circle. For this experiment, we chose an artificially-generated source: a white process with a Laplacian distribution [/z(x) = exp( -Ix!)]. In the frequency domain the deconvolving system looks as follows: (14) where D(z) = Wll (Z)W22(Z) - W12(Z)W21 (Z)). This leads to the following solution for the weight filters: Wll (z) = A 22 (Z) W22 (Z) = All (z) W21 (z) = -A21 (z) W12 (Z) = -A12(Z) (15) 762 T. Lee, A. 1. Bell and R. H. Lambert The learning rule we used was that of eq.9 and eq.ll with the logistic non-linearity, Yi = 1/ exp( -Ui). Fig.2A shows the four filters learnt by our IIR algorithm. The bottom row shows the inverting system convolved with the mixing system, proving that W * A is approximately the identity mapping. Delay learning is not demonstrated here, though for periodic signals like speech we observed that it is subject to local minima problems [8, 9]. (A) Feedback (IIR) learning (8) Feedforward (FIR) learning Figure 2: Top two rows: learned unmixing filters for (A) IIR learning on minimumphase mixing, and (B) FIR freq.-domain learning on non-minimum phase mixing. Bottom row: the convolved mixing and unmixing systems. The delta-like response indicates successful blind unmixing. In (B) this occurs acausally with a time-shift. 5 Back to the feedforward architecture. The feedback architecture is elegant but limited. It can only invert minimumphase mixing (all zeros are inside the unit circle meaning that all poles of the inverting system are as well). Unfortunately, real room acoustics usually involves non-minimum phase mixing. There does exist, however, a stable non-causal feedforward (FIR) inverse for nonminimum phase mixing systems. The learning rules for such a system can be formulated using the FIR polynomial matrix algebra as described by Lambert [5]. This may be performed in the time or frequency domain, the only requirements being that the inverting filters are long enough and their main energy occurs more-orless in their centre. This allows for the non-causal expansion of the non-minimum phase roots, causing the roughly symmetrical "flanged" appearance of the filters in Fig-.2B. For convenience, we formulate the infomax and natural gradient info max rules [2, 1] in the frequency domain: ~ W ex W-H + fft(y)XH ~ W ex (I + fft(y)UH)W (16) (17) where the H superscript denotes the Hermitian transpose (complex conjugate). In these rules, as in eq.14, W is a matrix of filters and U and X are blocks of multiBlind Separation of Delayed and Convolved Sources 763 sensor signal in the frequency domain. Note that the nonlinearity 1li = 8~i ~ still operates in the time domain and the fft is applied at the output. 6 Simulation results for the feedforward architecture To show the learning rule in eq.17 working, we altered the transfer function in eq.13 as follows: (18) This system is now non-minimum phase, having zeros outside the unit circle. The inverse system can be approximated by stable non-causal FIR filters. These were learnt using the learning rule of eq.17 (again, with the logistic non-linearity). The resulting learnt filters are shown in Fig.2B where the leading weights were chosen to be at half the filter size (M/2). Non-causality of the filters can be clearly observed for W12 and W21, where there are non-zero coefficients before the maximum amplitude weights. The bottom row of Fig.2B shows the successful separation by plotting the complete unmixing/mixing transfer function: W * A. 7 Experiments with real recordings To demonstrate separation in a real room, we set up two microphones and recorded firstly two people speaking and then one person speaking with music in the background. The microphones and the sources were both 60cm apart and 60cm from each other (arranged in a square), and the sampling was 16kHz. Fig.3A shows the two recordings of a person saying the digits "one" to "ten" while loud music plays in the background. The IIR system of eq.5, eq.9 and eq.ll was unable to separate these signals, presumably due to the non-mini mum-phase nature of the room transfer functions. However, the algorithm of eq.17, converged after 30 passes through the 10 second recordings. The filter lengths were 256 (corresponding to 16ms). The separated signals are shown in Fig.3B. Listening to them conveys a sense of almost-clean separation, though interference is audible. The results on the two people speaking were similar. An important application is in spontaneous speech recognition tasks where the best recognizer may fail completely in the presence of background music or competing speakers (as in the teleconferencing problem). To test this application, we fed into a speech recognizer, ten sentences recorded with loud music in the background and ten sentences recorded with a simultaneous speaker interference. After separation, the recognition rate increased considerably for both cases. These results are reported in detail in [6]. 8 Conclusions Starting with 'Natural gradient infomax' IIR learning rules for blind time delay adjustment, separation and deconvolution, we showed how these worked well on minimum-phase mixing, but not on non-minimum-phase mixing, as usually occurs in rooms. This led us to an FIR frequency domain infomax approach suggested by Lambert [5]. The latter approach shows much better separation of speech and music mixed in a real-room. Based on these techniques, it should now be possible to develop real-world applications. 764 T. Lee, A. 1. Bell and R. H. Lambert (A) Mixtures (8) Separations 1'1.11 .:M~i.t; ... :.rr'.~:" I ~ II~ 1.-+--• ....--: s-:-ec-~-r-; .--.--'-; -.. -.--r;-t-"'-~ -'1 I'.'.~!!s!';:~~.':.: .• :: : .1111 ~J l Figure 3: Real-room separation/deconvolution. (A) recorded mixtures (B) separated speech (spoken digits 1-10) and music. Acknowledgments T.W.L. is supported by the Daimler-Benz-Fellowship, and A.J.B. by a grant from the Office of Naval Research. We are grateful to Kari Torkkola for sharing his results with us, and to Jiirgen Fritsch, Terry Sejnowski and Alex Waibel for discussions and comments. References [1] Amari S-I. Cichocki A. & Yang H.H. 1996. A new learning algorithm for blind signal separation, Advances in Neural Information Processing Systems 8, MIT press. [2] Bell A.J. & Sejnowski T.J. 1995. An information maximisation approach to blind separation and blind deconvolution, Neural Computation, 7, 1129-1159 [3] Cardoso J-F. & Laheld B. 1996. Equivariant adaptive source separation, IEEE Trans. on Signal Proc., Dec. 1996 [4] Cichocki A., Amari S-I & Coo J. 1996. Blind separation of delayed and convolved signals with self-adaptive learning rate, in Proc. Intern. Symp. on Nonlinear Theory and Applications (NOLTA *96), Kochi, Japan. [5] Lambert R. 1996.Multichannel blind deconvolution: FIR matrix algebra and separation of multipath mixtures, PhD Thesis, University of Southern California, Department of Electrical Engineering, May 1996. [6] Lee T-W. & Orglmeister R. Blind source separation of real-world signals. submitted to Proc. ICNN, Houston, USA, 1997. [7] Platt J.C. & Faggin F. 1992. Networks for the separation of sources that are superimposed and delayed, in Moody J.E et al (eds) Advances in Neural Information Processing Systems 4, Morgan-Kaufmann [8] Torkkola K. 1996. Blind separation of delayed sources based on information maximisation, Proc IEEE ICASSP, Atlanta, May 1996. [9] Torkkola K. 1996. Blind separation of convolved sources based on information maximisation, Proc. IEEE Workshop on Neural Networks and Signal Processing, Kyota, Japan, Sept. 1996
1996
89
1,290
Dual Kalman Filtering Methods for Nonlinear Prediction, Smoothing, and Estimation Eric A. Wan ericwan@ee.ogi.edu Alex T. Nelson atnelson@ee.ogi.edu Department of Electrical Engineering Oregon Graduate Institute P.O. Box 91000 Portland, OR 97291 Abstract Prediction, estimation, and smoothing are fundamental to signal processing. To perform these interrelated tasks given noisy data, we form a time series model of the process that generates the data. Taking noise in the system explicitly into account, maximumlikelihood and Kalman frameworks are discussed which involve the dual process of estimating both the model parameters and the underlying state of the system. We review several established methods in the linear case, and propose severa! extensions utilizing dual Kalman filters (DKF) and forward-backward (FB) filters that are applicable to neural networks. Methods are compared on several simulations of noisy time series. We also include an example of nonlinear noise reduction in speech. 1 INTRODUCTION Consider the general autoregressive model of a noisy time series with both process and additive observation noise: x(k) y(k) I(x(k - 1), ... x(k - M), w) + v(k - 1) x(k) + r(k), (1) (2) where x(k) corresponds to the true underlying time series driven by process noise v(k), and 10 is a nonlinear function of past values of x(k) parameterized by w. 794 E. A. Wan and A. T. Nelson The only available observation is y(k) which contains additional additive noise r(k) . Prediction refers to estimating an x(k) given past observations. (For purposes of this paper we will restrict ourselves to univariate time series.) In estimation, x(k) is determined given observations up to and including time k. Finally, smoothing refers to estimating x(k) given all observations, past and future. The minimum mean square nonlinear prediction of x(k) (or of y(k)) can be written as the conditional expectation E[x(k)lx(k - 1)], where x(k) = [x(k), x(k 1),· .. x(O)] . If the time series x(k) were directly available, we could use this data to generate an approximation of the optimal predictor. However, when x(k) is not available (as is generally the case), the common approach is to use the noisy data directly, leading to an approximation of E[y(k)ly(k -1)]. However, this results in a biased predictor: E[y(k)ly(k-l)] = E[x(k)lx(k -1) +R(k -1)] i= E[x(k)lx(k-l)]. We may reduce the above bias in the predictor by exploiting the knowledge that the observations y(k) are measurements arising from a time series. Estimates x(k) are found (either through estimation or smoothing) such that Ilx(k) - x(k)11 < II x (k ) - y( k) II. These estimates are then used to form a predictor that approximates E[x(k)lx(k - 1)].1 In the remainder of this paper, we will develop methods for the dual estimation of both states x and weights Vi. We show how a maximum-likelihood framework can be used to relate several existing algorithms and how established linear methods can be extended to a nonlinear framework. New methods involving the use of dual Kalman filters are also proposed and experiments are provided to compare results. 2 DUAL ESTIMATION Given only noisy observations y(k), the dual estimation problem requires consideration of both the standard prediction (or output) errors ep(k) = y(k) - f(ic.(k-1)' w) as well as the observation (or input) errors eQ(k) = y(k) - x(k) . The minimum observation error variance equals the noise variance 0';. The prediction error, however, is correlated with the observation error since y(k) - f(x(k - 1)) = r(k - 1) + v(k), and thus has a minimum variance of 0'; + 0';. Assuming the errors are Gaussian, we may construct a log-likelihood function which is proportional to eT:E-1e, where eT = [eQ(O), eQ(l) .... eQ(N), ep(M), ep(M + 1), ... ep(N)], a vector of all errors up to time N, and o o o (3) Minimization of the log-likelihood function leads to the maximum-likelihood estimates for both x(k) and w. (Although we may also estimate the noise variances 0'; and 0';, we will assume in this paper that they are known.) Two general frameworks for optimization are available: lBecause models are trained on estimated data x(k), it is important that estimated data still be used for prediction of out-of training set (on-line) data. In other words, if our model was formed as an approximation of E[x(k)lx(k - 1)], then we should not provide it with y(k - 1) as an input in order to avoid a model mismatch. Dual Kalman Filtering Methods 795 2.1 Errors-In-Variables (EIV) Methods This method comes from the statistics literature for nonlinear regression (see Seber and Wild, 1989), and involves batch optimization of the cost function in Equation 3. Only minor modifications are made to account for the time series model. These methods, however, are memory intensive (E is approx. 2N >< 2N) and also do not accommodate new data in an efficient manner. Retraining is necessary on all the data in order to produce estimates for the new data points. If we ignore the cross correlation between the prediction and observation error, then E becomes a diagonal matrix and the cost function may be expressed as simply 2::=1 "Ye~(k) + e~(k), with "Y = (J';/((J'; + (J';). This is equivalent to the Gleaming (CLRN) cost function (Weigend, 1995), developed as a heuristic method for cleaning the inputs in neural network modelling problems. While this allows for stochastic optimization, the assumption in the time series formulation may lead to severely biased results. Note also that no estimate is provided for the last point x(N). When the model/ = wT x is known and linear, EIV reduces to a standard (batch) weighted least squares procedure which can be solved in closed form to generate a maximum-likelihood estimate of the noise free time series. However, when the linear model is unknown, the problem is far more complicated. The inner product of the parameter vector w with the vector x( k - 1) indicates a bilinear relationship between these unknown quantities. Solving for x( k) requires knowledge of w, while solving for w requires x(k). Iterative methods are necessary to solve the nonlinear optimization, and a Newton's-type batch method is typically employed. An EIV method for nonlinear models is also readily developed, but the computational expense makes it less practical in the context of neural networks. 2.2 Kalman Methods Kalman methods involve reformulation of the problem into a state-space framework in order to efficiently optimize the cost function in a recursive manner. At each time point, an optimal estimation is achieved by combining both a prior prediction and new observation. Connor (1994), proposed using an Extended Kalman filter with a neural network to perform state estimation alone. Puskorious and Feldkamp (1994) and others have posed the weight estimation in a state-space framework to allow Kalman training of a neural network. Here we extend these ideas to include the dual Kalman estimation of both states and weights for efficient maximum-likelihood optimization. We also introduce the use offorward-backward in/ormation filters and further explicate relationships to the EIV methods. A state-space formulation of Equations 1 and 2 is as follows: where x(k) y(k) [ x(k) 1 x(k - 1) x(k) = . ~(k - M + 1) = F[x(k - 1)] + Bv(k - 1) = Cx(k) + r(k) [ f(x(k), ... , x(k - M + 1), w) 1 F[x(k)] = ~(k) x(k - M + 2) (4) (5) B = [ il' (6) 796 E. A. Wan and A. T. Nelson and C = BT. If the model is linear, then f(x(k)) takes the form wT x(k), and F[x(k)] can be written as Ax(k), where A is in controllable canonical form. If the model is linear, and the parameters ware known, the Kalman filter (KF) algorithm can be readily used to estimate the states (see Lewis, 1986). At each time step, the filter computes the linear least squares estimate x(k) and prediction x-(k), as well as their error covariances, Px(k) and P.;(k). In the linear case with Gaussian statistics, the estimates are the minimum mean square estimates. With no prior information on x, they reduce to the maximum-likelihood estimates. Note, however, that while the Kalman filter provides the maximum-likelihood estimate at each instant in time given all past data, the EIV approach is a batch method that gives a smoothed estimate given all data. Hence, only the estimates x(N) at the final time step will match. An exact equivalence for all time is achieved by combining the Kalman filter with a backwards information filter to produce a forward-backward (FB) smoothing filter (Lewis, 1986).2 Effectively, an inverse covariance is propagated backwards in time to form backwards state estimates that are combined with the forward estimates. When the data set is large, the FB filter offers Significant computational advantages over the batch form. When the model is nonlinear, the Kalman filter cannot be applied directly, but requires a linearization of the nonlinear model at the each time step. The resulting algorithm is known as the extended Kalman filter (EKF) and effectively approximates the nonlinear function with a time-varying linear one. 2.2.1 Batch Iteration for Unknown Models Again, when the linear model is unknown, the bilinear relationship between the time series estimates, X, and the weight estimates, Vi requires an iterative optimization. One approach (referred to as LS-KF) is to use a Kalman filter to estimate x(k) with Vi fixed, followed by least-squares optimization to find Vi using the current x( k). Specifically, the parameters are estimated as Vi = (X~FXKF) -1 XKFY, where XKF is a matrix of KF state estimates, and Y is a 1 x N vector of observations. For nonlinear models, we use a feedforward neural network to approximate f(·), and replace the LS and KF procedures by backpropagation and extended Kalman filtering, respectively (referred to here as BP-EKF, see Connor 1994). A disadvantage of this approach is slow convergence, due to keeping a set of inaccurate estimates fixed at each batch optimization stage. 2.2.2 Dual Kalman Filter Another approach for unknown models is to concatenate both wand x into a joint state vector. The model and time series are then estimated simultaneously by applying an EKF to the nonlinear joint state equations (see Goodwin and Sin, 1994 for the linear case). This algorithm, however, has been known to have convergence problems. An alternative is to construct a separate state-space formulation for the underlying weights as follows: w(k) y(k) w(k -1) = f(ic.(k - 1), w(k)) + n(k), (7) (8) 2 A slight modification of the cost in Equation 3 is necessary to account for initial conditions in the Kalman form. Dual Kalman Filtering Methods 797 where the state transition is simply an identity matrix, and f(x(k-1), w(k)) plays the role of a time-varying nonlinear observation on w. When the unknown model is linear, the observation takes the form x(k _1)Tw(k). Then a pair of dual Kalman filters (DKF) can be run in parallel, one for state estimation, and one for weight estimation (see Nelson, 1976). At each time step, all current estimates are used. The dual approach essentially allows us to separate the non-linear optimization into two linear ones. Assumptions are that x and w remain uncorrelated and that statistics remain Gaussian. Note, however, that the error in each filter should be accounted for by the other. We have developed several approaches to address this coupling, but only present one here for the sake of brevity. In short, we write the variance of the noise n( k) as 0 p~ (k )OT + (J';. in Equation 8, and replace v(k - 1) by v(k - 1) + (w(k)T - wT(k))x(k - 1) in Equation 4 for estimation of x(k). Note that the ability to couple statistics in this manner is not possible in the batch approaches. We further extend the DKF method to nonlinear neural network models by introducing a dual extended Kalman filtering method (DEKF). This simply requires that Jacobians of the neural network be computed for both filters at each time step. Note, by feeding x(k) into the network, we are implicitly using a recurrent network. 2.2.3 Forward-Backward Methods All of the Kalman methods can be reformulated by using forward-backward (FB) Kalman filtering to further improve state smoothing. However, the dual Kalman methods require an interleaving of the forward and backward state estimates in order to generate a smooth update at each time step. In addition, using the FB estimates requires caution because their noncausal nature can lead to a biased w if they are used improperly. Specifically, for LS-FB the weights are computed as: w = (XRFXFB)-lXKFY ,where XFB is a matrix of FB (smooth) state estimates. Equivalent adjustments are made to the dual Kalman methods. Furthermore, a model of the time-reversed system is required for the nonlinear case. The explication and results of these algorithms will be appear in a future publication. 3 EXPERIMENTS Table 1 compares the different approaches on two linear time series, both when the linear model is known and when it is unknown. The least square (LS) estimation for the weights in the bottom row represents a baseline performance wherein no noise model is used. In-sample training set predictions must be interpreted carefully as all training set data is being used to optimize for the weights. We see that the Kalman-based methods perform better out of training set (recall the model-mismatch issuel ). Further, only the Kalman methods allow for on-line estimations (on the test set, the state-estimation Kalman filters continue to operate with the weight estimates fixed). The forward-backward method further improves performance over KF methods. Meanwhile, the clearning-equivalent cost function sacrifices both state and weight estimation MSE for improved in-sample prediction; the resulting test set performance is significantly worse. Several time series were used to compare the nonlinear methods, with the results summarized in Table 2. Conclusions parallel those for the linear case. Note, the DEKF method performed better than the baseline provided by standard backprop798 MLJ:<; CLRN KF FB EIV CLRN LS-KF LS-FB UK!" DFB LS E. A. Wan and A. T. Nelson Table 1: Comparison of methods for two linear models Model Known Tram 1 Test 1 Tram 2 Test 2 Est. Pred. Est. Pred. w J:<;st. Pred. J:<;st. Pred. .094 .322 1.09 .165 .558 1.32 .203 .134 1.08 .343 .342 1.32 .134 .559 .132 0.59 .197 .778 .221 0.85 .094 .559 .132 0.59 .165 .778 .221 0.85 Model Unknown Est. Pred. Est. Pred. w Est. Pred. Est. Pred. .172 .545 1.81 .278 .049 14.1 .138 .563 .139 .605 .134 .197 .778 .226 0.85 .099 .347 .136 .603 .281 .169 .612 .229 0.89 .135 .557 .133 .595 .212 .198 .779 .221 .863 .096 .329 .134 .596 .187 .165 .587 .221 .859 .886 1.09 .612 1.08 1.32 w w .122 11.28 .325 .369 .149 .065 0.590 MSE values for estimation (Est.), prediction (Pred.) and weights (w) (normalized to signal var.). 1 - AR(ll) model, (1'; = 4, (1'; = 1. 2000 training samples, 1000 testing samples. EIV and CLRN were not computed for the unknown model due to memory constraints. 2 - AR(5) model, (1'; = .7., (1'; = .5., 375 training, 125 testing. Table 2: Comparison of methods on nonlinear time series NNet 1 NNet 2 NNet 3 Tram Test Tram Test Tram Test Es. Pro Es . Pro Es. Pro Es. Pro Es. Pro Es. Pro BP-EKF .17 . 58 .15 .63 .08 .31 .08 .33 .16 .59 .17 .59 DEKF .14 .57 .13 .59 .07 .30 .06 .32 .14 .56 .14 .55 BP .95 .57 .95 .69 .22 .30 .29 .36 .92 .68 .92 .68 The series Nnet 1,2,3 are generated by autoregressive neural networks which exhibit limit cycle and chaotic behavior. (1'; = .16, (1'; = .81, 2700 training samples, 1300 testing samples. All network models fit using 10 inputs and 5 hidden units. Cross-validation was not used in any of the methods. agation (wherein no model of the noise is used). The DEKF method exhibited fast convergence, requiring only 10-20 epochs for training. A DEFB method is under development. The DEKF was tested on a speech signal corrupted with simulated bursting white noise (Figure 1). The method was applied to successive 64ms (512 point) windows of the signal, with a new window starting every 8ms (64 points). The results in the figure were computed assuming both (1'; and (1'; were known. The average SNR is improved by 9.94 dB. We also ran the experiment when (1'; and (1'; were estimated using only the noisy signal (Nelson and Wan, 1997), and acheived an SNR improvement of 8.50 dB. In comparison, available "state-of-the-art" techniques of spectral subtraction (Boll, 1979) and RASTA processing (Hermansky et al., 1995), achieve SNR improvements of only .65 and 1.26 dB, respectively. We extend the algorithms to the colored noise case in a second paper (Nelson and Wan, 1997). 4 CONCLUSIONS We have described various methods under a Kalman framework for the dual estimation of both states and weights of a noisy time series. These methods utilize both Dual Kalman Filtering Methods 799 Clean Speech Noise Noisy Speech IIIIiII Cleaned Speech -1Itf .... , ~. t .... ~ • ..------il'" ? ... -'" Figure 1: Cleaning Noisy Speech With The DEKF. 33,000 pts (5 sec.) shown. process and observation noise models to improve estimation performance. Work in progress includes extensions for colored noise, blind signal separation, forwardbackward filtering, and noise estimation. While further study is needed, the dual extended Kalman filter methods for neural network prediction, estimation, and smoothing offer potentially powerful new tools for signal processing applications. Acknowledgements This work was sponsored in part by NSF under grant ECS-9410823 and by ARPA/ AASERT Grant DAAH04-95-1-0485. References S.F. Boll. Suppression of acoustic noise in speech using spectral subtraction. IEEE ASSP-27, pp. 113-120. April 1979. J. Connor, R. Martin, L. Atlas. Recurrent neural networks and robust time series prediction. IEEE Tr. on Neural Networks. March 1994. F. Lewis. Optimal Estimation John Wiley & Sons, Inc. New York. 1986. G. Goodwin, K.S. Sin. Adaptive Filtering Prediction and Control. Prentice-Hall, Inc., Englewood Cliffs, NJ. 1994. H. Hermansky, E. Wan, C. Avendano. Speech enhancement based on temporal processing. ICASSP Proceedings. 1995. A. Nelson, E. Wan. Neural speech enhancement using dual extended Kalman filtering. Submitted to ICNN'97. L. Nelson, E. Stear. The simultaneous on-line estimation of parameters and states in linear systems. IEEE Tr. on Automatic Control. February, 1976. G. Puskorious, L. Feldkamp. Neural control of nonlinear dynamic systems with kalman filter trained recurrent networks. IEEE Tm. on NN, vol. 5, no. 2. 1994. G. Seber, C. Wild. Nonlinear Regression. John Wiley & Sons. 1989. A. Weigend, H.G. Zimmerman. Clearning. University of Colorado Computer Science Technical Report CU-CS-772-95. May, 1995.
1996
9
1,291
A New Approach to Hybrid HMMJANN Speech Recognition Using Mutual Information Neural Networks G. Rigoll, c. Neukirchen Gerhard-Mercator-University Duisburg Faculty of Electrical Engineering Department of Computer Science Bismarckstr. 90, Duisburg, Germany ABSTRACT This paper presents a new approach to speech recognition with hybrid HMM/ANN technology. While the standard approach to hybrid HMMI ANN systems is based on the use of neural networks as posterior probability estimators, the new approach is based on the use of mutual information neural networks trained with a special learning algorithm in order to maximize the mutual information between the input classes of the network and its resulting sequence of firing output neurons during training. It is shown in this paper that such a neural network is an optimal neural vector quantizer for a discrete hidden Markov model system trained on Maximum Likelihood principles. One of the main advantages of this approach is the fact, that such neural networks can be easily combined with HMM's of any complexity with context-dependent capabilities. It is shown that the resulting hybrid system achieves very high recognition rates, which are now already on the same level as the best conventional HMM systems with continuous parameters, and the capabilities of the mutual information neural networks are not yet entirely exploited. 1 INTRODUCTION Hybrid HMM/ANN systems deal with the optimal combination of artificial neural networks (ANN) and hidden Markov models (HMM). Especially in the area of automatic speech recognition, it has been shown that hybrid approaches can lead to very powerful and efficient systems, combining the discriminative capabilities of neural networks and the superior dynamic time warping abilities of HMM's. The most popular hybrid approach is described in (Hochberg, 1995) and replaces the component modeling the emission probabilities of the HMM by a neural net. This is possible, because it is shown Mutual In/ormation Neural Networks/or Hybrid HMMIANN Speech Recognition 773 in (Bourlard, 1994) that neural networks can be trained so that the output of the m-th neuron approximates the posterior probability p(QmIX). In this paper, an alternative method for constructing a hybrid system is presented. It is based on the use of discrete HMM's which are combined with a neural vector quantizer (VQ) in order to form a hybrid system. Each speech feature vector is presented to the neural network, which generates a firing neuron in its output layer. This neuron is processed as VQ label by the HMM's. There are the following arguments for this alternative hybrid approach: • The neural vector quantizer has to be trained on a special information theory criterion, based on the mutual information between network input and resulting neuron firing sequence. It will be shown that such a network is the optimal acoustic processor for a discrete HMM system, resulting in a profound mathematical theory for this approach. • Resulting from this theory, a formula can be derived which jointly describes the behavior of the HMM and the neural acoustic processor. In that way, both systems can be described in a unified manner and both major components of the hybrid system can be trained using a unified learning criterion. • The above mentioned theoretical background leads to the development of new neural network paradigms using novel training algorithms that have not been used before in other areas of neurocomputing, and therefore represent major challenges and issues in learning and training for neural systems. • The neural networks can be easily combined with any HMM system of arbitrary complexity. This leads to the combination of optimally trained neural networks with very powerful HMM's, having all features useful for speech recognition, e.g. triphones, function words, crossword triphones, etc .. Context-dependency, which is very desirable but relatively difficult to realize with a pure neural approach, can be left to the HMM's. • The resulting hybrid system has still the basic structure of a discrete system, and therefore has all the effective features associated with discrete systems, e.g. quick and easy training as well as recognition procedures, real-time capabilities, etc .. • The work presented in this paper has been also successfully implemented for a demanding speech recognition problem, the 1000 word speaker-independent continuous Resource Management speech recognition task. For this task, the hybrid system produces one of the best recognition results obtained by any speech recognition system. In the following section, the theoretical foundations of the hybrid approach are briefly explained. A unified probabilistic model for the combined HMMIANN system is derived, describing the interaction of the neural and the HMM component. Furthermore, it is shown that the optimal neural acoustic processor can be obtained from a special information theoretic network training algorithm. 2 INFORMATION THEORY PRINCIPLES FOR NEURAL NETWORK TRAINING We are considering now a neural network of arbitrary topology used as neural vector quantizer for a discrete HMM system. If K patterns are presented to the hybrid system during training, the feature vectors resulting from these patterns using any feature extraction method can be denoted as x(k), k=l...K. If these feature vectors are presented to the input layer of a neural network, the network will generate one firing neuron for each presentation. Hence, all K presentations will generate a stream of firing neurons with length K resulting from the output layer of the neural net. This label stream is denoted as Y=y(l) ... y(K). The label stream Y will be presented to the HMM's, which calculate the probability that this stream has been observed while a pattern of a certain class has been presented to the system. It is assumed, that M different classes Qm are active in the 774 G. Rigoll and C. Neukirchen system, e.g. the words or phonemes in speech recognition. Each feature vector ~(k) will belong to one of these classes. The class Om, to which feature vector ~(k) belongs is denoted as Q(k). The major training issue for the neural network can be now formulated as follows: How should the weights of the network be trained, so that the network produces a stream of firing neurons that can be used by the discrete HMM's in an optimal way? It is known that HMM's are usually trained with information theory methods which mostly rely on the Maximum Likelihood (ML) principle. If the parameters of the hybrid system (i.e. transition and emission probabilities and network weights) are summarized in the vector !!, the probability P!!(x(k)IQ(k» denotes the probability of the pattern X at discrete time k, under the assumption that it has been generated by the model representing class O(k), with parameter set !!. The ML principle will then try to maximize the joint probability of all presented training patterns ~(k), according to the following Maximum Likelihood function: fl* = arg max {~ i log P!! (K(k) I Q(k»j ~ k=1 (1) where !!* is the optimal parameter vector maximizing this equation. Our goal is to feed the feature vector ~ into a neural network and to present the neural network output to the Markov model. Therefore, one has to introduce the neural network output in a suitable manner into the above formula. If the vector ~ is presented to the network input layer, and we assume that there is a chance that any neuron Yn, n=1...N (with network output layer size N) can fire with a certain probability, then the output probability p(~IQ) in (1) can be written as: N N p(KIQ) = I p(x ,Y n IQ) = I p(y n IQ) . p(x Iy n,Q) n=1 n=1 (2) Now, the combination of the neural component with the HMM can be made more obvious: In (2), typically the probability P(YnIQ) will be described by the Markov model, in terms of the emission probabilities of the HMM. For instance, in continuous parameter HMM's, these probabilities are interpreted as weights for Gaussian mixtures. In the case of semi-continuous systems or discrete HMM's, these probabilities will serve as discrete emission probabilities of the codebook labels. The probability p(xIYn,Q) describes the acoustic processor of the system and is characterizing the relation between the vector ~ as input to the acoustic processor and the label Yn, which can be considered as the n-th output component of the acoustic processor. This n-th output component may characterize e.g. the n-th Gaussian mixture component in continuous parameter HMM's, or the generation of the n-th label of a vector quantizer in a discrete system. This probability is often considered as independent of the class 0 and can then be expressed as p(xIYn). It is exactly this probability, that can be modeled efficiently by our neural network. In this case, the vector X serves as input to the neural network and Yn characterizes the n-th neuron in the output layer of the network. Using Bayes law, this probability can be written as: P(YnIK) ' pW p(xl Y n) = p(y n) (3) yielding for (2): (4) Using again Bayes law to express Mutual Information Neural Networks for Hybrid HMMIANN Speech Recognition 775 (5) one obtains from (4): p(K) N p(KI.Q)= -(.Q) . L p(.Qlyn) ·p(ynlo!J p n=1 (6) We have now modified the class-dependent probability of the feature vector X in a way that allows the incorporation of the probability P(YnIX). This probability allows a better characterization of the behavior of the neural network, because it describes the probability of the various neurons Yn, if the vector X is presented to the network input. Therefore, these probabilities give a good description of the input/output behavior of the neural network. Eq. (6) can therefore be considered as probabilistic model for the hybrid system, where the neural acoustic processor is characterized by its input/output behavior. Two cases can be now distinguished: In the first case, the neural network is assumed to be a probabilistic paradigm, where each neuron fires with a certain probability, if an input vector is presented. In this case all neurons contribute to the information forwarded to the HMM's. As already mentioned, in this paper, the second possible case is considered, namely that only one neuron in the output layer fires and will be fed as observed label to the HMM. In this case, we have a deterministic decision, and the probability P(YnIX) describes what neuron Yn* fires if vector X is presented to the input layer. Therefore, this probability reduces to (7) Then, (6) yields: (8) Now, the class-dependent probability p(Xln) is expressed through the probability p(nIYn*), involving directly the firing neuron Yn*, when feature vector X is presented. One has now to turn back to (1), recalling the fact, that this equation describes the fact that the Markov models are trained with the ML criterion. It should also be recalled, that the entire sequence of feature vectors, x(k), k=l...K, results in a label stream of firing neurons Yn*(k), k=l...K, where Yn*(k) is the firing neuron if the k-th vector x(k) is presented to the neural network. Now, (8) can be substituted into (1) for each presentation k, yielding the modified ML criterion: { K p(x(k)) } 1( = arg;ax ::1 log P(Q (k)) . p(.Q(k) I Y n*,k)) ~ arg;ax {~, log p(x (k)) - ~109P(Q(k)) + ~IOg p(Q(k) I Y n.(k))} (9) Usually, in a continuous parameter system, the probability p(x) can be expressed as: N p(K) = LP(K,ly n) . p(y n) n=1 (10) and is therefore dependent of the parameter vector ft, because in this case, p(xIYn) can be interpreted as the probability provided by the Gaussian distributions, and the parameters of 776 G. Rigoll and C. Neukirchen the Gaussians will depend on ft. As just mentioned before, in a discrete system, only one firing neuron Yn* survives, resulting in the fact that only the n*-th member remains in the sum in (10). This would correspond to only one "firing Gaussian" in the continuous case, leading to the following expression for p(x): p(K) = p(x Iy nJ· p(y nJ = p(K,y nJ = p(y n"lx) . p(x) (11) Considering now the fact, that the acoustic processor is not represented by a Gaussian but instead by a vector quantizer, where the probability P(Yn*IX) of the firing neuron is equal to 1, then (11) reduces to p(~) = p(x) and it becomes obvious that this probability is not affected by any distribution that depends on the parameter vector ft. This would be different, if P(Yn*IX) in (11) would not have binary characteristics as in (7), but would be computed by a continuous function which in this case would depend on the parameter vector ft. Thus, without consideration of p(X), the remaining expression to be maximized in (9) reduces to: ,r( = arg;ax [~ ~IOg p(.Q( k)) + ! log p(.Q( k) I Y n·(k)) 1 = arg max [- E {log p(.o)} + E {log p(.o I y n")}] fJ.. (12) These expectations of logarithmic probabilities are also defined as entropies. Therefore, (9) can be also written as fl." = arg max {H (.0) - H(.o I Y)} fJ.. (13) This equation can be interpreted as follows: The term on the right side of (13) is also known as the mutual information I(n,Y~ between the probabilistic variables nand Y, i.e. : 1(.0, Y) =H(.o) - H (.01 Y) =H (Y) - H(YI.o) (14) Therefore, the final information theory-based training criterion for the neural network can be formulated as follows: The synaptic weights of the neural network should be chosen as to maximize the mutual information between the string representing the classes of the vectors presented to the network input layer during training and the string representing the resulting sequence of firing neurons in the output layer of the neural network. This can be also expressed as the Maximum Mutual Information (MMI) criterion for neural network training. This concludes the proof that MMI neural networks are indeed optimal acoustic processors for HMM's trained with maximum likelihood principles. 3 REALIZATION OF MMI TRAINING ALGORITHMS FOR NEURAL NETWORKS Training the synaptic weights of a neural network in order to achieve mutual information maximization is not easy. Two different algorithms have been developed for this task and can only be briefly outlined in this paper. A detailed description can be found in (Rigoll, 1994) and (Neukirchen, 1996). The first experiments used a single-layer neural network with Euclidean distance as propagation function. The first implementation of the MMI training paradigm has been realized in (Rigoll, 1994) and is based on a self-organizing procedure, starting with initial weights derived from k-means clustering of the training vectors, followed by an iterative procedure to modify the weights. The mutual information increases in a self-organizing way from a low value at the start to a much higher value after several iteration cycles. The second implementation has been realized Mutual Information Neural Networks for Hybrid HMMIANN Speech Recognition 777 recently and is described in detail in (Neukirchen, 1996). It is based on the idea of using gradient methods for finding the MMI value. This technique has not been used before, because the maximum search for finding the firing neuron in the output layer has prevented the calculation of derivatives. This maximum search can be approximated using the softmax function, denoted as sn for the n-th neuron. It can be computed from the activations Zl of all neurons as: N z IT "" Z I IT Sn=e n / £..Je /=1 (15) where a small value for parameter T approximates a crisp maximum selection. Since the string n in (14) is always fixed during training and independent of the parameters in ft, only the function H(nIY) has to be minimized. This function can also be expressed as M N H(!2 I Y) = - L L p(y n,!2m ) ·logp(!2m I Y n) m=1 n=1 m=1 n=1 (16) A derivative with respect to a weight Wlj of the neural network yields: aH (!21 Y) = JW/j (17) As shown in (Neukirchen, 1996), all the required terms in (17) can be computed effectively and it is possible to realize a gradient descend method in order to maximize the mutual information of the training data. The great advantage of this method is the fact that it is now possible to generalize this algorithm for use in all popular neural network architectures, including multilayer and recurrent neural networks. 4 RESULTS FOR THE HYBRID SYSTEM The new hybrid system has been developed and extensively tested using the Resource Management 1000 word speaker-independent continuous speech recognition task. First, a baseline discrete HMM system has been built up with all well-known features of a context-dependent HMM system. The performance of that baseline system is shown in column 2 of Table 1. The 1st column shows the performance of the hybrid system with the neural vector quantizer. This network has some special features not mentioned in the previous sections, e.g. it uses multiple frame input and has been trained on contextdependent classes. That means that the mutual information between the stream of firing neurons and the corresponding input stream of triphones has been maximized. In this way, the firing behavior of the network becomes sensitive to context-dependent units. Therefore, this network may be the only existing context-dependent acoustic processor, carrying the principle of triphone modeling from the HMM structure to the acoustic front end. It can be seen, that a substantially higher recognition performance is obtained with the hybrid system, that compares well with the leading continuous system (HTK, in column 3). It is expected, that the system will be further improved in the near future through various additional features, including full exploitation of multilayer neural VQ's 778 G. Rigoll and C. Neukirchen and several conventional HMM improvements, e.g. the use of crossword triphones. Recent results on the larger Wall Street Journal (WSJ) database have shown a 10.5% error rate for the hybrid system compared to a 13.4% error rate for a standard discrete system, using the 5k vocabulary test with bigram language model of perplexity 110. This error rate can be further reduced to 8.9% using crossword triphones and 6.6% with a trigram language model. This rate compares already quite favorably with the best continuous systems for the same task. It should be noted that this hybrid WSJ system is still in its initial stage and the neural component is not yet as sophisticated as in the RM system. 5 CONCLUSION A new neural network paradigm and the resulting hybrid HMMI ANN speech recognition system have been presented in this paper. The new approach performs already very well and is still perfectible. It gains its good performance from the following facts: (1) The use of information theory-based training algorithms for the neural vector quantizer, which can be shown to be optimal for the hybrid approach. (2) The possibility of introducing context-dependency not only to the HMM's, but also to the neural quantizer. (3) The fact that this hybrid approach allows the combination of an optimal neural acoustic processor with the most advanced context-dependent HMM system. We will continue to further implement various possible improvements for our hybrid speech recognition system. REFERENCES Rigoll, G. (1994) Maximum Mutual Information Neural Networks for Hybrid Connectionist-HMM Speech Recognition Systems, IEEE Transactions on Speech and Audio Processing, Vol. 2, No.1, Special Issue on Neural Networks for Speech Processing, pp. 175-184 Neukirchen, C. & Rigoll, G. (1996) Training of MMI Neural Networks as Vector Quantizers, Internal Report, Gerhard-Mercator-University Duisburg, Faculty of Electrical Engineering, available via http://www.fb9-tLuni-duisburg.de/veroeffentl.html Bourlard, H. & Morgan, N. (1994) Connectionist Speech Recognition: A Hybrid Approach, Kluwer Academic Publishers Hochberg, M., Renals, S., Robinson, A., Cook, G. (1995) Recent Improvements to the ABBOT Large Vocabulary CSR System, in Proc. IEEE-ICASSP, Detroit, pp. 69-72 Rigoll, G., Neukirchen, c., Rottland, J. (1996) A New Hybrid System Based on MMINeural Networks for the RM Speech Recognition Task, in Proc. IEEE-ICASSP, Atlanta Table 1: Comparison of recognition rates for different speech recognition systems RM SI word recognition rate with word pair grammar: correctness (accuracy) test set hybrid MMI-NN baseline k-means continuous pdf system system VQ system (HTK) Feb.'89 96,3 % (95,6 %) 94,3 % (93,6 %) 96,0 % (95,5 %) Oct.'89 95,4 % (94,5 %) 93,5 % (92,0 %) 95,4% (94,9 %) Feb.'91 96,7 % (95,9 %) 94,4% (93,5 %) 96,6% (96,0 %) Sep.'92 93,9 % (92,5 %) 90,7 % (88,9 %) 93,6 % (92,6 %) average 95,6 % (94,6 %) 93,2 % (92,0 %) 95,4% (94,7 %)
1996
90
1,292
Competition Among Networks Improves Committee Performance Paul W. Munro Department of Infonnation Science and Telecommunications University of Pittsburgh Pittsburgh PA 15260 munro@sis.pitt.edu Bambang Parman to Department of Health Infonnation Management University of Pittsburgh Pittsburgh PA 15260 parmanto+@pitt.edu ABSTRACT The separation of generalization error into two types, bias and variance (Geman, Bienenstock, Doursat, 1992), leads to the notion of error reduction by averaging over a "committee" of classifiers (Perrone, 1993). Committee perfonnance decreases with both the average error of the constituent classifiers and increases with the degree to which the misclassifications are correlated across the committee. Here, a method for reducing correlations is introduced, that uses a winner-take-all procedure similar to competitive learning to drive the individual networks to different minima in weight space with respect to the training set, such that correlations in generalization perfonnance will be reduced, thereby reducing committee error. 1 INTRODUCTION The problem of constructing a predictor can generally be viewed as finding the right combination of bias and variance (Geman, Bienenstock, Doursat, 1992) to reduce the expected error. Since a neural network predictor inherently has an excessive number of parameters, reducing the prediction error is usually done by reducing variance. Methods for reducing neural network complexity can be viewed as a regularization technique to reduce this variance. Examples of such methods are Optimal Brain Damage (Le Cun et. al., 1991), weight decay (Chauvin, 1989), and early stopping (Morgan & Boulard, 1990). The idea of combining several predictors to fonn a single, better predictor (Bates & Granger, 1969) has been applied using neural networks in recent years (Wolpert, 1992; Perrone, 1993; Hashem, 1994). Competition Among Networks Improves Committee Performance 593 2 REDUCING MISCLASSIFICATION CORRELATION Since committee errors occur when too many individual predictors are in error, committee performance improves as the correlation of network misclassifications decreases. Error correlations can be handled by using a weighted sum to generate a committee prediction; the weights can be estimated by using ordinary least squares (OLS) estimators (Hashem, 1994) or by using Lagrange multipliers (Perrone, 1993). Another approach (Parmanto et al., 1994) is to reduce error correlation directly by attempting to drive the networks to different minima in weight space, that will presumably have different generalization syndromes, or patterns of error with respect to a test set (or better yet, the entire stimulus space). 2.1 Data Manipulations Training the networks using nonidentical data has been shown to improve committee performance, both when the data sets are from mutually exclusive continuous regions (eg, Jacobs et al.,1991), or when the training subsets are arbitrarily chosen (Breiman, 1992; Parmanto, Munro, and Doyle, 1995). Networks tend to converge to different weight states, because the error surface itself depends on the training set; hence changing the data changes the error surface. 2.2 Auxiliary tasks Another way to influence the networks to disagree is to introduce a second output unit with a different task to each network in the committee. Thus, each network has two outputs, a primary unit which is trained to predict the class of the input, and a secondary unit, with some other task that is different than the tasks assigned to the secondary units of the other committee members. The success of this approach rests on the assumption that the decorrelation of the network errors will more than compensate for any degradation of performance induced on the primary task by the auxiliary task. The presence of a hidden layer in each network guarantees that the two output response functions share some weight parameters (i.e., the input-hidden weights), and so the learning of the secondary task influences the function learned by the primary output unit. Parmanto et al. (1994) acheived significant decorrelation and improved performance on a varoety of tasks using one of the input variables as the training signal for the secondary unit. Interestingly, the secondary task does not necessarily degrade performance on the primary task. Our studies, as well as those of Caruana (1995), show that extra tasks can facilitate learning time and generalization performance on an individual network. On the other hand, certain auxiliary tasks interfere with the primary task. We have found however, that even when the individual performance is degraded, committee performance is nevertheless enahnced (relative to a committee of single output networks) due to the magnitude of error decorrelation. 3 THE COMPETITIVE COMMITTEE An alternative to using a stationary task per se, such as replicating an input variable or projecting onto principal components (as was done in Parmanto et ai, 1994), is to use a signal that depends on the other networks, in such a manner that the functions computed by the secondary units are negatively correlated after training. This notion is reminiscent of competitive learning (Rumelhart and Zipser, 1986); that is, the functions computed by the secondary units will partition the stimulus space. Thus, a Competitive Committee Machine (CCM) is defined as a committee of neural 594 P. W Munro and B. Parmanto network classifiers, each with two output units: a primary unit trained according to the classification task, and a secondary unit participating in a competitive process with secondary units of the other networks in the committee; let the outputs of network i be denoted Pi and Si, respectively (see Figure 1). The network weights are modified according to the following variant of the back propagation procedure. When data item a from the training set is presented to the committee during training, with input vector XCl and known output classification value yCl (binary), the networks each process XCl simultaneously, and the P and S output units of each network respond. Each P-unit receives the identical training signal, yCl, that corresponds to the input item; the training signal to the S-units is zero for all networks except the network with the greatest S-unit response among the committee; the maximum Si among the networks in the committee receives a training signal of 1, and the others receive a training signal of O. where or mof are the errors attributed to the primary and secondary units respectively to adjust network weights with back propagationl . During the course of training, the Sunit's response is explicitly trained to become sensitive to a unique region (relative to the other networks' S-units) of the stimulus space. This training signal is different from typical "tasks" that are used to train neural networks in that it is not a static function of the input; instead, since it depends on the other networks in the committee, it has a dynamic qUality. 4 RESULTS Some experiments have been run using the sine wave classification task (Figure 2) of Geman and Bienenstock (1992). Comparisons of CCM perfonnance versus the baseline perfonnance of a committee with a simple average over a range of architectures (as indicated by the number of hidden units) are favorable (Figure 3). Also, note that the improvement is primarily attributable to descreased correlation, since the average individual perfonnance is not significantly affected. Visualization of the response of the individual networks to the entire stimulus space gives a complete picture of how the networks generalize and shows the effect of the competition (Figure 4). For this particular data set, the classes are easily separated in the central region (note that all the networks do well here). But at the edges, there is much more variance in the networks trained with competitive secondary units (Figure 5). 5 DISCUSSION Caruana (1995) has demontrated significant improvement on "target" classification tasks in individual networks by adding one or more supplementary output units trained to compute tasks related to the target task. The additional output unit added to each network IFor notational convenience, the derivative factor sometimes included in the definition of 8 is not included in this description of oP and oS. Competition Among Networks Improves Committee Performance (Training signal yfl COMMITIEE OUTPUT (AVERAGE OR VOTE) • s c_ . ----. • •• Input Variables (simultaneously presented to all networks) 595 Figure 1: A Competitive Committee Machine. Each of the K networks receives the srune input and produces two outputs, P and S. The P responses of all the networks are compared to a common training signal to compute an error value for backpropagation (dark dashed arrows); the P responses are combined (by vote or by sum) to determine the committee response. The S-unit responses are compared with each other, with the "winner" (highest response) receiving a training signal of 1, and the others receiving a training signal of O. Thus the training signal for network i is computed by comparing all S-unit responses, and then fed back to the S-units, hence the two-way arrows (gray). in the CCM merges a variant of Rumelhart and Zipser's (1986) competitive learning procedure with backpropagation, to form a novel hybrid of a supervised training technique with an unsupervised method. The training signal delivered to the secondary unit under CCM is more direct than an arbitrary task, in that it is defined explicitly in terms of dissociating response properties. Note that the training signals for the S-units differ from the P-unit training signals in two important respects: 1. Not static: The signal depends on the S-unit responses/rom the other networks ,md hence chcmges during the course of training. 2. Not uniform: It is not constant across the committee (whereas the P-unit training signal is.) 596 P W Munro and B. Parmanto Fi!.!ure 2. A ("/assif/ca[ion task. Training data (bottom) is srunpled from a classitication ta;k defined hy a s'inusoid (top) conupted hy noise (middle). Committee Performance :e ...... :! ..... ......... ........................... ...... -..................• 8 10 12 14 16 , hidden unllS Correlation o~----------.. c: 0 o ~ <.Q a; 0 ~ ~ 0>0 ~ N o •.................. " ..... . .... ..... -......... " ................. . o oL..--_________ --' 4 10 12 14 16 , hidden urI/IS Indiv. Performance . ...... _-. __ ._-.. o ba&.tlne " ...... cem 8 10 12 14 16 , hidden unns Percent Improvement ~r-----~~~~---; ..•............ _.-............ .. 4 6 /,..... ..•. .. ........... . o bas.-w • ,._ .. cem 8 10 12 14 16 , hidden unns Figure 3. Peliormallce of CCM. Committees of 5 networks were trained with competitive learning (CCM) and without (baseline). Each data point is an average over 5 simulations with different initial weights. Competition Among Networks Improves Committee Performance 597 Network.l (Error: 10.20%) Network '2 (Err",: 15.59"1.) r-==== Network .3 (Error: 15.250/.) Nelwork ,4 (Error: 15.54"/.) Network #5 (Error: 12.65%) Canmillee OUtput • Thtesholded (Error: 11.64) Figure 4. Generalization plots for a committee. The level of gray indicates the response for each network of a committee trained without competition. The panel on the lower right shows the (thresholded) committee output. The average pairwise correlation of the committee is 0.91. Network #1 (Error: 10.21%) Network,2 (Error: 9.83%) Network #3 (Error. 1683%) Nelwork'4 (Error: 14.88%) Figure 5. Gelleralization plots for a CCM committee. Comparison with Figure 4 shows much more variance among the committee at the edges. Note that the committee performs much better near the right and left ends of the stimulus space than does any individual network. This committee had an error rate of 8.11 % (cf 11.64% in the baseline case). 598 P. W. Munro and B. Pannanto The weighting of oS relative to oP is an important consideration; in the simulations above, the signal from the secondary unit was arbitrarily multiplied by a factor of 0.1. While we have not yet examined this systematically, it is assumed that this factor will modulate the tradeoff between degradation of the primary task and reduction of error correlation. References Bates, J.M., and Granger, C.W. (1969) "The combination of forecasts," Operation Research Quarterly, 20(4),451-468. Breiman, L, (1992) "Stacked Regressions", TR 367, Dept. of Statistics, Univ. of Cal. Berkeley. Caruana, R (1995) "Learning many related tasks at the same time with backpropagation," In: Advances in Neural Information Processing Systems 7. D. S. Touretsky, ed. Morgan Kaufmann. Chauvin, Y. (1989) "A backpropagation algorithm with optimal use of hidden units." In Touretzky D., (ed.), Advances in Neural Information Processing 1, Denver, 1988, Morgan Kaufmann. Geman, S., Bienenstock, E., and Doursat, R. (1992) "Neural networks and the bias/variance dilemma," Neural Computation 4, 1-58. Hashem, S. (1994). Optimal Linear Combinations of Neural Networks., PhD Thesis, Purdue University. Jacobs, R.A., Jordan, M.I., Nowlan, SJ., and Hinton, G.E. (1991) "Adaptive mixtures of local experts," Neural Computation, 3, 79-87 Le Cun, Y., Denker J. and Solla, S. (1990). Optimal Brain Damage. In D. Touretzky (Ed.) Advances in Neural Information Processing Systems 2, San Mateo: Morgan Kaufmann. 598-605. Morgan, N. & Boulard, H. (1990) Generalization and parameter estimation in feedforward nets: some experiments. In D. Touretzky (Ed.) Advances in Neural Information Processing Systems 2 San Mateo: Morgan Kaufmann. Parmanto, B., Munro, P.W., Doyle, H.R., Doria, C., Aldrighetti, L., Marino, I.R., Mitchel, S., and Fung, JJ. (1994) "Neural network classifier for hepatoma detection," Proceedings of the World Congress of Neural Networks Parmanto, B., Munro, P.W., Doyle, H.R. (1996) "Improving committee diagnosis with resampling techniques," In: D. S. Touretzky, M. C. Mozer, M. E. Hasselmo, eds. Advances in Neural Information Processing Systems 8. MIT Press: Cambridge, MA. Perrone, M.P. (1993) "Improving Regression Estimation: Averaging Methods for Variance Reduction with Extension to General Convex Measure Optimization," PhD Thesis, Department of Physics, Brown University. Rumelhart. D.E and Zipser, D. (1986) "Feature discovery by competitive learning," In: Rumelhart, D.E.and McClelland, J.E. (Eds.), Parallel Distributed Processing: Explorations in the Microstructure of Cognition. MIT Press, Cambridge, MA. Wolpert, D. (1992). Stacked generalization, Neural Networks, 5,241-259.
1996
91
1,293
Selective Integration: A Model for Disparity Estimation Michael S. Gray, Alexandre Pouget, Richard S. Zemel, Steven J. Nowlan, Terrence J. Sejnowski Departments of Biology and Cognitive Science University of California, San Diego La Jolla, CA 92093 and Howard Hughes Medical Institute Computational Neurobiology Lab The Salk Institute, P. O. Box 85800 San Diego, CA 92186-5800 Email: michael, alex, zemel, nowlan, terry@salk.edu Abstract Local disparity information is often sparse and noisy, which creates two conflicting demands when estimating disparity in an image region: the need to spatially average to get an accurate estimate, and the problem of not averaging over discontinuities. We have developed a network model of disparity estimation based on disparityselective neurons, such as those found in the early stages of processing in visual cortex. The model can accurately estimate multiple disparities in a region, which may be caused by transparency or occlusion, in real images and random-dot stereograms. The use of a selection mechanism to selectively integrate reliable local disparity estimates results in superior performance compared to standard back-propagation and cross-correlation approaches. In addition, the representations learned with this selection mechanism are consistent with recent neurophysiological results of von der Heydt, Zhou, Friedman, and Poggio [8] for cells in cortical visual area V2. Combining multi-scale biologically-plausible image processing with the power of the mixture-of-experts learning algorithm represents a promising approach that yields both high performance and new insights into visual system function. Selective Integration: A Model for Disparity Estimation 867 1 INTRODUCTION In many stereo algorithms, the local correlation between images from the two eyes is used to estimate relative depth (Jain, Kasturi, & Schunk [5]). Local correlation measures, however, convey no information about the reliability of a particular disparity measurement. In the model presented here, we introduce a separate selection mechanism to determine which locations of the visual input have consistent disparity information. The focus was on several challenging viewing situations in which disparity estimation is not straightforward. For example, can the model estimate the disparity of more than one object in a scene? Does occlusion lead to poorer disparity estimation? Can the model determine the disparities of two transparent surfaces? Does the model estimate accurately the disparities present in real world images? Datasets corresponding to these different conditions were generated and used to test the model. Our goal is to develop a neurobiologically plausible model of stereopsis that accurately estimates disparity. Compared to traditional cross-correlation approaches that try to compute a depth map for all locations in space, the mixture-of-experts model used here searches for sparse, reliable patterns or configurations of disparity stimuli that provide evidence for objects at different depths. This allows partial segmentation of the image to obtain a more compact representation of disparities. Local disparity estimates are sufficient in this case, as long as we selectively segment those regions of the image with reliable disparity information. The rest of the paper is organized as follows. First, we describe the architecture of the mixture-of-experts model. Second, we provide a brief qualitative description of the model's performance followed by quantitative results on a variety of datasets. In the third section, we compare the activity of units in the model to recent neurophysiological data. Finally, we discuss these findings, and consider remaining open questions. 2 MIXTURE-OF-EXPERTS MODEL The model of stereopsis that we have explored is based on the filter model for motion detection devised by Nowlan and Sejnowski [6]. The motion problem was readily adapted to stereopsis by changing the time domain of motion to the left/right image domain for stereopsis. Our model (Figure 1) consisted of several stages and computed its output using only feed-forward processing, as described below (see also Gray, Pouget, Zemel, Nowlan, and Sejnowski [2] for more detail). The output of the first stage (disparity energy filters) became the input to two different primary pathways: (1) the local disparity networks, and (2) the selection networks. The activation of each of the four disparity-tuned output units in the model was the product of the outputs of the two primary pathways (summed across space). An objective function based on the mixture-of-experts framework (Jacobs, Jordan, Nowlan, & Hinton [4]) was used to optimize the weights from the disparity energy units to the local disparity networks and to the selection networks. The weights to the output units from the local disparity and selection pathways were fixed at 1.0. Once the model was trained, we obtained a scalar disparity estimate from the model by computing a nonlinear least squares Gaussian fit to the four output values. The mean of the Gaussian was our disparity estimate. When two objects were present 868 M. S. Gray, A. Pouget, R. S. Zemel. S. 1. Nowlan and T. 1. Sejnowski Disparity Energy Filters .---~~r'---, Competition Low SF '--_---'----''------'1 Medium SF 1 i 1 High SF i --re; :??-r i Left Eye Retina Right Eye Retina Figure 1: The mixture-of-experts architecture. in the input, we fit the sum of two Gaussians to the four output values. 2.1 DISPARITY ENERGY FILTERS The retinal layer in the model consisted of one-dimensional right eye and left eye images, each 82 pixels in length. These images were the input to disparity energy filters, as developed by Ohzawa, DeAngelis, and Freeman [7]. At the energy filter layer, there were 51 receptive field (RF) locations which received input from partially overlapping regions of the retinae. At each of these RF locations, there were 30 energy units corresponding to 10 phase differences at 3 spatial frequencies. These phase differences were proportional to disparity. An energy unit consisted of 4 energy filter pairs, each of which was a Gabor filter. The outputs of the disparity energy units were normalized at each RF location and within each spatial frequency using a soft-max nonlinearity. 2.2 LOCAL DISPARITY NETWORKS In the local disparity pathway, there were 8 RF locations, and each received a weighted input from 9 disparity energy locations. Each RF location corresponded to a local disparity network and contained a pool of 4 disparity-tuned units. Neighboring locations received input from overlapping sets of disparity energy units. Weights were shared across all RF locations for each disparity. Soft-max competition occurred within each local disparity network (across disparity), and insured that only one disparity was strongly activated at each RF location. 2.3 SELECTION NETWORKS Like the local disparity networks, the selection networks were organized into a grid of 8 RF locations with a pool of 4 disparity-tuned units at each location. These 4 units represented the local support for each of the different disparity hypotheses. It is more useful to think of the selection networks, however, as 4 separate layers each Selective Integration: A Modelfor Disparity Estimation 869 of which responded to a specific disparity across all regions of the image. Like the local disparity pathway, neighboring RF locations received input from overlapping disparity energy units, and weights were shared across space for each disparity. In addition, the outputs of the selection network were normalized with the softmax operation. This competition, however, occurred separately for each of the 4 disparities in a global fashion across space. 3 RESULTS Figure 2 shows the pattern of activations in the model when presented with a single object at a disparity of 2.1 pixels. The visual layout of the model in this figure is identical to the layout in Figure 1. The stimulus appears at bottom, with the 3 disparity energy filter banks directly above it. On the left above the disparity energy filters are the local disparity networks. The selection networks are on the right. The summed output (across space) appears in the upper right corner of the figure. Note that the selection network for a 2 pixel disparity (2nd row from the bottom in the selection pathway) is active for the spatial location at far left. The corresponding location is also highly active in the local disparity pathway, and this combination leads to strong activation for a 2 pixel disparity in the output of the model. The mixture-of-experts model was optimized individually on a variety of different datasets and then tested on novel stimuli from the same datasets. The model's ability to discriminate among different disparities was quantified as the disparity threshold the disparity difference at which one can correctly see a difference in depth 75% of the time. Disparity thresholds for the test stimuli were computed using signal-detection theory (Green & Swets [3]). Sample stimuli and their disparity thresholds are shown in Table 1. The model performed best on single object stimuli (top row). This disparity threshold (0.23 pixels) was substantially less than the input resolution of the model (1 pixel) and was thus exhibiting stereo hyperacuity. The model also performed well when there were multiple, occluding objects (2nd row). When both the stimulus and the background were generated from a uniform random distribution, the disparity threshold rose to 0.55 pixels. The model estimated disparity accurately in random-dot stereograms and real world images. Binary stereograms containing two transparent surfaces, however, were a challenging stimulus, and the threshold rose to 0.83 pixels. Part of the difficulty with this stimulus (containing two objects) was fitting the sum of 2 Gaussians to 4 data points. We have compared our mixture-of-experts model (containing both a selection pathway and a local disparity pathway) with standard backpropagation and crosscorrelation techniques (Gray et al [2]). The primary difference is that the backpropagation and cross-correlation models have no separate selection mechanism. In essence, one mechanism must compute both the segmentation and the disparity estimation. In our tests with the back-propagation model, we found that disparity thresholds for single object stimuli had risen by a factor of 3 (to 0.74 pixels) compared to the mixture-of-experts model. Disparity estimation of the cross-correlation model was similarly poor. Thresholds rose by a factor of2 (compared to the mixtureof-experts model) for both single object stimuli and the noise stimuli (threshold = 0.46, 1.28 pixels, respectively). 870 M. S. Gray, A. Pouget, R. S. Zeme~ S. J. Nowlan and T. J. Sejnowski Desired Output Space Output Actual Output :~ :~ o 3 0.98 0.00 0.87 0.00 0.98 0.01 Low SF Med SF High SF Local Dlsparit\:l Nets 1.00 0.00 0.39 0.00 Selection Nets o 3 0.88 0.00 Disparit~ Energ~ '" ~'ffi l! • . , ~. 0.38 0.00 0.82 0.00 Input Stilllulus 0.99 0.50 Figure 2: The activity in the mixture-of-experts model in response to an input stimulus containing a single object at a disparity of 2.10 pixels. At bottom is the input stimulus. The 3 regions in the middle represent the output of the disparity energy filters. Above the disparity energy output are the two pathways of the model. The local disparity networks appear to the left and the selection networks are to the right. Both the local disparity networks and the selection networks receive topographically organized input from the disparity energy filters. The selection and local disparity networks are displayed so that the top row represents a disparity of 0 pixels, the next row a 1 pixel disparity, then 2 and 3 pixel disparities in the remaining rows. At the top left part of the figure is the desired output for the given input stimulus. In the top middle is the output for each local region of space. On the top right is the actual output of the model collapsed across space. The numbers at the bottom left of each part of the network indicate the maximum and minimum activation values within that part. White indicates maximum activation level, black is minimum. Selective Integration: A Model for Disparity Estimation 871 Stimulus Type Single Double Noise Random-Dot Transparent Real Sample Stimulus ;= ·Eft "'_"'_' .. ::JI •.• I. __ ".~."' __ .iftk" .. !, ___ ;r._. . ...... ::·~·~. i;:i;;:;:· "":·· i,,·:;'·::J Threshold 0.23 0.41 0.55 0.36 0.83 0.30 Table 1: Sample stimuli for each of the datasets, and corresponding disparity thresholds (in pixels) for the mixture-of-experts model. 4 COMPARISON WITH NEUROPHYSIOLOGICAL DATA To gain insight into the response properties of the selection units in our model, we mapped their activations as a function of space and disparity. Specifically, we Selection 1 measured the activation of a unit as a single high-contrast edge was moved across the spatial extent of the receptive field. At each spatial location, we tested all possible disparities. An example of this mapping is shown in Figure 3. This selection unit is sensitive to changes in disparity as we move across space. We refer to this property as disparity contrast. In other words, the selection unit learned that a reliable indicator for a given disparity is a change in o 0.2 0.4 0.6 0.8 i disparity across space. This type of detector Figure 3: Selection unit activity can be behaviorally significant, because disparity contrast may playa role in signaling object boundaries. These selection units could thus provide valuable information in the construction of a 3-D model of the world. Recent neurophysiological studies by von der Heydt, Zhou, Friedman, and Poggio [8] is consistent with this interpretation. They found that neurons of awake, behaving monkeys in area V2 responded to edges of 4° by 4° random-dot stereograms. Because random-dot stereograms have no monocular form cues, these neurons must be responding to edges in depth. This sensitivity to edges in a depth map corresponds directly to the response profile of the selection units. 5 DISCUSSION A major difficulty in estimating the disparities of objects in a visual scene in realistic circumstances (i.e., with clutter, transparency, occlusion, noise) is knowing which cues are most reliable and should be integrated, and which regions have ambiguous or unreliable information. Nowlan and Sejnowski [6] found that selection units learned to respond strongly to image regions that contained motion energy in several different directions. The role of those selection units is similar to layered analysis techniques for computing support maps in the motion domain (Darrell & Pentland [1]). The operation of the dual pathways in our model bears some similar872 M. S. Gray, A. Pouget, R. S. Zemel, S. J. Nowlan and T. 1. Sejnowski ities to the pathways developed in the motion model of Nowlan and Sejnowski [6]. In the stereo domain, we have found that our selection units develop into edge detectors on a disparity map. They thus responded to regions rich in disparity information, analogous to the salient motion information captured in the motion [6] selection units. We have also found that the model matches psychophysical data recorded by Westheimer and McKee [9] on the effects of spatial frequency filtering on disparity thresholds (Gray et al [2]). They found, in human psychophysical experiments, that disparity thresholds increased for any kind of spatial frequency filtering of line targets. In particular, disparity sensitivity was more adversely affected by high-pass filtering than by low-pass filtering. In summary, we propose that the functional division into local response and selection represents a general principle for image interpretation and analysis that may be applicable to many different visual cues, and also to other sensory domains. In our approach to this problem, we utilized a multi-scale neurophysiologically-realistic implementation of binocular cells for the input, and then combined it with a neural network model to learn reliable cues for disparity estimation. References [1] T. Darrell and A.P. Pentland. Cooperative robust estimation using layers of support. IEEE Transactions on Pattern Analysis and Machine Intelligence, 17(5):474-87,1995. [2] M.S. Gray, A. Pouget, R.S. Zemel, S.J. Nowlan, and T.J. Sejnowski. Reliable disparity estimation through selective integration. INC Technical Report 9602, Institute for Neural Computation, University of California, San Diego, 1996. [3] D.M. Green and J.A. Swets. Signal Detection Theory and Psychophysics. John Wiley and Sons, New York, 1966. [4] R.A. Jacobs, M.I. Jordan, S.J. Nowlan, and G.E. Hinton. Adaptive mixtures of local experts. Neural Computation, 3:79-87, 1991. [5] R. Jain, R. Kasturi, and B.G. Schunck. Machine Vision. McGraw-Hill, New York, 1995. [6] S.J. Nowlan and T.J. Sejnowski. Filter selection model for motion segmentation and velocity integration. Journal of the Optical Society of America A, 11(12):3177-3200, 1994. [7] I. Ohzawa, G.C. DeAngelis, and R.D. Freeman. Stereoscopic depth discrimination in the visual cortex: Neurons ideally suited as disparity detectors. Science, 249:1037-1041,1990. [8] R. von der Heydt, H. Zhou, H. Friedman, and G.F. Poggio. Neurons of area V2 of visual cortex detect edges in random-dot stereograms. Soc. Neurosci. Abs., 21:18, 1995. [9] G. Westheimer and S.P. McKee. Stereoscopic acuity with defocus and spatially filtered retinal images. Journal of the Optical Society of America, 70:772-777, 1980.
1996
92
1,294
The Neurothermostat: Predictive Optimal Control of Residential Heating Systems Michael C. Mozert, Lucky Vidmart, Robert H. Dodiert tDepartment of Computer Science tDepartment of Civil, Environmental, and Architectural Engineering University of Colorado, Boulder, CO 80309-0430 Abstract The Neurothermostat is an adaptive controller that regulates indoor air temperature in a residence by switching a furnace on or off. The task is framed as an optimal control problem in which both comfort and energy costs are considered as part of the control objective. Because the consequences of control decisions are delayed in time, the N eurothermostat must anticipate heating demands with predictive models of occupancy patterns and the thermal response of the house and furnace. Occupancy pattern prediction is achieved by a hybrid neural net / look-up table. The Neurothermostat searches, at each discrete time step, for a decision sequence that minimizes the expected cost over a fixed planning horizon. The first decision in this sequence is taken, and this process repeats. Simulations of the Neurothermostat were conducted using artificial occupancy data in which regularity was systematically varied, as well as occupancy data from an actual residence. The Neurothermostat is compared against three conventional policies, and achieves reliably lower costs. This result is robust to the relative weighting of comfort and energy costs and the degree of variability in the occupancy patterns. For over a quarter century, the home automation industry has promised to revolutionize our lifestyle with the so-called Smart House@ in which appliances, lighting, stereo, video, and security systems are integrated under computer control. However, home automation has yet to make significant inroads, at least in part because software must be tailored to the home occupants. Instead of expecting the occupants to program their homes or to hire someone to do so, one would ideally like the home to essentially program itself by observing the lifestyle of the occupants. This is the goal of the Neural Network House (Mozer et al., 1995), an actual residence that has been outfitted with over 75 sensorsincluding temperature, light, sound, motion-and actua.tors to control air heating, water heating, lighting, and ventilation. In this paper, we describe one research 954 M. C. Mozer. L. Vidmar and R. H. Dodier project within the house, the Neurothermostat, that learns to regulate the indoor air temperature automatically by observing and detecting patterns in the occupants' schedules and comfort preferences. We focus on the problem of air heating with a whole-house furnace, but the same approach can be taken with alternative or multiple heating devices, and to the problems of cooling and ventilation. 1 TEMPERATURE REGULATION AS AN OPTIMAL CONTROL PROBLEM Traditionally, the control objective of air temperature regulation has been to minimize energy consumption while maintaining temperature within an acceptable comfort margin during certain times of the day and days of the week. This is sensible in commercial settings, where occupancy patterns follow simple rules and where energy considerations dominate individual preferences. In a residence, however, the desires and schedules of occupants need to be weighted equally with energy considerations. Consequently, we frame the task of air temperature regulation as a problem of maximizing occupant comfort and minimizing energy costs. These two objectives clearly conflict, but they can be integrated into a unified framework via an optimal control aproach in which the goal is to heat the house according to a policy that minimizes the cost J = lim 1 l:!~~"+ 1 [e( ut) + m( xt}], "-+00 " 0 where time, t, is quantized into nonoverlapping intervals during which we assume all environmental variables remain constant, to is the interval ending at the current time, Ut is the control decision for interval t (e.g., turn the furnace on), e(u) is the energy cost associated with decision u, Xt is the environmental state during interval t, which includes the indoor temperature and the occupancy status of the home, and m(x) is the misery of the occupant given state x. To add misery and energy costs, a common currency is required. Energy costs are readily expressed in dollars. We also determi'ne misery in dollars, as we describe later. While we have been unable to locate any earlier work that combined energy and comfort costs in an optimal control framework, optimal control has been used in a variety of building energy system control applications (e.g., Henze & Dodier, 1996; Khalid & Omatu, 1995), 2 THE NEUROTHERMOSTAT Figure 1 shows the system architecture of the Neurothermostat and its interaction with the environment. The heart of the Neurothermostat is a controller that, at time intervals of 6 minutes, can switch the house furnace on or off. Because the consequences of control decisions are delayed in time, the controller must be predictive to anticipate heating demands. The three boxes in the Figure depict components that predict or model various aspects of the environment. We explain their purpose via a formal description of the controller operation. The controller considers sequences of '" decisions, denoted u, and searches for the sequence that minimizes the expected total cost, i u , over the planning horizon of ",6 minutes: where the expectation is computed over future states of the environment conditional on the decision sequence u. The energy cost in an interval depends only on the control decision during that interval. The misery cost depends on two components The Neurothermostat 955 ,_ .r( . • ' ·. , .... . , . ._~ " .:J . ~ I;"; • EnvlrcJni;imt ... = -~':.~ Figure 1: The Neurothermostat and its interaction with the environment of the state-the house occupancy status, o(t) (0 for empty, 1 for occupied), and the indoor air temperature, hu(t): mu(o(t), hu(t» = p[o(t) = 1] m(l, hu(t» + p[o(t) = 0] m(O, hu(t» Because the true quantities e, hu, m, and p are unknown, they must be estimated. The house thermal model of Figure 1 provides e and hu, the occupant comfort cost model provides m, and the home occupancy predictor provides p. We follow a tradition of using neural networks for prediction in the context of building energy system control (e.g., Curtiss, Kreider, & Brandemuehl, 1993; Ferrano & Wong, 1990; Miller & Seem, 1991), although in our initial experiments we require a network only for the occupancy prediction. 2.1 PREDICTIVE OPTIMAL CONTROLLER We propose a closed-loop controller that combines prediction with fixed-horizon planning, of the sort proposed by Clarke, Mohtadi, and 'lUffs (1987). At each time step, the controller considers all possible decision sequences over the pl~nning horizon and selects the sequence that minimizes the expected total cost, J. The first decision in this sequence is then performed. After b minutes, the planning and execution process is repeated. This approach assumes that beyond the planning horizon, all costs are independent of the first decision in the sequence. While dynamic programming is an efficient search algorithm, it requires a discrete state space. Wishing to avoid quantizing the continuous variable of indoor temperature, and the errors that might be introduced, we performed performed exhaustive search through the possible decision sequences, which was tractable due to relatively short horizons and two additional domain constraints. First, because the house occupancy status can reasonably be assumed to be independent of the indoor temperature, p need not be recalculated for every possible decision sequence. Second, the current occupancy status depends on the recent occupancy history. Consequently, one needs to predict occupancy patterns over the planning horizon, o E {O, l}'" to compute p. However, because most occupancy sequences are highly improbable, we find that considering only the most likely sequences-those containing at most two occupancy state transitions-produces the same decisions as doing the search over the entire distribution, reducing the cost from 0(2"') to O(K;2). 2.2 OCCUPANCY PREDICTOR The basic task of the occupancy predictor is to estimate the probability that the occupant will be home b minutes in the future. The occupancy predictor can be run iteratively to estimate the probability of an occupancy pattern. If occupants follow a deterministic daily schedule, a look up table indexed by time of day and current occupancy state should capture occupancy patterns. We thus use a look up table to encode whatever structure possible, and a neural network 956 M. C. Mozer. L. Vidmar and R. H. Dodier to encode residual structure. The look up table divides time into fixed 6 minute bins. The neural network consisted of the following inputs: current time of day; day of the week; average proportion of time home was occupied in the 10, 20, and 30 minutes from the present time of day on the previous three days and on the same day of the week during the past four weeks; and the proportion of time the home was occupied during the past 60, 180, and 360 minutes. The network, a standard three-layer architecture, was trained by back propagation. The number of hidden units was chosen by cross validation. 2.3 THERMAL MODEL OF HOUSE AND FURNACE A thermal model of the house and furnace predicts future indoor temperature(s) as a function of the outdoor temperature and the furnace operation. While one could perform system identification using neural networks, a simple parameterized resistance-capacitance (RC) model provides a reasonable first-order approximation. The RC model assumes that: the inside of the house is at a uniform temperature, and likewise the outside; a homogeneous flat wall separates the inside and outside, and this wall has a thermal resistance R and thermal capacitance C; the entire wall mass is at the inside temperature; and the heat input to the inside is Q when the furnace is running or zero otherwise. Assuming that the outdoor temperature, denoted g, is constant, the the indoor temperature at time t, hu(t), is: hu(t) = hu(t - 1) exp( -606 / RC) + (RQu(t) + 9 )(1 - exp( -606 / RC)), where hu(to) is the actual indoor temperature at the current time. Rand C were determined by architectural properties of the Neural Network House to be 1.33 Kelvins/kilowatt and 16 megajoules/Kelvin, respectively. The House furnace is rated at 133,000 Btu/hour and has 92.5% fuel use efficiency, resulting in an output of Q = 36.1 kilowatts. With natural gas at $.485 per CCF, the cost of operating the furnace, e, is $.7135 per hour. 2.4 OCCUPANT COMFORT COST MODEL In the Neural Network House, the occupant expresses discomfort by adjusting a set point temperature on a control panel. However, for simplicity, we assume in this work the setpoint is a constant, A. When the home is occupied, the misery cost is a monotonic function of the deviation of the actual indoor temperature from the set point. When the home is empty, the misery cost is zero regardless of the temperature. There is a rich literature directed at measuring thermal comfort in a given environment (i.e., dry-bulb temperature, relative humidity, air velocity, c~othing insulation, etc.) for the average building occupant (e.g., Fanger, 1972; Gagge, Stolwijk, & Nishi, 1971). Although the measurements indicate the fraction of the population which is uncomfortable in a particular environment, one might also interpret them as a measure of an individual's level of discomfort. As a function of dry-bulb temperature, this curve is roughly parabolic. We approximate it with a measure of misery in a 6-minute period as follows: A ( h) _ _6_ max(0,1.x-hl-£)2 m 0, - oa 24 x 60 25 . The first term, 0, is a binary variable indicating the home occupancy state. The second term is a conversion factor from arbitrary "misery" units to dollars. The third term scales the misery cost from a full day to the basic update interval. The fourth term produces the parabolic relative misery function, scaled so that a temperature difference of 5° C produces one unit of misery, with a deadband region from A - ( to A + L We have chosen the conversion factor a using an economic perspective. Consider the lost productivity that results from trying to work in a home that is 5° C colder The Neurothermostat 957 than desired for a 24 hour period. Denote this loss p, measured in hours. The cost in dollars of this loss is then a = 'YP, where'Y is the individual's hourly salary. With this approch, a can be set in a natural, intuitive manner. 3 SIMULATION METHODOLOGY In all experiments we report below, fJ = 10 minutes, K, = 12 steps (120 minute planning horizon), ,\ = 22.5° C, f = 1, and 'Y = 28 dollars per hour. The productivity loss, p, was varied from 1 to 3 hours. We report here results from the Neurothermostat operating in a simulated environment, rather than in the actual Neural Network House. The simulated environment incorporates the house/furnace thermal model described earlier and occupants whose preferences follow the comfort cost model. The outdoor temperature is assumed to remain a constant 0° C. Thus, the Neurothermostat has an accurate model of its environment, except for the occupancy patterns, which it must predict based on training data. This allows us to evaluate the performance of the Neurothermostat and the occupancy model as occupancy patterns are varied, uncontaminated by the effect of inaccuracy in the other internal models. We have evaluated the Neurothermostat with both real and artificial occupancy data. The real data was collected from the Neural Network House with a single resident over an eight month period, using a simple algorithm based on motion detector output and the opening and closing of outside doors. The artificial data was generated by a simulation of a single occupant. The occupant would go to work each day, later on the weekends, would sometimes come home for lunch, sometimes go out on weekend nights, and sometimes go out of town for several days. To examine performance of the Neurothermostat as a function of the variability in the occupant's schedule, the simulation model included a parameter, the variability index. An index of 0 means that the schedule is entirely deterministic; an index of 1 means that the schedule was very noisy, but still contained statistical regularities. The index determined factors such as the likelihood and duration of out-of-town trips and the variability in departure and return times. 3.1 ALTERNATIVE HEATING POLICIES In addition to the Neurothermostat, we examined three nonadaptive control policies. These policies produce a setpoint at each time step, and the furnace is switched on if the temperature drops below the setpoint and off if the temperature rises above the setpoint. (We need not be concerned about damage to the furnace by cycling because control decisions are made ten minutes apart.) The constant-temperature policy produces a fixed setpoint of 22.5° C. The occupancy-triggered policy produces a set point of 18° C when the house is empty, 22.5° C when the house is occupied. The setback-thermostat policy lowers the setpoint from 22.5° C to 18° C half an hour before the mean work departure time for that day of the week, and raises it back to 22.5° C half an hour before the mean work return time for that day of the week. The setback temperature for the occupancy-triggered and setback-thermostat policies was determined empirically to minimize the total cost. 4 RESULTS 4.1 OCCUPANCY PREDICTOR Performance of three different predictors was evaluated using artificial data across a range of values for the variability index. For each condition, we generated eight training/test sets of artificial data, each training and test set consisting of 150 days of data. Table 1 shows the normalized mean squared error (MSE) on the test set, averaged over the eight replications. The normalization involved dividing the MSE for each replication by the error obtained by predicting the future occupancy state 958 M. C. Mozer, L. Vidmar and R. H. Dodier Table 1: Normalized MSE on Test Set for Occupancy Prediction-Artificial Data variability index 0 .25 .50 .75 1 lookup table .49 .81 .94 .92 .94 neural net .02 .63 .83 .86 .91 lookup table + neural net .02 .60 .78 .77 .74 Productivity Loss = 1.0 hr. Productivity Loss ,. 3.0 hr. 8.2 ~ 10 " ~ -,'*"- ./ ~ / 9~--- ___ , _ . _ . _ ./ O~ :; .- ~'"-"" ;"" ---i 8 --"." '0 " ~7.8 ~.'~~~ ... .. .. :..::::.. .... ~.:::..: :" .. ... .. . 115 r---_ 87.6 . ~.~ ____ _ c ." -----= 7.4 __ .-' . ., ... - . . ~ 7.2_ o 0.25 0.5 0.75 o 0.25 0.5 0.75 Variability Index Variability Index Figure 2: Mean cost per day incurred by four control policies on (artificial) test data as a function of the data's variability index for p = 1 (comfort lightly weighted, left panel) and p = 3 (comfort heavily weighted, right panel). is the same as the present state. The main result here is that the combination of neural network and look up table perform better than either component in isolation (ANOVA: F(l, 7) = 1121,p < .001 for combination vs. table; F(l, 7) = 64,p < .001 for combination vs. network), indicating that the two components are capturing different structure in the data. 4.2 CONTROLLER WITH ARTIFICIAL OCCUPANCY DATA Having trained eight occupancy predictors with different (artificial data) training sets, we computed misery and energy costs for the Neurothermostat on the corresponding test sets. Figure 2 shows the mean total cost per day as a function of variability index, control policy, and relative comfort cost. The robust result is that the Neurothermostat outperforms the three nonadaptive control policies for all levels of the variability index and for both a wide range of values of p. Other patterns in the data are noteworthy. Costs for the Neurothermostat tend to rise with the variability index, as one would expect because the occupant's schedule becomes less predictable. The constant-temperature policy is worst if occupant comfort is weighted lightly, and begins to approach the Neurothermostat in performance as comfort costs are increased. If comfort costs overwhelm energy costs, then the constant-temperature policy and the Neurothermostat converge. 4.3 CONTROLLER WITH REAL OCCUPANCY DATA Eight months of real occupancy data collected in the Neural Network House beginning in September 1994 was also used to generate occupancy models and test controllers. Three training/test splits were formed by training on five consecutive months and testing on the next month. Table 2 shows the mean daily cost for the four controllers. The Neurothermostat significantly outperforms the three nonadaptive controllers, as it did with the artificial data. 5 DISCUSSION The simulation studies reported here strongly suggest that adaptive control of residential heating and cooling systems is worthy of further investigation. One is The Neurothermostat 959 Table 2: Mean Daily Cost Based on Real Occupancy Data productivity loss p=1 p=3 Neurothermostat $6.77 $7.05 constant temperature $7.85 $7.85 occupancy triggered $7.49 $8.66 setback thermostat $8.12 $9.74 tempted to trumpet the conclusion that adaptive control lowers heating costs, but before doing so, one must be clear that the cost being lowered is a combination of comfort and energy costs. If one is merely interested .in lowering energy costs, then simply shut off the furnace. A central contribution of this work is thus the framing of the task of air temperature regulation as an optimal control problem in which both comfort and energy costs are considered as part of the control objective. A common reaction to this research project is, "My life is far too irregular to be predicted. I don't return home from work at the same time every day." An important finding of this work is that even a highly nondeterministic schedule contains sufficient statistical regularity to be exploited by a predictive controller. We found this for both artificial data with a high variability index and real occupancy data. A final contribution of our work is to show that for periodic data such as occupancy patterns that follow a weekly schedule, the combination of a look up table to encode the periodic structure and a neural network to encode the residual structure can outperform either method in isolation. Acknowledgements Support for this research has come from Lifestyle Technologies, NSF award IRI9058450, and a CRCW grant-in-aid from the University of Colorado. This project owes its existence to the dedication of many students, particularly Marc Anderson, Josh Anderson, Paul Kooros, and Charles Myers. Our thanks to Reid Hastie and Gary McClelland for their suggestions on assessing occupant misery. References Clarke, D. W., Mohtadi, C., & Tuffs, P. S. (1987). Generalized predictive control-Part I. The basic algorithm. Automatica, 29, 137- 148. Curtiss, P., Kreider, J. F., & Brandemuehl, M. J . (1993). Local and global control of commercial building HVAC systems using artificial neural networks. Proceedings of the American Control Conference, 9, 3029-3044. Fanger, P. O. (1972). Thermal comfort. New York: McGraw-Hill. Ferrano, F. J., & Wong, K. V. (1990). Prediction of thermal storage loads using a neural network. ASHRAE Transactions, 96, 723-726. Gagge, A. P., Stolwijk, J . A. J., & Nishi, Y. (1971). An effective temperature scale based on a simple model of human physiological regulatory response. ASHRAE Transactions, 77,247- 262. Henze, G. P., & Dodier, R. H. (1996). Development of a predictive optimal controller for thermal energy storage systems. Submitted for publication. Khalid, M., & Omatu, S. (1995). Temperature regulation with neural networks and alternative control schemes. IEEE Transactions on Neural Networks, 6, 572-582. Miller, R. C., & Seem, J. E. (1991). Comparison of artificial neural networks with traditional methods of predicting return time from night or weekend setback. ASHRAE Transactions, 97, 500- 508. Mozer, M. C., Dodier, R. H., Anderson, M., Vidmar, L., Cruickshank III, R. F., & Miller, D. (1995). The neural network house: An overview. In L. Niklasson & M. Boden (Eds.), Current trends in connectionism (pp. 371-380). Hillsdale, NJ: Erlbaum.
1996
93
1,295
Softening Discrete Relaxation Andrew M. Finch, Richard C. Wilson and Edwin R. Hancock Department of Computer Science, University of York, York, Y01 5DD, UK Abstract This paper describes a new framework for relational graph matching. The starting point is a recently reported Bayesian consistency measure which gauges structural differences using Hamming distance. The main contributions of the work are threefold. Firstly, we demonstrate how the discrete components of the cost function can be softened. The second contribution is to show how the softened cost function can be used to locate matches using continuous non-linear optimisation. Finally, we show how the resulting graph matching algorithm relates to the standard quadratic assignment problem. 1 Introduction Graph matching [6, 5, 7, 2, 3, 12, 11J is a topic of central importance in pattern perception. The main computational issues are how to compare inexact relational descriptions (7J and how to search efficiently for the best match [8J. These two issues have recently stimulated interest in the connectionist literature (9, 6, 5, lOJ. For instance, Simic [9], Suganathan et al. (101 and Gold et ai. [6, 51 have addressed the issue of how to expressively measure relational distance. Both Gold and Rangarajan (61 and Suganathan et al [101 have shown how non-linear optimisation techniques such as mean-field annealing [lOJ and graduated assignment [61 can be applied to find optimal matches. In a recent series of papers we have developed a Bayesian framework for relational graph matching [2, 3, 11, 121. The novelty resides in the fact that relational consistency is gauged by a probability distribution that uses Hamming distance to measure structural differences between the graphs under match. This new framework has not only been used to match complex infra-red (3J and radar imagery [11], it has also been used to successfully control a graph-edit process (12J of the sort originally proposed by Sanfeliu and Fu (71. The optimisation of this relational consistency measure has hitherto been confined to the use of discrete update procedures [11, 2, 3]. Examples include discrete relaxation [7, 11], simulated annealing Softening Discrete Relaxation 439 [4, 3] and genetic search [2]. Our aim in this paper is to consider how the optimisation of the relational consistency measure can be realised by continuous means [6, 10]. Specifically we consider how the matching process can be effected using a non-linear technique similar to mean-field annealing [IOJ or graduated assignment [6]. In order to achieve this goal we must transform our discrete cost function [11] into a form suitable for optimisation by continuous techniques. The key idea is to exploit the apparatus of statistical physics [13] to compute the effective Gibbs potentials for our discrete relaxation process. The potentials are in-fact weighted sums of Hamming distance enumerated over the consistent relations of the model graph. The quantities of interest in the optimisation process are the derivatives of the global energy function computed from the Gibbs potentials. In the case of our weighted sum of Hamming distance, these derivatives take on a particularly interesting form which provides an intuitive insight into the dynamics of the update process. An experimental evaluation of the technique reveals not only that it is successful in matching noise corrupted graphs, but that it significantly outperforms the optimisation of the standard quadratic energy function. 2 Relational Consistency Our overall goal in this paper is to formulate a non-linear optimisation technique for matching relational graphs. We use the notation G = (V, E) to denote the graphs under match, where V is the set of nodes and E is the set of edges. Our aim in matching is to associate nodes in a graph G D = (V D , ED) representing data to be matched against those in a graph G M = (V M , EM) representing an available relational model. Formally, the matching is represented by a function f : VD -T VM from the nodes in the data graph G D to those in the model graph G M. We represent the structure of the two graphs using a pair of connection matrices. The connection matrix for the data graph consists of the binary array { 1 if (a, b) E ED Dab = 0 otherwise while that for the model graph is { 1 if (a, (3) E EM MOl{3 = 0 otherwise (1) (2) The current state of match between the two graphs is represented by the function f : V D -T V M· In others words the statement f (a) = a means that the node a E V D is matched to the node a E V M. The binary representation of the current state of match is captured by a set of assignment variables which convey the following meaning _ {1 if f{a) = a Saa o otherwise (3) The basic goal of the matching process is to optimise a consistency-measure which gauges the structural similarity of the matched data graph and the model graph. In a recent series of papers, Wilson and Hancock [11, 12] have shown how consistency of match can be modelled using a Bayesian framework. The basic idea is to construct a probability distribution which models the effect of memoryless matching errors in generating departures from consistency between the data and model graphs. Suppose that Sa = aU {(3I(a, (3) E EM} represents the set of nodes that form the immediate contextual neighbourhood of the node a in the model graph. 440 A. M. Finch, R. C. Wilson and E. R. Hancock Further suppose that ra = f(a) U {f(b)l(a,b) E ED} represents the set of matches assigned to the contextual neighbourhood of the node a E VD of the data graph. Basic to Wilson and Hancock's modelling of relational consistency is to regard the complete set of model-graph relations as mutually exclusive causes from which the potentially corrupt matched model-graph relations arise. As a result, the probability of the matched configuration r a can be expressed as a mixture distribution over the corresponding space of model-graph configurations p(ra) = L p(raISa)P(Sa) (4) aEVM The modelling of the match confusion probabilities p(r alSa) draws on the assumption that the error process is independent of location. This allows p(raISa) to be factorised over its component matches. Individual label errors are further assumed to act with a memoryless probability Pe . With these ingredients the probability of the matched neighbourhood r a reduces to [11, 12] p(ra) = I~~I 2: exp[-ItH(a,a)] aEVM (5) where Ka = (1- Pe)lfal and the exponential constant is related to the probability of label errors, i.e. It = In (l-;,~e ). Consistency of match is gauged by the "Hamming distance", H(a, a) between the matched relation r a and the set of consistent neighbourhood structures Sa, 'Va E VM from the model graph. According to our binary representation of the matching process, the distance measure is computed using the connectivity matrices and the assignment variables in the following manner H(a, a) = 2: 2: Ma{3Dab(l - Sb{3) (6) bEVD {3EVM The probability distribution p(r a) may be regarded as providing a natural way of modelling departures from consistency at the neighbourhood level. Matching consistency is graded by Hamming distance and controlled hardening may be induced by reducing the label-error probability Pe towards zero. 3 The Effective Potential for Discrete Relaxation We commence the development of our graduated assignment approach to discrete relaxation by computing an effective Gibbs potential U(r a) for the matching configuration r a. In other words, we aim to replace the compound exponential probability distribution appearing in equation (5) by the single Gibbs distribution (7) Our route to the effective potential is provided by statistical physics. If we represent p(r a) by an equivalent Gibbs distribution with an identical partition function, then the equilibrium configurational potential is related to the partial derivative of the log-probability with respect to the coupling constant It in the following manner [13] Softening Discrete Relaxation u(r a) = _ 8ln p(r a) 8J.t Upon substituting for p(r a) from equation (5) 2: H(a, a) exp[ -J.tH(a, a)] u(ra) = _a_E~VM~ ________________ __ 2: exp[-J.tH(a,a)] aEVM 441 (8) (9) In other words the neighbourhood Gibbs potentials are simply weighted sums of Hamming distance between the data and model graphs. In fact the local clique potentials display an interesting barrier property. The potential is concentrated at Hamming distance H ~ ~. Both very large and very small Hamming distances contribute insignificantly to the energy function, i.e. limH-to H exp[-J.tH] = 0 and limH-too H exp[-J.tH] = o. With the neighbourhood matching potentials to hand, we construct a global "matching-energy" [; = 2:aEVD U(r a) by summing the contributions over the nodes of the data graph. 4 Optimising the Global Cost Function We are now in a position to develop a continuous update algorithm by softening the discrete ingredients of our graph matching potential. The idea is to compute the derivatives of the global energy given in equation (10) and to effect the softening process using the soft-max idea of Bridle [1]. 4.1 Softassign The energy function represented by equations (9) and (10) is defined over the discrete matching variables Saa. The basic idea underpinning this paper is to realise a continuous process for updating the assignment variables. The optimal step-size is determined by computing the partial derivatives of the global matching energy with respect to the assignment variables. We commence by computing the derivatives of the contributing neighbourhood Gibbs potentials, i.e. where ~aa = exp(-J.tH(a, a)] 2:aIEVM exp[-J.tH(a, a l )] (11) To further develop this result, we must compute the derivatives of the Hamming distances. From equation (6) it follows that 8H(a,a) _ M D 8 - a{3 ab Sb{3 (12) It is now a straightforward matter to show that the derivative of the global matching energy is equal to 442 A. M. Finch, R. C. Wilson and E. R. Hancock We would like our continuous matching vanables to remain constrained to lie within the range [0, 1]. Rather than using a linear update rule, we exploit Bridle's soft-max ansatz [1). In doing this we arrive at an update process which has many features in common with the well-known mean-field equations of statistical physics exp[-~~] T OSaa Sao. +- -----'::.........,[;:---:--0-3-:"""""""""] L exp -~_£ T OSaa' a'EVM (14) The mathematical structure of this update process is important and deserves further comment. The quantity eaa defined in equation (11) naturally plays the role of a matching probability. The first term appearing under the square bracket in equation (13) can therefore be thought of as analogous to the optimal update direction for the standard quadratic cost function [10,6); we will discus this relationship in more detail in Section 4.2. The second term modifies this principal update direction by taking into account the weighted fluctuations in the Hamming distance about the effective potential or average Hamming distance. If the average fluctuation is zero, then there is no net modification to the update direction. When the net fluctuation is non-zero, the direction of update is modified so as to compensate for the movement of the mean-value of the effective potential. This corrective tracking process provides an explicit mechanism for maintaining contact with the minimum of the effective potential under rescaling effects induced by changes in the value of the coupling constant p. Moreover, since the fluctuation term is itself proportional to p, this has an insignificant effect for Pe ~ ~ but dominates the update process when Pe -+ 0. 4.2 Quadratic Assignment Problem Before we proceed to experiment with the new graph matching process, it is interesting to briefly review the standard quadratic formulation of the matching problem investigated by Simic (9], Suganathan et al (to] and Gold and Rangarajan (6]. The common feature of these algorithms is to commence from the quadratic cost function (15) In this case the derivative of the global cost function is linear in the assignment variables, i.e. (16) This step size is equivalent to that appearing in equation (14) provided that p = 0, i.e. Pe -+ !. The update is realised by applying the soft-max ansatz of equation (14). In the next section, we will provide some experimental comparison with the resulting matching process. However, it is important to stress that the update process adopted here is very simplistic and leaves considerable scope for further refinement. For instance, Gold and Rangarajan (6] have exploited the doubly stochastic properties of Sinckhorn matrices to ensure two-way symmetry in the matching process. Softening Discrete Relaxation 443 5 Experiments and Conclusions Our main aim in this Section is to compare the non-linear update equations with the optimisation of the quadratic matching criterion described in Section 4.2. The data for our study is provided by synthetic Delaunay graphs. These graphs are constructed by generating random dot patterns. Each random dot is used to seed a Voronoi cell. The Delaunay triangulation is the region adjacency graph for the Voronoi cells. In order to pose demanding tests of our matching technique, we have added controlled amounts of corruption to the synthetic graphs. This is effected by deleting and adding a specified fraction of the dots from the initial random patterns. The associated Delaunay graph is therefore subject to structural corruption. We measure the degree of corruption by the fraction of surviving nodes in the corrupted Delaunay graph. Our experimental protocol has been as follows. For a series of different corruption levels, we have generated a sample of 100 random graphs. The graphs contain 50 nodes each. According to the specified corruption level, we have both added and deleted a predefined fraction of nodes at random locations in the initial graphs so as to maintain their overall size. For each graph we measure the quality of match by computing the fraction of the surviving nodes for which the assignment variables indicate the correct match. The value of the temperature T in the update process has been controlled using a logarithmic annealing schedule of the form suggested by Geman and Geman (41. We initialise the assignment variables uniformly across the set of matches by setting Saa = JM , "ta, 0:. We have compared the results obtained with two different versions of the matching algorithm. The first of these involves updating the softened assignment variables by applying the non-linear update equation given in (14). The second matching algorithm involves applying the same optimisation apparatus to the quadratic cost function defined in equation (15) in a simplified form of the quadratic assignment algorithm [6, 101. Figure 1 shows the final fraction of correct matches for both algorithms. The data curves show the correct matching fraction averaged over the graph samples as a function of the corruption fraction. The main conclusions that can be drawn from these plots is that the new matching technique described in this paper significantly outperforms its conventional quadratic counterpart described in Section 4.2. The main difference between the two techniques resides in the fact that our new method relies on updating with derivatives of the energy function that are non-linear in the assignment variables. To conclude, our main contribution in this paper has been to demonstrate how the discrete Bayesian relational consistency measure of Wilson and Hancock (111 can be cast in a form that is amenable to continuous non-linear optimisation. We have shown how the method relates to the standard quadratic assignment algorithm extensively studied in the connectionist literature [6, 9, 101. Moreover, an experimental analysis reveals that the method offers superior performance in terms of noise control. References [1] Bridle J.S. "Training stochastic model recognition algorithms can lead to maximum mutual information estimation of parameters" NIPS2, pp. 211-217, 1990. 444 ~' .. -.. "", 0.8 c: 0 '13 til u: 0.6 i 0 0.4 () iii c: u:: 0.2 0 0 A. M. Finch, R. C. Wilson and E. R. Hancock ..... Quadratic Assignment -···~· ....... ~oftened Discrete Relaxation ..... . ~ ", ". '.' •.. ~.--............... " •.. ' ...... .,.~---- ---i; \ \ \ \ '. \ .... 0.2 0.4 0.6 0.8 Fraction of Graph Corrupt Figure 1: Experimental comparison: softened discrete relaxation (dotted curve); matching using the quadratic cost function (solid curve). [2] Cross A.D.J., RC.Wilson and E.R Hancock, "Genetic search for structural matching", Proceedings ECCV96, LNCS 1064, pp. 514-525, 1996. [3] Cross A.D.J . and E.RHancock, "Relational matching with stochastic optimisation" IEEE Computer Society International Symposium on Computer Vision, pp. 365-370, 1995. [4] Geman S. and D. Geman, "Stochastic relaxation, Gibbs distributions and Bayesian restoration of images," IEEE PAMI, PAMI-6 , pp.721- 741, 1984. [5] Gold S., A. Rangarajan and E. Mjolsness, "Learning with pre-knowledge: Clustering with point and graph-matching distance measures", Neural Computation, 8, pp. 787804, 1996. [6] Gold S. and A. Rangarajan, "A graduated assignment algorithm for graph matching", IEEE PAMI, 18, pp. 377-388, 1996. [7] Sanfeliu A. and Fu K.S., "A distance measure between attributed relational graphs for pattern recognition", IEEE SMC, 13, pp 353-362, 1983. [8] Shapiro L. and RM.Haralick, "Structural description and inexact matching", IEEE PAM!, 3, pp 504-519, 1981. [9] Simic P., "Constrained nets for graph matching and other quadratic assignment problems", Neural Computation, 3 , pp. 268- 281, 1991. [10] Suganathan P.N., E.K. Teoh and D.P. Mital, "Pattern recognition by graph matching using Potts MFT networks", Pattern Recognition, 28, pp. 997-1009, 1995. [11] Wilson RC., Evans A.N. and Hancock E.R, "Relational matching by discrete relaxation", Image and Vision Computing, 13, pp. 411-421, 1995. [12] Wilson RC and Hancock E.R, "Relational matching with dynamic graph structures" , Proceedings of the Fifth International Conference on Computer Vision, pp. 450-456, 1995. [13] Yuille A., "Generalised deformable models, statistical physics and matching problems", Neural Computation, 2, pp. 1-24, 1990.
1996
94
1,296
A comparison between neural networks and other statistical techniques for modeling the relationship between tobacco and alcohol and cancer Tony Plate BC Cancer Agency 601 West 10th Ave, Epidemiology Vancouver BC Canada V5Z 1L3 tap@comp.vuw.ac.nz Joel Bert Dept of Chemical Engineering University of British Columbia 2216 Main Mall Vancouver BC Canada V6T 1Z4 Pierre Band BC Cancer Agency 601 West 10th Ave, Epidemiology Vancouver BC Canada V5Z 1L3 John Grace Dept of Chemical Engineering University of British Columbia 2216 Main Mall Vancouver BC Canada V6T 1Z4 Abstract Epidemiological data is traditionally analyzed with very simple techniques. Flexible models, such as neural networks, have the potential to discover unanticipated features in the data. However, to be useful, flexible models must have effective control on overfitting. This paper reports on a comparative study of the predictive quality of neural networks and other flexible models applied to real and artificial epidemiological data. The results suggest that there are no major unanticipated complex features in the real data, and also demonstrate that MacKay's [1995] Bayesian neural network methodology provides effective control on overfitting while retaining the ability to discover complex features in the artificial data. 1 Introduction Traditionally, very simple statistical techniques are used in the analysis of epidemiological studies. The predominant technique is logistic regression, in which the effects of predictors are linear (or categorical) and additive on the log-odds scale. An important virtue of logistic regression is that the relationships identified in the 968 T. Plate, P. Band, J. Bert and J. Grace data can be interpreted and explained in simple terms, such as "the odds of developing lung cancer for males who smoke between 20 and 29 cigarettes per day are increased by a factor of 11.5 over males who do not smoke". However, because of their simplicity, it is difficult to use these models to discover unanticipated complex relationships, i.e., non-linearities in the effect of a predictor or interactions between predictors. Interactions and non-linearities can of course be introduced into logistic regressions, but must be pre-specified, which tends to be impractical unless there are only a few variables or there are a priori reasons to test for particular effects. Neural networks have the potential to automatically discover complex relationships. There has been much interest in using neural networks in biomedical applications; witness the recent series of articles in The Lancet, e.g., Wyatt [1995] and Baxt [1995]. However, there are not yet sufficient comparisons or theory to come to firm conclusions about the utility of neural networks in biomedical data analysis. To date, comparison studies, e.g, those by Michie, Spiegelhalter, and Taylor [1994], Burke, Rosen, and Goodman [1995]' and Lippmann, Lee, and Shahian [1995], have had mixed results, and Jefferson et aI's [1995] complaint that many "successful" applications of neural networks are not compared against standard techniques appears to be justified. The intent of this paper is to contribute to the body of useful comparisons by reporting a study of various neural-network and statistical modeling techniques applied to an epidemiological data analysis problem. 2 The data The original data set consisted of information on 15,463 subjects from a study conducted by the Division of Epidemiology and Cancer Prevention at the BC Cancer Agency. In this study, detailed questionnaire reported personal information, lifetime tobacco and alcohol use, and lifetime employment history for each subject. The subjects were cancer patients in BC with diagnosis dates between 1983 and 1989, as ascertained by the population-based registry at the BC Cancer Agency. Six different tobacco and alcohol habits were included: cigarette (c), cigar (G), and pipe (p) smoking, and beer (B), wine (w), and spirit drinking (s). The models reported in this paper used up to 27 predictor variables: age at first diagnosis (AGE), and 26 variables related to alcohol and tobacco consumption. These included four variables for each habit: total years of consumption (CYR etc), consumption per day or week (CDAY, BWK etc), years since quitting (CYQUIT etc), and a binary variable indicating any indulgence (CSMOKE, BDRINK etc). The remaining two binary variables indicated whether the subject ever smoked tobacco or drank alcohol. All the binary variables were non-linear (threshold) transforms of the other variables. Variables not applicable to a particular subject were zero, e.g., number of years of smoking for a non-smoker, or years since quitting for a smoker who did not quit. Of the 15,463 records, 5901 had missing information in some of the fields related to tobacco or alcohol use. These were not used, as there are no simple methods for dealing with missing data in neural networks. Of the 9,562 complete records, a randomly selected 3,195 were set aside for testing, leaving 6,367 complete records to be used in the modeling experiments. There were 28 binary outcomes: the 28 sites at which a subject could have cancer (subjects had cancers at up to 3 different sites). The number of cases for each site varied, e.g., for LUNGSQ (Lung Squamous) there were 694 cases among the complete records, for ORAL (Oral Cavity and Pharynx) 306, and for MEL (Melanoma) 464. All sites were modeled individually using carefully selected subjects as controls. This is common practice in cancer epidemiology studies, due to the difficulty of collecting an unbiased sample of non-cancer subjects for controls. Subjects with Neural Networks in Cancer Epidemiology 969 cancers at a site suspected of being related to tobacco usage were not used as controls. This eliminated subjects with any sites other than COLON, RECTUM, MEL (Melanoma), NMSK (Non-melanoma skin), PROS (Prostate), NHL (Non-Hodgkin's lymphoma), and MMY (Multiple-Myeloma), and resulted in between 2959 and 3694 controls for each site. For example, the model for LUNGSQ (lung squamous cell) cancer was fitted using subjects with LUNGSQ as the positive outcomes (694 cases), and subjects all of whose sites were among COLON, RECTUM, MEL, NMSK, PROS, NHL, or MMY as negative outcomes (3694 controls). 3 Statistical methods A number of different types of statistical methods were used to model the data. These ranged from the non-flexible (logistic regression) through partially flexible (Generalized Additive Models or GAMs) to completely flexible (classification trees and neural networks). Each site was modeled independently, using the log likelihood of the data under the binomial distribution as the fitting criterion. All of the modeling, except for the neural networks and ridge regression, was done using the the S-plus statistical software package [StatSci 1995]. For several methods, we used Breiman's [1996] bagging technique to control overfitting. To "bag" a model, one fits a set of models independently on bootstrap samples. The bagged prediction is then the average of the predictions of the models in the set. Breiman suggests that bagging will give superior predictions for unstable models (such as stepwise selection, pruned trees, and neural networks). Preliminary analysis revealed that the predictive power of non-flexible models could be improved by including non-linear transforms of some variables, namely AGESQ and the binary indicator variables SMOKE, DRINK, CSMOKE, etc. Flexible models should be able to discover useful non-linear transforms for themselves and so these derived variables were not included in the flexible models. In order to allow comparisons to test this, one of non-flexible models (ONLYLIN-STEP) also did not use any of these derived variables. Null model: (NULL) The predictions of the null model are just the frequency of the outcome in the training set. Logistic regression: The FULL model used the full set of predictor variables, including a quadratic term for age: AGESQ. Stepwise logistic regression: A number of stepwise regressions were fitted, differing in the set of variables considered. Outcome-balanced lO-fold cross validation was used to select the model size giving best generalization. The models were as follows: AGE-STEP (AGE and AGESQ); CYR-AGE-STEP (CYR, AGE and AGESQ)j ALCCYR-AGE-STEP (all alcohol variables, CYR, AGE and AGESQ); FULL-STEP (all variables including AGESQ); and ONLYLIN-STEP (all variables except for the derived binary indicator variables SMOKE, CSMOKE, etc, and only a linear AGE term). Ridge regression: (RIDGE) Ridge regression penalizes a logistic regression model by the sum of the squared parameter values in order to control overfitting. The evidence framework [MacKay 1995] was used to select seven shrinkage parameters: one for each of the six habits, and one for SMOKE, DRINK, AGE and AGESQ. Generalized Additive Models: GAMs [Hastie and Tibshirani 1990] fit a smoothing spline to each parameter. GAMs can model non-linearities, but not interactions. A stepwise procedure was used to select the degree (0,1,2, or 4) of the smoothing spline for each parameter. The procedure started with a model having a smoothing spline of degree 2 for each parameter, and stopped when the Ale statistic could 970 T. Plate, P. Band, 1. Bert and 1. Grace not reduced any further. Two stepwise GAM models were fitted: GAM-FULL used the full set of variables, while GAM-CIG used the cigarette variables and AGE. Classification trees: [Breiman et al. 1984] The same cross-validation procedure as used with stepwise regression was used to select the best size for TREE, using the implementation in S-plus, and the function shrink.treeO for pruning. A bagged version with 50 replications, TREE-BAGGED, was also used. After constructing a tree for the data in a replication, it was pruned to perform optimally on the training data not included in that replication. Ordinary neural networks: The neural network models had a single hidden layer of tanh functions and a small weight penalty (0.01) to prevent parameters going to infinity. A conjugate-gradient procedure was used to optimize weights. For the NNORD-H2 model, which had no control on complexity, a network with two hidden units was trained three times from different small random starting weights. Of these three, the one with best performance on the training data was selected as "the model". The NN-ORD-HCV used common method for controlling overfitting in neural networks: 10fold CV for selecting the optimal number of hidden units. Three random starting points for each partition were used calculate the average generalization error for networks with one, two and three hidden units Three networks with the best number of hidden units were trained on the entire set of training data, and the network having the lowest training error was chosen. Bagged neural networks with early stopping: Bagging and early stopping (terminating training before reaching a minimum on training set error in order to prevent overfitting) work naturally together. The training examples omitted from each bootstrap replication provide a validation set to decide when to stop, and with early stopping, training is fast enough to make bagging practical. 100 networks with two hidden units were trained on separate bootstrap replications, and the best 50 (by their performance on the omitted examples) were included in the final bagged model, NN-ESTOP-BAGGED. For comparison purposes, the mean individual performance of these early-stopped networks is reported as NN-ESTOP-AVG. N eur~l networks with Bayesian regularization: MacKay's [1995] Bayesian evidence framework was used to control overfitting in neural networks. Three random starts for networks with 1, 2, 3 or 4 hidden units and three different sets of regularization (penalty) parameters were used, giving a total of 36 networks for each site. The three possibilities for regularization parameters were: (a) three penalty parameters - one for each of input to hidden, bias to hidden, and hidden to output; (b) partial Automatic Relevance Determination (ARD) [MacKay 1995] with seven penalty parameters controlling the input to hidden weights - one for each habit and one for AGE; and (c) full ARD, with one penalty parameter for each of the 19 inputs. The "evidence" for each network was evaluated and the best 18 networks were selected for the equally-weighted committee model NN-BAYES-CMTT. NN-BAYES-BEST was the single network with the maximum evidence. 4 Results and Discussion Models were compared based on their performance on the held-out test data, so as to avoid overfitting bias in evaluation. While there are several ways to measure performance, e.g., 0-1 classification error, or area under the ROC curve (as in Burke, Rosen and Goodman [1995]), we used the test-set deviance as it seems appropriate to compare models using the same criterion as was used for fitting. Reporting performance is complicated by the fact that there were 28 different modeling tasks (Le., sites), and some models did better on some sites and worse on others. We report some overall performance figures and some pairwise comparisons of models. Neural Networks in Cancer Epidemiology NULL AGE-STEP CYR-AGE-STEP ALC-CYR-AGE-STEP ONLYLIN-STEP FULL-STEP FULL RIDGE GAM-CIG GAM-FULL TREE TREE-BAGGED NN-ORD-H2 NN-ORD-HCV NN-ESTOP-AVG NN-ESTOP-BAGGED NN -BAYES-BEST NN-BAYES-CMTT -10 ALL 6618/3299 • • • • • • • • • • • • • • • • 0 10 20 -10 0 10 20 • -10 971 -10 0 10 20 MEL 322/142 r • • • • • • • • • • • • • • • • • • • • • • • :. • • • • • • • • • 0 10 20 Figure 1: Percent improvement in deviance on test data over the null model. Figure 1 shows aggregate deviances across sites (Le., the sum of the test deviance for one model over the 28 sites) and deviances for selected sites. The horizontal scale in each column indicates the percentage reduction in deviance over the null model. Zero percent (the dotted line) is the same performance as the null model, and 100% would be perfect predictions. Numbers below the column labels are the number of positive outcomes in the training and test sets, respectively. The best predictions for LUNGSQ can reduce the null deviance by just over 25%. It is interesting to note that much of the information is contained in AGE and CYR: The CYR-AGE-STEP model achieved a 7.1% reduction in overall deviance, while the maximum reduction (achieved by NN-BAYES-CMTT) was only 8.3%. There is no single threshold at which differences in test-set deviance are "significant", because of strong correlations between predictions of different models. However, the general patterns of superiority apparent in Figure 1 were repeated across the other sites, and various other tests indicate they are reliable indicators of general performance. For example, the best five models, both in terms of aggregate deviance across all sites and median rank of performance on individual sites, were, in order NN-BAYES-CMTT, RIDGE, NN-ESTOP-BAGGED, GAM-CIG, and FULL-STEP. The ONLYLIN-STEP model ranked sixth in median rank, and tenth in aggregate deviance. Although the differences between the best flexible models and the logistic models were slight, they were consistent. For example, NN-BAYES-CMTT did better than FULL-STEP on 21 sites, and better than ONLYLIN-STEP on 23 sites, while FULL-STEP drew with ONLYLIN-STEP on 14 sites and did better on 9. If the models had no effective difference, there was only a 1.25%. chance of one model doing better than the other 21 or more times out of 28. Individual measures of performance were also consistent with these findings. For example, for LUNGSQ a bootstrap test of test-set deviance revealed that the predictions of NN-BA YES-CMTT were on average better than those of ONLYLIN-STEP in 99.82% of res amp led test sets (out of 10,000), while the predictions of NN-BAYES-CMTT beat FULL-STEP in 93.75% of replications 972 T. Plate, P. Band, J. Bert and J. Grace and FULL-STEP beat ONLYLIN-STEP in 98.48% of replications. These results demonstrate that good control on overfitting is essential for this task. Ordinary neural networks with no control on overfitting do worse than guessing (i.e., the null model). Even when the number of hidden units is chosen by cross-validation, the performance is still worse than a simple two-variable stepwise logistic regression (CYR-AGE-STEP). The inadequacy of the simple Ale-based stepwise procedure for choosing the complexity of G AMs is illustrated by the poor performance of the GAM-FULL model (the more restricted GAM-CIG model does quite well). The effective methods for controlling overfitting were bagging and Bayesian regularization. Bagging improved the performance of trees and early-stopped neural networks to good levels. Bayesian regularization worked very well with neural networks and with ridge regression. Furthermore, examination of the performance of individual networks indicates that networks with fine-grained ARD were frequently superior to those with coarser control on regularization. 5 Artificial sites with complex relationships The very minor improvement achieved by neural networks and trees over logistic models provokes the following question: are complex relationships are really relatively unimportant in this data, or is the strong control on overfitting preventing identification of complex relationships? In order to answer this question, we created six artificial "sites" for the subjects. These were designed to have very similar properties to the real sites, while possessing non-linear effects and interactions. The risk models for the artificial sites possessed a underlying trend equal to half that of a good logistic model for LUNGSQ, and one of three more complex effects: FREQ, a frequent non-linear (threshold) effect (BWK > 1) affecting 4,334 of the 9,562 subjects; RARE, a rare threshold effect (BWK > 10), affecting 1,550 subjects; and INTER, an interaction (BYR . GYR) affecting 482 subjects. For three of the artificial sites the complex effect was weak (LO), and for the other three it was strong (HI). For each subject and each artificial site, a random choice as to whether that subject was a positive case for that site was made, based on probability given by the model for the artificial site. Models were fitted to these sites in the same way as to other sites and only subjects without cancer at a smoking related site were used as controls. NULL FREQ-TRUE RARE-TRUE INTER-TRUE ONLYLIN-STEP FULL-STEP TREE-BAGGED NN-ESTOP-BAGGED NN-BA YES-CMTT FREQ-LO 263;128 • • • • • • • • • 0204060 o 20 40 60 o 20 40 60 o 20 40 60 FREQ-HI RA~f-LO RARE-HI INTER-LO INTER-HI 440/2lO 402 218 482/253 245/115 564/274 ~ ~ ~ ~ ~ • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • 0204060 0204060 Figure 2: Percent improvement in deviance on test data for the artificial sites. For comparison purposes, logistic models containing the true set of variables, including non-linearities and interactions, were fitted to the artificial data. For example, the model RARE-TRUE contained the continuous variables AGE, AGESQ, CDAY, CYR, and CYQUIT, and the binary variables SMOKE and BWK> 10-, Neural Networks in Cancer Epidemiology 973 Figure 2 shows performance on the artificial data. The neural networks and bagged trees were very effective at detecting non-linearities and interactions. Their performance was at the same level as the appropriate true models, while the performance of simple models lacking the ability to fit the complexities (e.g., FULL-STEP) was considerably worse. 6 Conclusions For predicting the risk of cancer in our data, neural networks with Bayesian estimation of regularization parameters to control overfitting performed consistently but only slightly better than logistic regression models. This appeared to be due to the lack of complex relationships in the data: on artificial data with complex relationships they performed markedly better than logistic · models. Good control of overfitting is essential for this task, as shown by the poor performance of neural networks with the number of hidden units chosen by cross-validation. Given their ability to not overfit while still identifying complex relationships we expect that neural networks could prove useful in epidemiological data-analysis by providing a method for checking that a simple statistical model is not missing important complex relationships. Acknowledgments This research was funded by grants from the Workers Compensation Board of British Columbia, NSERC, and IRIS, and conducted at the BC Cancer Agency. References Baxt, w. G. 1995. Application of artificial neural networks to clinical medicine. The Lancet, 346:1135-1138. Breiman, L. 1996. Bagging predictors. Machine Learning, 26(2):123- 140. Breiman, L., Friedman, J ., Olshen, R. , and Stone, C. 1984. Classification and Regression Trees. Wadsworth, Belmont, CA. Burke, H. , Rosen, D., and Goodman, P. 1995. Comparing the prediction accuracy of artificial neural networks and other statistical methods for breast cancer survival. In Tesauro, G., Touretzky, D. S., and Leen, T. K., editors, Advances in Neural Information Processing Systems 7, pages 1063- 1067, Cambridge, MA. MIT Press. Hastie, T. J. and Tibshirani, R. J . 1990. Generalized additive models. Chapman and Hall, London. Jefferson, M. F., Pendleton, N., Lucas, S., and Horan, M. A. 1995. Neural networks (letter). The Lancet, 346:1712. Lippmann, R., Lee, Y., and Shahian, D. 1995. Predicting the risk of complications in coronary artery bypass operations using neural networks. In Tesauro, G., Touretzky, D. S., and Leen, T . K., editors, Advances in Neural Information Processing Systems 7, pages 1055-1062, Cambridge, MA. MIT Press. MacKay, D. J. C. 1995. Probable networks and plausible predictions - a review of practical Bayesian methods for supervised neural networks. Network: Computation in Neural Systems, 6:469- 505. Michie, D., Spiegelhalter, D., and Taylor, C. 1994. Machine Learning, Neural and Statistical Classification. Ellis Horwood, Hertfordshire, UK. StatSci 1995. S-Plus Guide to Statistical and Mathematical Analyses, Version 3.3. StatSci, a division of MathSoft, Inc, Seattle. Wyatt, J. 1995. Nervous about artificial neural networks? (commentary). The Lancet, 346:1175-1177.
1996
95
1,297
An Apobayesian Relative of Winnow Nick Littlestone NEC Research Institute 4 Independence Way Princeton, NJ 08540 Abstract Chris Mesterharm NEC Research Institute 4 Independence Way Princeton, NJ 08540 We study a mistake-driven variant of an on-line Bayesian learning algorithm (similar to one studied by Cesa-Bianchi, Helmbold, and Panizza [CHP96]). This variant only updates its state (learns) on trials in which it makes a mistake. The algorithm makes binary classifications using a linear-threshold classifier and runs in time linear in the number of attributes seen by the learner. We have been able to show, theoretically and in simulations, that this algorithm performs well under assumptions quite different from those embodied in the prior of the original Bayesian algorithm. It can handle situations that we do not know how to handle in linear time with Bayesian algorithms. We expect our techniques to be useful in deriving and analyzing other apobayesian algorithms. 1 Introduction We consider two styles of on-line learning. In both cases, learning proceeds in a sequence of trials. In each trial, a learner observes an instance to be classified, makes a prediction of its classification, and then observes a label that gives the correct classification. One style of on-line learning that we consider is Bayesian. The learner uses probabilistic assumptions about the world (embodied in a prior over some model class) and data observed in past trials to construct a probabilistic model (embodied in a posterior distribution over the model class). The learner uses this model to make a prediction in the current trial. When the learner is told the correct classification of the instance, the learner uses this information to update the model, generating a new posterior to be used in the next trial. In the other style of learning that we consider, the attention is on the correctness of the predictions rather than on the model of the world. The internal state of the An Apobayesian Relative o!Winnow 205 learner is only changed when the learner makes a mistake (when the prediction fails to match the label). We call such an algorithm mistake-driven. (Such algorithms are often called conservative in the computational learning theory literature.) There is a simple way to derive a mistake-driven algorithm from anyon-line learning algorithm (we restrict our attention in this paper to deterministic algorithms). The derived algorithm is just like the original algorithm, except that before every trial, it makes a record of its entire state, and after every trial in which its prediction is correct, it resets its state to match the recorded state, entirely forgetting the intervening trial. (Typically this is actually implemented not by making such a record, but by merely omitting the step that updates the state.) For example, if some algorithm keeps track of the number of trials it has seen, then the mistake-driven version of this algorithm will end up keeping track of the number of mistakes it has made. Whether the original or mistake-driven algorithm will do better depends on the task and on how the algorithms are evaluated. We will start with a Bayesian learning algorithm that we call SBSB and use this procedure to derive a mistake-driven variant, SASB. Note that the variant cannot be expected to be a Bayesian learning algorithm (at least in the ordinary sense) since a Bayesian algorithm would make a prediction that minimizes the Bayes risk based on all the available data, and the mistake-driven variant has forgotten quite a bit. We call such algorithms apobayesian learning algorithms. This name is intended to suggest that they are derived from Bayesian learning algorithms, but are not themselves Bayesian. Our algorithm SASB is very close to an algorithm of [CHP96). We study its application to different tasks than they do, analyzing its performance when it is applied to linearly separable data as described below. In this paper instances will be chosen from the instance space X = {a, l}n for some n. Thus instances are composed of n boolean attributes. We consider only two category classifications tasks, with predictions and labels chosen from Y = {a, I}. We obtain a' bound on the number of mistakes SASB makes that is comparable to bounds for various Winnow family algorithms given in [Lit88,Lit89). As for those algorithms, the bound holds under the assumption that the points labeled 1 are linearly separable from the points labeled 0, and the bound depends on the size 8 of the gap between the two classes. (See Section 3 for a definition of 8.) The mistake bound for SASB is 0 ( /or log ~ ). While this bound has an extra factor of log ~ not present in the bounds for the Winnow algorithms, SASB has the advantage of not needing any parameters. The Winnow family algorithms have parameters, and the algorithms' mistake bounds depend on setting the parameters to values that depend on 8. (Often, the value of 8 will not be known by the learner.) We expect the techniques used to obtain this bound to be useful in analyzing other apobayesian learning algorithms. A number of authors have done related research regarding worst-case on-line loss bounds including [Fre96,KW95,Vov90). Simulation experiments involving a Bayesian algorithm and a mistake-driven variant are described in [Lit95). That paper provides useful background for this paper. Note that our present analysis techniques do not apply to the apobayesian algorithm studied there. The closest of the original Winnow family algorithms to SASB appears to be the Weighted MaJority algorithm [LW94], which was analyzed for a case similar to that considered in this paper in [Lit89). One should get a roughly correct impression of SASB if 206 N. Littlestone and C. Mesterharm one thinks of it as a version of the Weighted Majority algorithm that learns its parameters. In the next section we describe the Bayesian algorithm that we start with. In Section 3 we discuss its mistake-driven apobayesian variant. Section 4 mentions some simulation experiments using these algorithms, and Section 5 is the conclusion. 2 A Bayesian Learning Algorithm To describe the Bayesian learning algorithm we must specify a family of distributions over X x Y and a prior over this family of distributions. We parameterize the distributions with parameters ((h, ... , 8n+ l ) chosen from e = [0, 1 ]n+l. The parameter 8n+1 gives the probability that the label is 1, and the parameter 8i gives the probability that the ith attribute matches the label. Note that the probability that the ith attribute is 1 given that the label is 1 equals the probability that the ith attribute is 0 given that the label is O. We speak of this linkage between the probabilities for the two classes as a symmetry condition. With this linkage, the observation of a point from either class will affect the posterior distribution for both classes. It is perhaps more typical to choose priors that allow the two classes to be treated separately, so that the posterior for each class (giving the probability of elements of X conditioned on the label) depends only on the prior and on observations from that class. The symmetry condition that we impose appears to be important to the success of our analysis of the apobayesian variant of this algorithm. (Though we impose this condition to derive the algorithm, it turns out that the apobayesian variant can actually handle tasks where this condition is not satisfied.) We choose a prior on e that gives probability 1 to the set of all elements () = (81, ... , 8n +l ) E e for which at most one of 81 , ... ,8n does not equal !. The prior is uniform on this set. Note that for any () in this set only a single attribute has a probability other than ~ of matching the label, and thus only a single attribute is relevant. Concentrating on this set turns out to lead to an apobayesian algorithm that can, in fact, handle more than one relevant attribute and that performs particularly well when only a small fraction of the attributes are relevant. This prior is related to to the familiar Naive Bayes model, which also assumes that the attributes are conditionally independent given the labels. However, in the typical Naive Bayes model there is no restriction to a single relevant attribute and the symmetry condition linking the two classes is not imposed. Our prior leads to the following algorithm. (The name SBSB stands for "Symmetric Bayesian Algorithm with Singly-variant prior for Bernoulli distribution.") Algorithm SBSB Algorithm SBSB maintains counts Si of the number of times each attribute matches the label, a count M of the number of times the label is 1, and a count t of the number of trials. Initialization Si t- 0 for i = 1, ... ,n M t-O tt-O Prediction Predict 1 given instance (Xl, ... ,xn ) if and only if (M + 1) f= XiCSi+l)+Clixi)(t-Si+1) > (t - M + 1) f= (1-Xi)(Si+1~+XiCt-si+l) i=l (S,) i=l (S.) Update M t- M + y, t t- t + 1, and for each i, if Xi = Y then Si t- Si + 1 An Apobayesian Relative of Winnow 207 3 An Apobayesian Algorithm We construct an apobayesian algorithm by converting algorithm SBSB into a mistake-driven algorithm using the standard conversion given in the introduction. We call the resulting learning algorithm SASBj we have replaced "Bayesian" with "Apobayesian" in the acronym. In the previous section we made assumptions made about the generation of the instances and labels that led to SBSB and thence to SASB. These assumptions have served their purpose and we now abandon them. In analyzing the apobayesian algorithm we do not assume that the instances and labels are generated by some stochastic process. Instead we assume that the instance-label pairs in all of the trials are linearly-separable, that is, that there exist some WI, ., . ,Wn , and c such that for every instance-label pair (x, y) we have E~=I WiXi ;::: c when y = 1 and 2:~=1 WiXi ::; c when y = O. We actually make a somewhat stronger assumption, given in the following theorem, which gives our bound for the apobayesian algorithm. Theorem 1 Suppose that 'Yi ;::: 0 and "Ii ;::: 0 for i = 1, ... , n, and that 2:~=1 'Yi + "I i = 1. Suppose that 0 ::; bo < bi ::; 1 and let 8 = bi - bo. Suppose that algorithm SASB is run on a sequence of trials such that the instance x and label y in each trial satisfy 2:~=1 'YiXi + "Ii (1 - Xi) ::; bo if y = 0 and 2:~=1 'YiXi + "Ii (1 - Xi) ;::: bi if y = 1. Then the number of mistakes made by SASB will be bounded by * log 8; . We have space to say only a little about how the derivation of this bound proceeds. Details are given in [Lit96]. In analyzing SASB we work with an abstract description of the associated algorithm SBSB. This algorithm starts with a prior on e as described above. We represent this with a density Po. Then after each trial it calculates a new posterior density Pt(O) = t-d8kP(X'YI~k, where Pt is the density after trial t and P(x, ylO) is the pt-d )P(x,y ) conditional probability ofthe instance x and label y observed in trial t given O. Thus we can think of the algorithm as maintaining a current distribution on e that is initially the prior. SASB is similar, but it leaves the current distribution unchanged when a mistake is not made. For there to exist a finite mistake bound there must exist some possible choice for the current distribution for which SASB would make perfect predictions, should it ever arrive at that distribution. We call any such distribution leading to perfect predictions a possible target distribution. It turns out that the separability condition given in Theorem 1 guarantees that a suitable target distribution exists. The analysis proceeds by showing that for an appropriate choice of a target density p the relative entropy of the current distribution with respect to the target distribution, J p( 0) log(p( 0) / Pt (0)), decreases by at least some amount R > 0 whenever a mistake is made. Since the relative entropy is never negative, the number of mistakes is bounded by the initial relative entropy divided by R. This form of analysis is very similar to the analysis of the various members of the Winnow family in [Lit89,Lit91]. The same technique can be applied to other apobayesian algorithms. The abstract update of Pt given above is quite general. The success of the analysis depends on conditions on Po and P(x, ylO) that we do not have space here to discuss. 208 p = 0.01 k = 1 n = 20 250r---~--~----~--~--~ 200 f/) 150 ~ S .~ :e 100 50 'Optimal' 'SBSB' ... . 'SASB' -'SASB + voting' _.,/ /' / / / ,/ /' /'" '.' / -.---:: .. -. / ...... <-... ./::: .. . . / / / / / '" / / / / , / / o~--~--~----~--~--~ o 2000 4000 6000 8000 10000 Trials N. LittLestone and C. Mesterharm p = 0.1 k = 5 n = 20 250 .-----.-----,-----,---~--~ 200 100 'Optimal' 'SBSB' .. .. // . 'SASB' -.,: 'SASB + voting" -. - / / . /' / / ,/ .,/ ,/ / / / ,/ , 50 . /' . / /' ,/ . / / :'/.",,~ r O""'---...L-------I.-----'----~--~ o 2000 4000 6000 8000 1 0000 Trials Figure 1: Comparison of SASB with SBSB 4 Simulation Experiments The bound of the previous section was for perfectly linearly-separable data. We have also done some simulation experiments exploring the performance of SASB on non-separable data and comparing it with SBSB and with various other mistakedriven algorithms. A sample comparison of SASB with SBSB is shown in Figure 1. In each experimental run we generated 10000 trials with the instances and labels chosen randomly according to a distribution specified by (h = '" = Ok = 1 - p, Ok+l = ... = 0n+l = .5 where 01, ••. ,On+l are interpreted as specified in Section 2, n is the number of attributes, and n, p, and k are as specified at the top of each plot. The line labeled "optimal" shows the performance obtained by an optimal predictor that knows the distribution used to generate the data ahead of time, and thus does not need to do any learning. The lines labeled "SBSB" and "SASB" show the performance of the corresponding learning algorithms. The lines labeled "SASB + voting" show the performance of SASB with the addition of a voting procedure described in [Lit95]. This procedure improves the asymptotic mistake rate of the algorithms. Each line on the graph is the average of 30 runs. Each line plots the cumulative number of mistakes made by the algorithm from the beginning of the run as a function of the number of trials. In the left hand plot, there is only 1 relevant attribute. This is exactly the case that SBSB is intended for, and it does better than SASB. In right hand plot, there are 5 relevant attributes; SBSB appears unable to take advantage of the extra information present in the extra relevant attributes, but SASB successfully does. Comparison of SASB and previous Winnow family algorithms is still in progress, and we defer presenting details until a clearer picture has been obtained. SASB and the Weighted Majority algorithm often perform similarly in simulations. Typically, as one would expect, the Weighted Majority algorithm does somewhat better than An Apobayesian Relative of Winnow 209 SASB when its parameters are chosen optimally for the particular learning task, and worse for bad choices of parameters. 5 Conclusion Our mistake bounds and simulations suggest that SASB may be a useful alternative to the existing algorithms in the Winnow family. Based on the analysis style and the bounds, SASB should perhaps itself be considered a Winnow family algorithm. Further experiments are in progress comparing SASB with Winnow family algorithms run with a variety of parameter settings. Perhaps of even greater interest is the potential application of our analytic techniques to a variety of other apobayesian algorithms (though as we have observed earlier, the techniques do not appear to apply to all such algorithms) . We have already obtained some preliminary results regarding an interpretation of the Perceptron algorithm as an apobayesian algorithm. We are interested in looking for entirely new algorithms that can be derived in this way and also in better understanding the scope of applicability of our techniques. All of the analyses that we have looked at depend on symmetry conditions relating the probabilities for the two classes. It would be of interest to see what can be said when such symmetry conditions do not hold. In simulation experiments [Lit95], a mistake-driven variant of the standard Naive Bayes algorithm often does very well, despite the absence of such symmetry in the prior that it is based on. Our simulation experiments and also the analysis of the related algorithm Winnow [Lit91] suggest that SASB can be expected to handle some instance-label pairs inside of the separating gap or on the wrong side, especially if they are not too far on the wrong side. In particular it appears to be able to handle data generated according to the distributions on which SBSB is based, which do not in general yield perfectly separable data. It is of interest to compare the capabilities of the original Bayesian algorithm with the derived apobayesian algorithm. When the data is stochastically generated in a manner consistent with the assumptions behind the original algorithm, the original Bayesian algorithm can be expected to do better (see, for example, Figure 1). On the other hand, the apobayesian algorithm can handle data beyond the capabilities of the original Bayesian algorithm. For example, in the case we consider, the apobayesian algorithm can take advantage of the presence of more than one relevant attribute, even though the prior behind the original Bayesian algorithm assumes a single relevant attribute. Furthermore, as for all of the Winnow family algorithms, the mistake bound for the apobayesian algorithm does not depend on details of the behavior of the irrelevant attributes (including redundant attributes). Instead of using the apobayesian variant, one might try to construct a Bayesian learning algorithm for a prior that reflects the actual dependencies among the attributes and the labels. However, it may not be clear what the appropriate prior is. It may be particularly unclear how to model the behavior of the irrelevant attributes. Furthermore, such a Bayesian algorithm may end up being computationally expensive. For example, attempting to keep track of correlations among all pairs of attributes may lead to an algorithm that needs time and space quadratic in the number of attributes. On the other hand, if we start with a Bayesian algorithm that 210 N. Littlestone and C. Mesterharm uses time and space linear in the number of attributes we can obtain an apobayesian algorithm that still uses linear time and space but that can handle situations beyond the capabilities of the original Bayesian algorithm. Acknowledgments This paper has benefited from discussions with Adam Grove. References [CHP96] Nicolo Cesa-Bianchi, David P. Helmbold, and Sandra Panizza. On bayes methods for on-line boolean prediction. In Proceedings of the Ninth Annual Conference on Computational Learning Theory, pages 314-324, 1996. [Fre96] Yoav Freund. Predicting a binary sequence almost as well as the optimal biased coin. In Proceedings of the Ninth Annual Conference on Computational Learning Theory, pages 89-98, 1996. [KW95] J. Kivinen and M. K. Warmuth. Additive versus exponentiated gradient updates for linear prediction. In Proc. 27th ACM Symp. on Theory of Computing, pages 209-218, 1995. [Lit88] N. Littlestone. Learning quickly when irrelevant attributes abound: A new linear-threshold algorithm. Machine Learning, 2:285-318, 1988. [Lit89] N. Littlestone. Mistake Bounds and Logarithmic Linear-threshold Learning Algorithms. PhD thesis, Tech. Rept. UCSC-CRL-89-11, Univ. of Calif., Santa Cruz, 1989. [Lit91] N. Littlestone. Redundant noisy attributes, attribute errors, and linearthreshold learning using Winnow. In Proc. 4th Annu. Workshop on Comput. Learning Theory, pages 147- 156. Morgan Kaufmann, San Mateo, CA, 1991. [Lit95] N. Littlestone. Comparing several linear-threshold learning algorithms on tasks involving superfluous attributes. In Proceedings of the XII International conference on Machine Learning, pages 353- 361, 1995. [Lit96] N. Littlestone. Mistake-driven bayes sports: Bounds for symmetric apobayesian learning algorithms. Technical report, NEC Research Institute, Princeton, NJ, 1996. [LW94] N. Littlestone and M. K. Warmuth. The weighted majority algorithm. Information and Computation, 108:212-261, 1994. [Vov90] Volodimir G. Vovk. Aggregating strategies. In Proceedings of the 1990 Workshop on Computational Learning Theory, pages 371-383, 1990.
1996
96
1,298
LSTM CAN SOLVE HARD LO G TIME LAG PROBLEMS Sepp Hochreiter Fakultat fur Informatik Technische Universitat Munchen 80290 Miinchen, Germany Abstract Jiirgen Schmidhuber IDSIA Corso Elvezia 36 6900 Lugano, Switzerland Standard recurrent nets cannot deal with long minimal time lags between relevant signals. Several recent NIPS papers propose alternative methods. We first show: problems used to promote various previous algorithms can be solved more quickly by random weight guessing than by the proposed algorithms. We then use LSTM, our own recent algorithm, to solve a hard problem that can neither be quickly solved by random search nor by any other recurrent net algorithm we are aware of. 1 TRIVIAL PREVIOUS LONG TIME LAG PROBLEMS Traditional recurrent nets fail in case 'of long minimal time lags between input signals and corresponding error signals [7, 3]. Many recent papers propose alternative methods, e.g., [16, 12, 1,5,9]. For instance, Bengio et ale investigate methods such as simulated annealing, multi-grid random search, time-weighted pseudo-Newton optimization, and discrete error propagation [3]. They also propose.an EM approach [1]. Quite a few papers use variants of the "2-sequence problem" (and ('latch problem" ) to show the proposed algorithm's superiority, e.g. [3, 1, 5, 9]. Some papers also use the "parity problem", e.g., [3, 1]. Some of Tomita's [18] grammars are also often used as benchmark problems for recurrent nets [2, 19, 14, 11]. Trivial versus non-trivial tasks. By our definition, a "trivial" task is one that can be solved quickly by random search (RS) in weight space. RS works as follows: REPEAT randomly initialize the weights and test the resulting net on a training set UNTIL solution found. 474 s. Hochreiter and J. Schmidhuber Random search (RS) details. In all our RS experiments, we randomly initialize weights in [-100.0,100.0]. Binary inputs are -1.0 (for 0) and 1.0 (for 1). Targets are either 1.0 or 0.0. All activation functions are logistic sigmoid in [0.0,1.0]. We use two architectures (AI, A2) suitable for many widely used "benchmark" problems: Al is a fully connected net with 1 input, 1 output, and n biased hidden units. A2 is like Al with n = 10, but less densely connected: each hidden unit sees the input unit, _ the output unit, and itself; the output unit sees all other units; all units are biased. All activations are set to 0 at each sequence begin. We will indicate where we also use different architectures of other authors. All sequence lengths are randomly chosen between 500 and 600 (most other authors facilitate their problems by using much shorter training/test sequences). The "benchmark" problems always require to classify two types of sequences. Our training set consists of 100 sequences, 50 from class 1 (target 0) and 50 from class 2 (target 1). Correct sequence classification is defined as "absolute error at sequence end below 0.1". We stop the search once a random weight matrix correctly classifies all training sequences. Then we test on the test set (100 sequences). All results below are averages of 10 trials. In all our simulations below, RS finally classified all test set sequences correctly; average final absolute test set errors were always below 0.001 in most cases below 0.0001. "2-sequence p~oblem"(and "latch problem") [3, 1, 9]. The task is to observe and classify input sequences. There are two classes. There is only one input unit or input line. Only the first N real-valued sequence elements convey relevant information about the class. Sequence elements at positions t > N(we use N = 1) are generated by a Gaussian with mean z~ro and variance 0.2. The first sequence element is 1.0 (-1.0) for class 1 (2). Tar#t at sequence end is 1.0 (0.0) for class 1 (2) (the latch problem is a simple version of the 2-sequence problem that allows for input tuning instead of weight tuning). Bengio et ale's results. For the 2-sequence problem, the best method among the six tested by Bengio et all [3] was multigrid random search (sequence lengths 50 100; N and stopping criterion undefined), which solved the problem after 6,400 sequence presentations, with final classification error 0.06. In more recent work, Bengio andFrasconi reported that an EM-approach [1] solves the problem within 2,900 trials. RS results. RS with architecture A2 (AI, n = 1) solves the problem within only 718 (1247) trials on average. Using an architecture with only 3 parameters (as in Bengio et at's architecture for the latch problem [3]), the problem was solved within only 22 trials on average, due to tiny parameter space. According to our definition above, the problem is trivial. RS outperforms Bengio et al.'s methods in every respect: (1) many fewer trials required, (2) much less computation time per trial. Also, in most cases (3) the solution quality is better (less error). It should be mentioned, however, that different input representations and different types of noise may lead to worse RS performance (Yoshua Bengio, personal communication, 1996). "Parity problem". The parity task [3, 1] requires to classify sequences with several 100 elements (only l's or -1's) according to whether the number of l's is even or odd. The target at sequence end is 1.0 for odd and 0.0 for even. LSTM can Solve Time Problems 475 Bengio et al."s results. For sequences with only 25-50 steps, among the six methods tested in [3] only simulated annealing was reported to achieve final classification error of 0.000 (within about 810,000 triaIs-- the authors did not mention the precise stopping criterion). A method called "discrete error BP" took about 54,000 trials to achieve final classification error 0.05. In more recent work [1], for sequences with 250-500 steps, their EM-approach took about 3,400 trials to achieve -final classification error 0.12. RS results. RS with Al (n = 1) solves the problem within only 2906 trials on average. RS with A2 solves it within 2797 trials. We also ran another experiment with architecture A2, but without self-connections for hidden units. RS solved the problem within 250 trials on average. Again it should be mentioned that different input representations and noise types may lead to worse RS performance (Yoshua Bengio, personal communication, 1996). Tomita grammars. Many authors also use Tomita's grammars [18] to test their algorithms. See, e.g., [2, 19, -14, 11, 10]. Since we already tested parity problems above, we now focus on a few "parity-free" Tomita grammars (nr.s #1, #2, #4). Previous work facilitated the problems by restricting sequence length. E.g., in [11], maximal test (training) sequence length is 15 (10). Reference [11] reports the number of sequences required for convergence (for various first and second order nets with 3 to 9 units): Tomita #1: 23,000 - 46,000; Tomita #2: 77,000 - 200,000; Tomita #4: 46,000 - 210,000. RS, however, clearly outperforms the methods in [11]. The average results are: Tomita #1: 182 (AI, n = 1) and 288 (A2), Tomita #2: 1,511 (AI, n = 3) and 17,953 (A2), Tomita #4: 13,833 (AI, n =2) and 35,610 (A2). Non-trivial tasks / Outline of remainder. Solutions of non-trivial tasks are sparse in weight space. They require either many free parameters (e.g., input weights) or high weight precision, such that RS becomes infeasible. To solve such tasks we need a novel method called "Long Short-Term Memory", or LSTM for short [8]. Section 2 will briefly review LSTM. Section 3 will show results on a task that cannot be solved at all by any other recurrent net learning algorithm we are aware of. The task involves distributed, high-precision, continuous-valued representations and long minimaltime lags there are no short time lag training exemplars facilitating learning. 2 LONG SHORT-TERM MEMORY Memory cells and gate units: basic ideas. LSTM's basic unit is called a memory cell. Within each memory cell, there is a linear unit with a fixed-weight self-connection (compare Mozer's time constants [12]). This enforces constant, nonexploding, non-vanishing error flow within the memory cell. A multiplicative input gate unit learns to protect the constant error flow within the memory cell from perturbation by irrelevant inputs. Likewise, a multiplicative output gate unit learns to protect other units from perturbation by currently irrelevant memory contents . stored in the memory cell. The gates learn-to open and close access to constant error flow. Why is constant error flow important¥ For instance, with conventional "backprop through time" (BPTT, e.g., [20]) or RTRL (e.g., [15]), error signals ''flowing backwards in time" tend to vanish: the temporal evolution of the backpropagated 476 S. Hochreiter and1. Schmidhuber error exponentially depends on the size ofthe weights. For the first theoretical error flow analysis see [7]. See [3] for a more recent, independent, essentially identical analysis. LSTM details. In what follows, Wuv denotes the weight on the connection from unit v to unit u. netu(t), yU(t) are net input and activation ofunit u (with activation function fu) at time t. For all non-input units that aren't memory cells (e.g. output units), we have yU(t) = fu(netu(t)), where netu(t) = Lv wuvyV(t - 1). The jth memory cell is denoted Cj. Each memory cell is built around a central linear unit with a fixed self-connection (weight 1.0) and identity function as activation function (see definition of SCj below). In addition to netCj(t) =Lu WCjuyU(t - 1), c; also gets input from a special unit out; (the "output gate"), and from another special unit in; (the "input gate"). in; 's activation at time t is denoted by ifnj(t). out; 's activation at time t is denoted by youtj(t). in;, outj are viewed as ordinary hidden units. We have youtj(t) = !outj(netoutj(t)), yin;(t) = finj(netinj(t)), where netoutj(t) = Lu WoutjuyU(t - 1), netinj(t) = Lu WinjuyU(t - 1). The summation indices u may stand for input units, gate units, memory cells, or even conventional hidden units if there are any (see also paragraph on "network topology" below). All these different types of units may convey useful information about the current state of the net. For instance, an input gate (output gate) may use inputs from other memory cells to decide whether to store (access) certain information in its memory cell. There even may be recurrent self-connections like WCjCj . It is up to the user to define the network topology. At time t, Cj'S output yCj(t) is computed in a sigma-pi-likefashion: yCj(t) =youtj(t)h(scj(t)), where SCj(O) =0, SCj(t) = SCj(t -1) + ifnj(t)g (netCj(t)) for t > o. The differentiable function 9 scales netCj . The differentiable function h scales memory cell outputs computed from the internal state SCj. Why gate units? in; controls the error flow to memory cell Cj 's input connections WCju • out; controls the error flow from unit j's output connections..Error signals trapped within a memory cell cannot change - but different error signals flowing into the ceil (at different times) via its output gate may get superimposed. The output gate will have to learn which errors to trap in its memory cell, by appropriately scaling them. Likewise, the input gate will have to learn when to release errors. Gates open and close access to constant error flow. Network topology. There is one input, one hidden, 'and one output layer.. The fully self-connected hidden layer contains memory cells and corresponding gate units (for convenience, we refer to both memory cells and gate units as hidden units located in the hidden layer). The hidden layer may also contain "conventional" hidden units providing inputs to gate uni~s and memory cells. All units (except for gate units) in a~llayers have directed connections (serve as inputs) to all units in higher layers. Memory.cell blocks. S memory cells sharing one input gate a:nd one output gate form a "memory cell block of size S".. They can facilitate information storage. Learning with excellent computational complexitysee details in appendix of [8]. We use a variant of RTRL which properly takes into account the altered (sigma-pi-like) dynamics caused by input and output gates. Ho~ever, to ensure constant error backprop, like with truncated BP.TT [20], errors arriving at "memory LSTM can Solve Time Lag Problems 477 cell net inputs" (for cell Cj, this includes nete ., netin ., netout .) do not get propagated . J J J back further in time (although they do serve to change the incoming weights). Only within memory cells, errors are propagated back through previous internal states SCj. This enforces constant error flow within memory cells. Thus only the derivatives 8s c d ~ nee to be stored and updated. Hence, the algorithm is very efficient, and LSTM's update complexity per time step is excellent in comparison to other approaches such as RTRL: given n units and a fixed number ofoutput units, LSTM's update complexity per time step is at most O(n2), just like BPTI's. 3 EXPERIMENT: ADDING PROBLEM Our previous experimental comparisons (on widely used benchmark problems) with.RTRL (e.g., .[15]; results compared to the ones in [17]), Recurrent CascadeCorrelation [6], Elman nets (results compared to the ones in [4]), and Neural Sequence Chunking [16], demonstrated that LSTM leads to many more successful runs than its competitors, and learns much faster [8]. The following task, though, is more difficult than the above benchmark problems: it cannot be solved at all in reasonable time by RS (we tried various architectures) nor any other recurrent net learning algorithm we are aware of (see [13] for ·an overview). The experiment will show that LSTM can solve non-trivial, complex long time lag problems involving distributed, high-precision, continuous-valued representations. Task. Each element of each input sequence is a pair consisting of two components. The first component is a real value randomly chosen from the interval [-1,1]. The second component is ~ither 1.0, 0.0, or -1.0, and is used as a marker: at the end of each sequence, the task is to output the sum of the first components of those pairs that are marked by second components .equal to 1.0. The value T is used to determine average sequence length, which is a randomly chosen int~ger between T and T + 'fa. With a given sequence, exactly two pairs are marked as follows: we first randomly select and mark one of the first ten pairs (whose first component is called ,Xl). Then we randomly select and mark one of the first ~ -1 still unmarked pairs (whose first component is called X2). The second components of the remaining pairs are zero except for the first and final pair, whose second components are -1 (Xl is set to zero in the rare case where the first pair of the sequence got marke~. An error signal is generated only at the sequence end: the target is 0.5 + X)4~O 2 (the sum Xl + X2 scaled to the interval [0,1]). A sequence was processed correctly if the absolute error at the sequence end is below 0.04. Architecture. We use a 3-layer net with 2 input units, 1 output unit, and 2 memory cell blocks of size 2 (a cell block size of 1 works well, too). The output layer receives connections only from memory cells. Memory cells/ gate units receive inputs from memory cells/gate units (fully connected hidden layer). Gate units (finj,fout;) and output units are sigmoid in [0,1]. h is sigmoid in [-1,1], and 9 is sigmoid in [-2,2]. State drift versus initial bias. Note that the task requires to store the precise values of real numbers for long durations the system must learn to protect memory cell contents against even minor "internal state drifts". Our simple but highly effective way of solving drift problems at the. beginning oflearning is to initially bias the input gate in; towards zero. There is no need for fine tuning initial bias: with 478 S. Hochreiter and J. Schmidhuber sigmoid l~gistic activation functions, the precise initial bias hardly matters because vastly different initial bias values produce almost the same near-zero activations. In fact, the system itself learns to generate the most appropriate input gate bias. To study the significance of the drift problem, we bias all non-input units, thus artificially inducing internal state drifts. Weights (including bias weights) are randomly initialized in the range [-0.1,0.1]. The first (second) input gate bias is initialized with -3.0 (-6.0) (recall that the precise- initialization values hardly matters, as confirmed by additional experiments) .. Training / Testing. The learning rate is 0.5. Training examples are generated on-line. Training is stopped if the average training error is below 0.01, and the 2000 most recent sequences were processed correctly (see definition above). Results. With a test set consisting of 2560 randomly chosen.sequences, the average test set error was always below 0.01, and there were never more than 3 incorrectly processed sequences. The following results are means of 10 trials: For T = 100 (T =500, T = 1000), training was stopped after 74,000 (209,000; 853,000) training sequences, and then only 1 (0, 1) of the test sequences was not processed correctly. For T = 1000, the number of required training examples varied between 370,000 and 2,020,000, exceeding 700,000 in only 3 cases. The experiment demonstrates even for very long minimal time lags: (1) LSTM is able to work well with distributed representations. (2) LSTM is able to perform calculations involving high-precision, continuous values. Such tasks are impossible to solve within reasonable time by other algorithms: the main problem of gradientbased approaches (including TDNN, pseudo Newton) is their inability to deal with very long minimal time lags (vanishing gradient). A main problem of "global" and "discrete" approaches (RS, Bengio's and Frasconi's EM-approach, discrete error propagation) is their inability to deal with high-precision, continuous values. Other experiments. In [8] LSTM is used to solve numerous additional tasks that cannot be solved by other recurrent net learning algorithm we are aware of. For instance, LSTM can extract information conveyed by the temporal order of widely separated inputs. LSTM also can learn real-valued, conditional expectations of strongly delayed, noisy targets, given the inputs. Conclusion. For non-trivial tasks (where RS is infeasible), we recommend LSTM. 4 ACKNOWLEDGMENTS This work was supported by DFG grant SCRM 942/3-1 from "Deutsche Forschungsgemeinschaft" . References [1] Y. Bengio and P. Frasconi. Credit assignment through time: Alternatives to backpropagation. In J. D. Cowan, G. Tesauro, and J. Alspector, editors, Advances in Neural Information Processing Systems 6, pages 75-82. San Mateo, CA: Morgan Kaufmann, 1994. [2] Y. Bengio and P. Frasconi. An input output HMM architecture. In G. Tesauro, D. S. Touretzky, and T. K. Leen, editors, Advances in Neural Information Processing Systems 7, pages 427-434. MIT Press, Cambridge MA, 1995. LSTM can Solve 4 ............ oIIL.J'IV"'I.)IIi; Time Lag Problems 479 [3] Y. Bengio, P. Simard, and P. Frasconi. Learning long-term dependencies with gradient descent is difficult. IEEE Transactions on Neural Networks, 5(2):157-166, 1994. [4] A. Cleeremans, D. Servan-Schreiber, and J. L. McClelland. Finite-state automata and simple recurrent networks. Neural Computation, 1:372-381, 1989. [5] S. EI Hihi and Y. Bengio. Hierarchical recurrent neural networks for long-term dependencies. In Advances in Neural Information·Processing Systems 8, 1995. to appear. [6] S. E. Fahlman. The recurrent cascade-correlation learning algorithm. In R. P. Lippmann, J .. E. Moody, and D. S. Touretzky, editors, Advances in Neural Information Processing Systems 9, pages 190-196. San Mateo, CA: Morgan Kaufmann, 1991. [7] J. Hochreiter. Untersuchungen zu dynamischen neuronalen Netzen. Diploma thesis, Institut fiir Informatik, Lehrstuhl Prof. Brauer, Technische Universitat Munchen, 1991. See www7.informatik.tu-muenchen.de/-hochreit. [8] S. Hochreiter and J. Schmidhuber. Long short-term memory. Technical Report FKI207-95, Fakultii.t fur Informatik, Technische Universitat Munchen, 1995. Revised 1996 (see www.idsia.ch/-juergen, www7.informatik.tu-muenchen.dej-hochreit). [9] T. Lin, B. G. Horne, P. Tino, and C. L. Giles. Learning long-term dependencies is not as difficult with NARX recurrent neural networks. Technical Report UMIACSTR-95-78 and CS-TR-3500, Institute for Advanced Computer Studies, University of Maryland, College Pa.rk, MD 20742, 1995. [10] P. Manolios and R. Fanelli. First-order recurrent neural networks and deterministic finite state automata. Neural Computation, 6:1155-1173, 1994. [11] C. B. Miller and C. L. Giles. Experimental comparison of the effect of order in recurrent neural networks. International Journal of Pattern Recognition and Artificial Intelligence, 7(4):849-872, 1993. [12] M. C. Mozer. Induction of multiscale temporal structure. In J. E. Moody, S. J. Hanson, and R. P. Lippman, editors, Advances in Neural Information Processing Systems 4, pages 275-282. San Mateo, CA: Morgan Kaufmann, 1992. [13] B. A. Pearlmutter. Gradient calculations for dynamic recurrent neural networks: A survey. IEEE Transactions on Neural Networks, 6(5):1212-1228, 1995. [14] J. B. Pollack. The induction of dynamical recognizers. Machine Learning, 7:227-252, 1991. [15] A. J. Robinson and F. Fallside. The utility driven dynamic error propagation network. Technical Report CUEDjF-INFENGjTR.1, Cambridge University Engineering Department, 1987. [16] J. H. Schmidhuber. Learning complex, extended sequences using the principle of history compression. Neural Computation, 4(2):234-242, 1992. [17] A. W. Smith and D. Zipser. Learning sequential structures with the real-time recurrent learning algorithm. International Journal of Neural Systems, 1(2):125-131, 1989. [18] M. Tomita. Dynamic construction of finite automata from examples using hillclimbing. In Proceedings of the Fourth Annual Cognitive Science Conference, pages 105-108. Ann Arbor, MI, 1982. [19] R. L. Watrous and G. M. Kuhn. Induction of finite-state automata using secondorder recurrent networks. In J. E. Moody, S. J. Hanson, and R. P. Lippman, editors, Advances in Neural Information Processing Systems 4, pages 309-316. San Mateo, CA: Morgan Kaufmann, 1992. [20] R. J. Williams and J. Peng. An efficient gradient-based algorithm for on-line training of recurrent network trajectories. Neural Computation, 4:491-501, 1990.
1996
97
1,299
488 Solutions to the XOR Problem Frans M. Coetzee * eoetzee@eee.emu.edu Department of Electrical Engineering Carnegie Mellon University Pittsburgh, PA 15213 Virginia L. Stonick ginny@eee.emu.edu Department of Electrical Engineering Carnegie Mellon University Pittsburgh, PA 15213 Abstract A globally convergent homotopy method is defined that is capable of sequentially producing large numbers of stationary points of the multi-layer perceptron mean-squared error surface. Using this algorithm large subsets of the stationary points of two test problems are found. It is shown empirically that the MLP neural network appears to have an extreme ratio of saddle points compared to local minima, and that even small neural network problems have extremely large numbers of solutions. 1 Introduction The number and type of stationary points of the error surface provide insight into the difficulties of finding the optimal parameters ofthe network, since the stationary points determine the degree of the system[l]. Unfortunately, even for the small canonical test problems commonly used in neural network studies, it is still unknown how many stationary points there are, where they are, and how these are divided into minima, maxima and saddle points. Since solving the neural equations explicitly is currently intractable, it is of interest to be able to numerically characterize the error surfaces of standard test problems. To perform such a characterization is non-trivial, requiring methods that reliably converge and are capable of finding large subsets of distinct solutions. It can be shown[2] that methods which produce only one solution set on a given trial become inefficient (at a factorial rate) at finding large sets of multiple distinct solutions, since the same solutions are found repeatedly. This paper presents the first provably globally convergent homotopy methods capable of finding large subsets of the Currently with Siemens Corporate Research, Princeton NJ 08540 488 Solutions to the XOR Problem 411 stationary points of the neural network error surface. These methods are used to empirically quantify not only the number but also the type of solutions for some simple neural networks. 1.1 Sequential Neural Homotopy Approach Summary We briefly acquaint the reader with the principles of homotopy methods, since these approaches differ significantly from standard descent procedures. Homotopy methods solve systems of nonlinear equations by mapping the known solutions from an initial system to the desired solution of the unsolved system of equations. The basic method is as follows: Given a final set of equations /(z) = 0, xED ~ ?Rn whose solution is sought, a homotopy function h : D x T -+ ?Rn is defined in terms of a parameter T ETC ?R, such that { g(z) h(z, T) = /(z) when T = 0 when T = 1 where the initial system of equations g(z) = 0 has a known solution. For optimization problems f(x) = \7 x€2(x) where €2(x) is the error meaSUre. Conceptually, h(z, T) = 0 is solved numerically for z for increasing values of T, starting at T = 0 at the known solution, and incrementally varying T and correcting the solution x until T = 1, thereby tracing a path from the initial to the final solutions. The power and the problems of homotopy methods lie in constructing a suitable function h. Unfortunately, for a given f most choices of h will fail, and, with the exception of polynomial systems, no guaranteed procedures for selecting h exist. Paths generally do not connect the initial and final solutions, either due to non-existence of solutions, or due to paths diverging to infinity. However, if a theoretical proof of existence of a suitable trajectory can be constructed, well-established numerical procedures exist that reliably track the trajectory. The following theorem, proved in [2], establishes that a suitable homotopy exists for the standard feed-forward backpropagation neural networks: Theorem 1.1 Let €2 be the unregularized mean square error (MSE) problem for the multi-layer perceptron network, with weights f3 E ?Rn . Let f3 0 E U C ?Rn and a EVe ?Rn , where U and V are open bounded sets. Then except for a set of measure zero (f3, a) E U x V, the solutions ({3, T) of the set of equations h(f3, T) = (1- T)(f3 - f3o) + TD(J (€2 + J.t'¢'(IIf3 - aWn = 0 (1) where J.t > 0 and'¢' : ?R -+ ?R satisfies 2'¢'"(a2)a2 + '¢"(a2 ) > 0 as a -+ 00, form noncrossing one dimensional trajectories for all T E ?R, which are bounded V T E [0, 1]. Furthermore, the path through (f3o, 0) connects to at least one solution ({3* ,1) of the regularized MSE error problem (2) On T E [0,1] the approach corresponds to a pseudo-quadratic error surface being deformed continuously into the final neural network error surfacel . Multiple solu1 The common engineering heuristic whereby some arbitrary error surface is relaxed into another error surface generally does not yield well defined trajectories. (3 412 F M. Coetzee and V. L. Stonick tions can be obtained by choosing different initial values f30. Every desired solution (3" is accessible via an appropriate choice of a, since f30 = (3" suffices. Figure 1 qualitatively illustrates typical paths obtained for this homotopy2. The paths typically contain only a few solutions, are disconnected and diverge to infinity. A novel two-stage homotopy[2, 3] is used to overcome these problems by constructing and solving two homotopy equations. The first homotopy system is as described above. A synthetic second homotopy solves an auxiliary set of equations on a nonEuclidean compact manifold (sn (0; R) x A, where A is a compact subset of R) and is used to move between the disconnected trajectories of the first homotopy. The method makes use of the topological properties of the compact manifold to ensure that the secondary homotopy paths do not diverge. 4 +infty 3 2 0 0 -1 -infty -2 -infty 0 +infty -3 T -4 -4 -2 0 2 4 Figure 1: (a) Typical homotopy trajectories, illustrating divergence of paths and multiple solutions occurring on one path. (b) Plot of two-dimensional vectors used as training data for the second test problem (Yin-Yang problem). 2 Test Problems The test problems described in this paper are small to allow for (i) a large number of repeated runs, and (ii) to make it possible to numerically distinguish between solutions. Classification problems were used since these present the only interesting small problems, even though the MSE criterion is not necessarily best for classification. Unlike most classification tasks, all algorithms were forced to approximate the stationary point accurately by requiring the 11 norm of the gradient to be less than 10- 10 , and ensuring that solutions differed in h by more than 0.01. The historical XOR problem is considered first. The data points (-1, -1), (1,1)' '(-1,1) and (1,-1) were trained to the target values -0.8,-0.8, 0.8 and 0.8. A network with three inputs (one constant), two hidden layer nodes and one output node were used, with hyperbolic tangent transfer functions on the hidden and final 2Note that the homotopy equation and its trajectories exist outside the interval T = [0,1]. 488 Solutions to the XOR Problem 413 nodes. The regularization used fL = 0.05, tjJ(x) = x and a = 0 (no bifurcations were found for this value during simulations) . This problem was chosen since it is small enough to serve as a benchmark for comparing the convergence and performance of the different algorithms. The second problem, referred to as the Yin-Yang problem, is shown in Figure 1. The problem has 23 and 22 data points in classes one and two respectively, and target values ±0.7. Empirical evidence indicates that the smallest single hidden layer network capable of solving the problem has five hidden nodes. We used a net with three inputs, five hidden nodes and one output. This problem is interesting since relatively high classification accuracy is obtained using only a single neuron, but a 100% classification performance requires at least five hidden nodes and one of only a few global weight solutions. The stationary points form equivalence classes under renumbering of the weights or appropriate interchange of weight signs. For the XOR problem each solution class contains up to 22 2! = 8 distinct solutions; for the Yin-Yang network, there are 25 5! = 3840 symmetries. The equivalence classes are reported in the following sections. 3 Test Results A Ribak-Poliere conjugate gradient (CG) method was used as a control since this method can find only minima, as contrasted to the other algorithms, all of which are attracted by all stationary points. In the second algorithm, the homotopy equation (1) was solved by following the main path until divergence. A damped Newton (DN) method and the twa-stage homotopy method completed the set of four algorithms considered. The different algorithms were initialized with the same n random weights f30 E sn-l(o; v'2n). 3.1 Control- The XOR problem The total number and classification of the solutions obtained for 250 iterations on each algorithm are shown in Table 1. Table 1: Number of equivalence class solutions obtained. XOR Problem Algorithm # Solutions #Minima # Maxima #Saddle Points CG 17 17 0 0 DN 44 6 0 38 One Stage 28 16 0 12 Two Stage 61 17 0 44 I Total DIstmct 61 17 0 44 The probability of finding a given solution on a trial is shown in Figure 2. The twa-stage homotopy method finds almost every solution from every initial point. In contrast to the homotopy approaches, the Newton method exhibits poor convergence, even when heavily damped. The sets of saddle points found by the DN algorithm and the homotopy algorithms are to a large extent disjoint, even though the same initial weights were used. For the Newton method solutions close to the initial point are typically obtained, while the initial point for the homotopy algaPi Pi Pi 414 0.5 0.4 0.3 0.2 0.1 0 0 0.5 0.4 0.3 0.2 0.1 o o 0.5 0.4 0.3 0.2 0.1 0 0 0.8 0.6 0.4 0.2 0 0 50%-t 50%-t 50%-t 50%-t 10 20 r1r">~n. ~. 10 20 10 20 10 20 75%-t 30 40 75%-t 30 40 75%-t 30 40 75%-t 30 40 Equivalence Class F. M. Coetzee and V. L Stonick 50 .Jl • ..nnnn.. 50 50 50 lOO%-t 60 lOO%-t 60 lOO%-t 60 lOO%-t 60 Conjugate Gradient Newton Single Stage Homotopy Two Stage Homotopy Figure 2: Probability of finding equivalence class i on a trial. Solutions have been sorted based on percentage ofthe training set correctly classified. Dark bars indicate local minima, light bars saddle points. XOR problem 488 Solutions to the XOR Problem 415 Table 2: Number of solutions correctly classifying x% of target data. Classification 25 % 50 % 75 % 100 % Minimum 17 17 4 4 Saddle 44 44 20 0 Total Distinct 61 61 24 4 rithm might differ significantly from the final solution. This difference illustrates that homotopy arrives at solutions in a fundamentally different way than descent approaches. Based on these results we conclude that the two-stage homotopy meets its objective of significantly increasing the number of solutions produced on a single trial. The homotopy algorithms converge more reliably than Newton methods, in theory and in practise. These properties make homotopy attractive for characterizing error surfaces. Finally, due to the large number of trials and significant overlap between the solution sets for very different algorithms, we believe that Tables 1-2 represent accurate estimates for the number and types of solutions to the regularized XOR problem. 3.2 Results on the Yin-Yang problem The first three algorithms for the Yin-Yang problem were evaluated for 100 trials. The conjugate gradient method showed excellent stability, while the Newton method exhibited serious convergence problems, even with heavy damping. The two-stage algorithm was still producing solutions when the runs were terminated after multiple weeks of computer time, allowing evaluation of only ten different initial points. Table 3: Number of equivalence class solutions obtained. Yin-Yang Problem Algorithm # Solutions #Minima # Maxima #Saddle Points Conjugate Gradient 14 14 0 0 Damped Newton 10 0 0 10 One Stage Homotopy 78 15 0 63 Two Stage Homotopy 1633 12 0 1621 Total Dlstmct 1722 28 0 1694 Table 4: Number of solutions correctly classifying x% of target data. Classification 75 80 90 95 96 97 98 99 100% Minimum 28 28 28 26 26 5 5 2 2 Saddle 1694 1694 1682 400 400 13 13 3 3 Total Distinct 1722 1722 1710 426 426 18 18 5 5 The results in Tables 3-4 for the number of minima are believed to be accurate, due to verification provided by the conjugate gradient method. The number of saddle 416 F. M. Coetzee and V. L Stonick points should be seen as a lower bound. The regularization ensured that the saddle points were well conditioned, i.e. the Hessian was not rank deficient, and these solutions are therefore distinct point solutions. 4 Concl usions The homotopy methods introduced in this paper overcome the difficulties of poor convergence and the problem of repeatedly finding the same solutions. The use of these methods therefore produces significant new empirical insight into some extraordinary unsuspected properties of the neural network error surface. The error surface appears to consist of relatively few minima, separated by an extraordinarily large 'number of saddle points. While one recent paper by Goffe et al [4] had given some numerical estimates based on which it was concluded that a large number of minima in neural nets exist (they did not find a significant number of these), this extreme ratio of saddle points to minima appears to be unexpected. No maxima were discovered in the above runs; in fact none appear to exist within the sphere where solutions were sought (this seems likely given the regularization). The numerical results reveal astounding complexity in the neural network problem. If the equivalence classes are complete, then 488 solutions for the XOR problem are implied, of which 136 are minima. For the Yin-Yang problem, 6,600,000+ solutions and 107,250+ minima were characterized. For the simple architectures considered, these numbers appear extremely high. We are unaware of any other system of equations having these remarkable properties. Finally, it should be noted that the large number of saddle points and the small ratio of minima to saddle points in neural problems can create tremendous computational difficulties for approaches which produce stationary points, rather than simple minima. The efficiency of any such algorithm at producing solutions will be negated by the fact that, from an optimization perspective, most of these solutions will be useless. Acknowledgements. The partial support of the National Science Foundation by grant MIP-9157221 is gratefully acknowledged. References [1] E. H. Rothe, Introduction to Various Aspects of Degree Theory in Banach Spaces. Mathematical Surveys and Monographs (23), Providence, Rhode Island: American Mathematical Society, 1986. ISBN 0-82218-1522-9. [2] F. M. Coetzee, Homotopy Approaches for the Analysis and Solution of Neural Network and Other Nonlinear Systems of Equations. PhD thesis, Carnegie Mellon University, Pittsburgh, PA, May 1995. [3] F. M. Coetzee and V. L. Stonick, "Sequential homotopy-based computation of multiple solutions to nonlinear equations," in Proc. IEEE ICASSP, (Detroit), IEEE, May 1995. [4] W. L. Goffe, G. D. Ferrier, and J. Rogers, "Global optimization of statistical functions with simulated annealing," Jour. Econometrics, vol. 60, no. 1-2, pp. 65-99, 1994.
1996
98