index
int64 0
20.3k
| text
stringlengths 0
1.3M
| year
stringdate 1987-01-01 00:00:00
2024-01-01 00:00:00
| No
stringlengths 1
4
|
|---|---|---|---|
1,000
|
A Dynamical Model of Context Dependencies for the Vestibulo-Ocular Reflex Olivier J.M.D. Coenen* Terrence J. Sejnowskit Computational Neurobiology Laboratory Howard Hughes Medical Institute The Salk Institute for Biological Studies 10010 North Torrey Pines Road La Jolla, CA 92037, U.S.A. Departments oftBiology and *tPhysics University of California, San Diego La Jolla, CA 92093, U.S.A {olivier,terry}@salk.edu Abstract The vestibulo-ocular reflex (VOR) stabilizes images on the retina during rapid head motions. The gain of the VOR (the ratio of eye to head rotation velocity) is typically around -1 when the eyes are focused on a distant target. However, to stabilize images accurately, the VOR gain must vary with context (eye position, eye vergence and head translation). We first describe a kinematic model of the VOR which relies solely on sensory information available from the semicircular canals (head rotation), the otoliths (head translation), and neural correlates of eye position and vergence angle. We then propose a dynamical model and compare it to the eye velocity responses measured in monkeys. The dynamical model reproduces the observed amplitude and time course of the modulation of the VOR and suggests one way to combine the required neural signals within the cerebellum and the brain stem. It also makes predictions for the responses of neurons to multiple inputs (head rotation and translation, eye position, etc.) in the oculomotor system. 1 Introduction The VOR stabilizes images on the retina during rapid head motions: Rotations and translations of the head in three dimensions must be compensated by appropriate rotations of the eye. Because the head's rotation axis is not the same as the eye's rotation axis, the calculations for proper image stabilization of an object must take into account diverse variables such as object distance from each eye, 90 O. J. M. D. COENEN, T. J. SEJNOWSKI gaze direction, and head translation (Viire et al., 1986). The stabilization is achieved by integrating infonnation from different sources: head rotations from the semicircular canals of the inner ear, head translations from the otolith organs, eye positions, viewing distance, as well as other context infonnation, such as posture (head tilts) or activity (walking, running) (Snyder and King, 1992; Shelhamer et al.,1992; Grossman et al., 1989). In this paper we concentrate on the context modulation of the VOR which can be described by the kinematics of the reflex, i.e. eye position, eye vergence and head translation. 2 The Vestibulo-Ocular Reflex: Kinematic Model Definition of Vectors Coordinate System Eye position Vector Head Top View Target Object Gaze Vector Gaze Angle Interocular Distance Rotation Axis Semicircular Canals and Otoliths ¥ ~_--+_ Origin of coordinate syste,,-, (arbitrary) Figure 1: Diagram showing the definition of the vectors used in the equation of the kinematic model of the vestibulo-ocular reflex. The ideal VOR response is a compensatory eye movement which keeps the image fixed on the retina for any head rotations and translations. We therefore derived an equation for the eye rotation velocity by requiring that a target remains stationary on the retina. The velocity of the resulting compensatory eye rotation can be written as (see fig. 1): w = -Oe + 1:1 x [Dej x Oe - To;] (1) where Oe is the head rotation velocity sensed by the semicircular canals, TOj is the head translation velocity sensed by the otoliths, Dej == (e - OJ), e is a constant vector specifying the location of an eye in the head, OJ is the position of either the left or right otolith, fJ and Igl are the unit vector and amplitude of the gaze vector: fJ gives the eye position (orientation of the eye relative to the head), and Igl gives the distance from the eye to the object, and the symbol x indicates the cross-product between two vectors. wand Oe are rotation vectors which describe the instantaneous angUlar velocity of the eye and head, respectively. A rotation vector lies along the instantaneous axis of rotation; its magnitude indicates the speed of rotation around the axis, and its direction is given by the righthand screw rule. A motion of the head combining rotation (0) and translation (T) is sensed as the combination of a rotation velocity Oe measured by the semicircular canals and a translation velocity To sensed by the otoliths. The rotation vectors are equal (0 = Oe), and the translation velocity vector as measured by the otoliths is given by: TOj = OOj x 0 + T, where OOj == (a - OJ), and a is the position vector of the axis of rotation. A Dynarnical Model of Context Dependencies for the Vestibula-Ocular Reflex 91 The special case where the gaze is horizontal and the rotation vector is vertical (horizontal head rotation) has been studied extensively in the literature. We used this special case in the sirnulations. In that case w rnay be sirnplify by writing its equation with dot products. Since 9 and slc are then perpendicular (9 . fie = 0). the first term of the following expression in brackets is zero: (2) The sernicircular canals decornpose and report acceleration and velocity of head rotation fi by its cornponents along the three canals on each side of the head fie : horizontal. anterior and posterior. The two otolith organs on each side report the dynamical inertial forces generated during linear rnotion (translation) in two perpendicular plane. one vertical and the other horizontal relative to the head. Here we assurne that a translation velocity signal (To) derived frorn or reported by the otolith afferents is available. The otoliths encode as well the head orientation relative to the gravity vector force. but was not included in this study. To cornplete the correspondence between the equation and a neural correlate. we need to determine a physiological source for 9 and I!I. The eye position 9 is assurned to be given by the output of the velocity-to-position transformation or so-called "neural integrator" which provides eye position information and which is necessary for the activation of the rnotoneuron to sustain the eye in a fixed position. The integrator for horizontal eye position appears to be located in the nucleus prepositus hypoglossi in the pons. and the vertical integrator in the rnidbrain interstitial nucleus of Cajal. (Crawford. Cadera and Vilis. 1991; Cannon and Robinson. 1987). We assurne that the eye position is given as the coordinates of the unit vector 9 along the ~ and 1; of fig. 1. The eye position depends on the eye velocity according to '* = 9 x w. For the special case w(t) = w(t)z. i.e. for horizontal head rotation. the eye position coordinates are given by: 91 (t) = 91 (0) + f~ iJ2( r )w( r) dr 92(t) = 92(0) f~ 91(r)w(r)dr (3) This is a set of two negatively coupled integrators. The "neural integrator" therefore does not integrate the eye velocity directly but a product of eye position and eye velocity. The distance frorn eye to target I!I can be written using the gaze angles in the horizontal plane of the head: Right eye: 1 19RT (4) Left eye: 1 19LT (5) where «() R - () L) is the vergence angle. and I is the interocular distance; the angles are rneasured frorn a straight ahead gaze. and take on negative values when the eyes are turned towards the right. Within the oculornotor systern. the vergence angle and speed are encoded by the rnesencephalic reticular formation neurons (Judge and Curnrning. 1986; Mays. 1984). The nucleus reticularis tegrnenti pontis with reciprocal connections to the flocculus. oculornotor vermis. paravermis of the cerebellurn also contains neurons which activity varies linearly with vergence angle (Gamlin and Clarke. 1995). We conclude that it is possible to perform the cornputations needed to obtain an ideal VOR with signals known to be available physiologically. 92 Dynamical Model Overview Nod_ PftpoIItao IIyposIoooI O. J. M. D. COENEN, T. J. SEJNOWSKI Figure 2: Anatomical connections considered in the dynamical model. Only the left side is shown, the right side is identical and connected to the left side only for the calculation of vergence angle. The nucleus prepositus hypoglossi and the nucleus reticularis tegmenti pontis are meant to be representative of a class of nuclei in the brain stem carrying eye position or vergence signal. All connections are known to exist except the connection between the prepositus nucleus to the reticularis nucleus which has not been verified. Details of the cerebellum are in fig. 3 and of the vestibular nucleus in fig. 4. 3 Dynamical Model Snyder & King (1992) studied the effect of viewing distance and location of the axis of rotation on the VOR in monkeys; their main results are reproduced in fig. 5. In an attempt to reproduce their data and to understand how the signals that we have described in section 2 may be combined in time, we constructed a dynamical model based on the kinematic model. Its basic anatomical structure is shown in fig. 2. Details of the model are shown in fig. 3, and fig. 4 where all constants are written using a millisecond time scale. The results are presented in fig. 5. The dynamical variables represent the change of average firing rate from resting level of activity. The firing rate of the afferents has a tonic component proportional to the velocity and a phasic component proportional to the acceleration of movement. Physiologically, the afferents have a wide range of phasic and tonic amplitudes. This is reflected by a wide selection of parameters in the numerators in the boxes of fig. 3 and fig. 4. The Laplace transform of the integration operator in equation (3) of the eye position coordinates is ~. Following Robinson (1981), we modeled the neural integrator with a gain and a time constant of 20 seconds. We therefore replaced the pure integrator ~ with 20~~~~1 in the calculations of eye position. The term 1 in fig. 3 is calculated by using equations (4) and (5), and by using the integrator 9 20~o:!~1 on the eye velocity motor command to find the angles (h and (JR. The dynamical model is based on the assumption that the cerebellum is required for context modulation, and that because of its architecture, the cerebellum is more likely to implement complex functions of multiple signals than other relevant nuclei. The major contributions of vergence and eye position modulation on the VOR are therefore mediated by the cerebellum. Smaller and more transient contributions from eye position are assumed to be mediated through the vestibular nucleus as shown in fig. 4. The motivation for combining eye position as in fig. 4 are, first, the evidence for eye response oscillations; second, the theoretical consideration that linear movement information (To) is useless without eye position information for proper VOR. The parameters in the dynamical model were adjusted by hand after observing the behavior of the different components of the model and noting how these combine to produce the oscillations observed A Dynamical Model of Context Dependencies for the Vestibulo-Ocular Reflex Vestibular Semicirtular c..l O -----t 401+1 r-----®--f--..j 300+1 x OIolith 0Igan Cerebellum VHlibabr Nuc1tul 93 Figure 3: Contribution of the cerebellum to the dynamical model. Filtered velocity inputs from the canals and otoliths are combined with eye position according to equation (2). These calculations could be performed either outside the cerebellum in one or multiple brain stem nuclei (as shown) or possibly inside the cerebellum. The only output is to the vestibular nucleus. The Laplace notation is used in each boxes to represent a leaky integrator with a time constant. input derivative and input gain. The term oe are the coordinates of the vector oe shown in fig. 1. The x indicates a multiplication. The term! multiplies each inputs individually. The open arrows indicate inhibitory (negative) connections. VHlibalu Semicimtlu c.w Cere ... lIum O--'----t~l---+--®----t~~ X Figure 4: Contribution of the vestibular nucleus to the dynamical model. Three pathways in the vestibular nucleus process the canal and otolith inputs to drive the eye. The first pathway is modulated by the output of the cerebellum through a FIN (Flocculus Target Neuron). The second and third pathways report transient information from the inputs which are combined with eye position in a manner identical to fig. 3. The location of these calculations is hypothetical. in the data. Even though the number of parameters in the model is not small. it was not possible to fit any single response in fig. 5 without affecting most of the other eye responses. This puts severe limits on the set of parameters allowed in the model. The dynamical model suggests that the oscillations present in the data reflect: 1) important acceleration components in the neural signals. both rotational and linear, 2) different time delays between the canal and otolith signal processing. and 3) antagonistic or synergistic action of the canal and otolith signals with different axes of rotation, as described by the two terms in the bracket of equation (2). 4 Discussion By fitting the dynamical model to the data, we tested the hypothesis that the VOR has a response close to ideal taking into account the time constraints imposed by the sensory inputs and the neural networks performing the computations. The vector computations that we used in the model may not 94 ~ w O. J. M. D. COENEN, T. J. SEJNOWSKI Dynamical Model Responses vs Experimental Data 80 -20 LOMtIOftof .... 01 rotMIon -,a.-om -400~----~5~0------~ 10=0~ Time (m.) 80 60 40 20 -20 T .......... ~ .-40oL-----~ 5~ 0 ----~1~0~0- Time (m.) Figure 5: Comparison between the dynamical model and monkey data. The dotted lines show the effect of viewing distance and location of the axis of rotation on the VOR as recorded by Snyder & King (1992) from monkeys in the dark. The average eye velocity response (of left and right eye) to a sudden change in head velocity is shown for different target distances (left) and rotational axes (right). On the left, the location of the axis of rotation was in the midsagittal plane 12.5 cm behind the eyes (-12.5 cm), and the target distance was varied between 220 cm and 9 cm. On the right, the target di stance was kept constant at 9 cm in front of the eye, and the location of the axis of rotation was varied from 14 cm behind t04cm in front of the eyes (-14cm to 4cm) in the midsagittal plane. The solid lines show the model responses. The model replicates many characteristics of the data. On the left the model captures the eye velocity fluctuations between 20-50 ms, followed by a decrease and an increase which are both modulated with target distance (50-80 ms). The later phase of the response (80-100 ms) is almost exact for 220 cm, and one peak is seen at the appropriate location for the other distances. On the right the closest fits were obtained for the 4 cm and 0 cm locations. The mean values are in good agreement and the waveforms are close, but could be shifted in time for the other locations of the axis of rotations. Finally, the latest peak ( ..... lOOms) in the data appears in the model for -14 cm and 9 cm location. be the representation used in the oculomotor system. Mathematically, the vector representation is only one way to describe the computations involved. Other representations exist such as the quaternion representation which has been studied in the context of the saccadic system (Tweed and Vilis, 1987; see also Handzel and Flash, 1996 for a very general representation). Detailed comparisons between the model and recordings from neurons will be require to settle this issue. Direct comparison between Purkinje cell recordings (L.H. Snyder & W.M. King, unpublished data) and predictions of the model could be used to determine more precisely the different inputs to some Purkinje cells. The model can therefore be an important tool to gain insights difficult to obtain directly with experiments. The question of how the central nervous system learns the transformations that we described still remains. The cerebellum may be one site of learning for these transformations, and its output may modulate the VOR in real time depending on the context. This view is compatible with the results of Angelaki and Hess (1995) which indicate that the cerebellum is required to correctly perform an otolith transformation. It is also consistent with adaptation results in the VOR. To test this hypothesis, we have been working on a model of the cerebellum which learns to anticipate sensory inputs and feedbacks, and use these signals to modulate the VOR. The learning in the cerebellum and vestibular nuclei is mediated by the climbing fibers which report a reinforcement signal of the prediction error (Coenen and Sejnowski. in preparation). A Dynamical Model of Context Dependencies for the Vestibulo-Ocular Reflex 95 5 Conclusion Most research on the VOR has assumed forward gaze focussed at infinity. The kinematics of offcenter gaze and fixation at finite distance necessitates nonlinear corrections that require the integration of a variety of sensory inputs. The dynamical model studied here is a working hypothesis for how these corrections could be computed and is generally consistent with what is known about the cerebellum and brain stem nuclei. We are, however, far from knowing the mechanisms underlying these computations, or how they are learned through experience. 6 Acknowledgments The first author was supported by a McDonnell-Pew Graduate Fellowship during this research. We would like to thank Paul Viola for helpful discussions. References Angelaki, D. E. and Hess, B. J. (1995). Inertial representation of angular motion in the vestibular system of rhesus monkeyus. II. Otolith-controlled transformation that depends on an intact cerebellar nodulus. Journal of Neurophysiology, 73(5): 1729-1751. Cannon, S. C. and Robinson, D. A. (1987). Loss of the neural integrator of the oculomotor system from brain stem lesions in monkey. Journal of Neurophysiology, 57(5):1383-1409. Crawford, J. D., Cadera, W., and Vilis, T. (1991). Generation of torsional and vertical eye position signals by the interstitial nucleus of Cajal. Science, 252:1551-1553. Gamlin, P. D. R. and Clarke, R. J. (1995). Single-unit activity in the primate nucleus reticularis tegmenti pontis related to vergence and ocular accomodation. Journal of Neurophysiology, 73(5):2115-2119. Grossman, G. E., Leigh, R. J., Bruce, E. N., Huebner, W. P.,and Lanska, D.J. (1989). Performanceofthe human vestibu1oocu1ar reflex during locomotion. Journal of Neurophysiology, 62(1 ):264-272. Handzel, A. A. and Flash, T. (1996). The geometry of eye rotations and listing's law. In Touretzky, D., Mozer, M., and Hasselmo, M., editors, Advances in Neural Information Processing Systems 8, Cambridge, MA. MIT Press. Judge, S. J. and Cumming, B. G. (1986). Neurons in the monkey midbrain with activity related to vergence eye movement and accomodation. Journal of Neurophysiology, 55:915-930. Mays, L. E. (1984). Neural control of vergence eye movements: Convergence and divergence neurons in midbrain. Journal of Neurophysiology, 51:1091-1108. Robinson, D. A. (1981). The use of control systems analysis in the neurophysiology of eye movements. Ann. Rev. Neurosci., 4:463-503. Shelhamer, M., Robinson, D. A., and Tan, H. S. (1992). Context-specific adaptation of the gain of the vestibuloocular reflex in humans. Journal of Vestibular Research, 2:89-96. Snyder, L. H. and King, W. M. (1992). Effect of viewing distance and location ofthe axis of head rotation on the monkey's vestibuloocular reflex I. eye movement response. Journal of Neurophysiology, 67(4):861-874. Tweed, D. and Vilis, T. (1987). Implications of rotational kinematics for the oculomotor system in three dimensions. Journal of Neurophysiology, 58(4):832-849. Viire, E., Tweed, D., Milner, K., and Vilis, T. (1986). A reexamination of the gain ofthe vestibuloocular reflex. Journal of Neurophysiology, 56(2):439-450.
|
1995
|
101
|
1,001
|
Constructive Algorithms for Hierarchical Mixtures of Experts S.R.Waterhouse A.J.Robinson Cambridge University Engineering Department, Trumpington St., Cambridge, CB2 1PZ, England. Tel: [+44] 1223 332754, Fax: [+44] 1223 332662, Email: srw1001.ajr@eng.cam.ac.uk Abstract We present two additions to the hierarchical mixture of experts (HME) architecture. By applying a likelihood splitting criteria to each expert in the HME we "grow" the tree adaptively during training. Secondly, by considering only the most probable path through the tree we may "prune" branches away, either temporarily, or permanently if they become redundant. We demonstrate results for the growing and path pruning algorithms which show significant speed ups and more efficient use of parameters over the standard fixed structure in discriminating between two interlocking spirals and classifying 8-bit parity patterns. INTRODUCTION The HME (Jordan & Jacobs 1994) is a tree structured network whose terminal nodes are simple function approximators in the case of regression or classifiers in the case of classification. The outputs of the terminal nodes or experts are recursively combined upwards towards the root node, to form the overall output of the network, by "gates" which are situated at the non-terminal nodes. The HME has clear similarities with tree based statistical methods such as Classification and Regression Trees (CART) (Breiman, Friedman, Olshen & Stone 1984). We may consider the gate as replacing the set of "questions" which are asked at each branch of CART. From this analogy, we may consider the application of the splitting rules used to build CART. We start with a simple tree consisting of two experts and one gate. After partially training this simple tree we apply the splitting criterion to each terminal node. This evaluates the log-likelihood increase by splitting each expert into two experts and a gate. The split which yields the best increase in log-likelihood is then added permanently to the tree. This process of training followed by growing continues until the desired modelling power is reached. Constructive Algorithms for Hierarchical Mixtures of Experts 585 Figure 1: A simple mixture of experts. This approach is reminiscent of Cascade Correlation (Fahlman & Lebiere 1990) in which new hidden nodes are added to a multi-layer perceptron and trained while the rest of the network is kept fixed. The HME also has similarities with model merging techniques such as stacked regression (Wolpert 1993), in which explicit partitions of the training set are combined. However the HME differs from model merging in that each expert considers the whole input space in forming its output. Whilst this allows the network more flexibility since each gate may implicitly partition the whole input space in a "soft" manner, it leads to unnecessarily long computation in the case of near optimally trained models. At anyone time only a few paths through a large network may have high probability. In order to overcome this drawback, we introduce the idea of "path pruning" which considers only those paths from the root node which have probability greater than a certain threshold. CLASSIFICATION USING HIERARCHICAL MIXTURES OF EXPERTS The mixture of experts, shown in Figure 1, consists of a set of "experts" which perform local function approximation. The expert outputs are combined by a gate to form the overall output. In the hierarchical case, the experts are themselves mixtures of further experts, thus extending the architecture in a tree structured fashion. Each terminal node or "expert" may take on a variety of forms, depending on the application. In the case of multi-way classification, each expert outputs a vector Yj in which element m is the conditional probability of class m (m = 1 ... M) which is computed using the soft max function: P(CmI x(n>, Wj) = exp(w~jx(n») It exp(w~.kX(n») k=1 where Wj = [wlj W2j ... WMj] is the parameter matrix for expert j and Ci denotes class i. The outputs of the experts are combined using a "gate" which sits at the nonterminal nodes. The gate outputs are estimates of the conditional probability of selecting the daughters of the non-terminal node given the input and the path taken to that node from the root node. This is once again computed using the softmax function: P(Zj I ~ (.), ~) = exp( ~ J ~(.» It <xp( ~f ~(.» where'; = [';1';2 ... ';f] is the parameter matrix for the gate, and Zj denotes expert j . 586 S. R. WATERHOUSE, A. 1. ROBINSON The overall output is given by a probabilistic mixture in which the gate outputs are the mixture weights and the expert outputs are the mixture components. The probability of class m is then given by: ] P(cmlz(n),8) = I: P(zilz(n), ~)P(Cmlz(n), Wi). i=1 A straightforward extension of this model also gives us the conditional probability ht) of selecting expert j given input zen) and correct class Ck, In order to train the HME to perform classification we maximise the log likelihood L = l:~=1 l:~=1 t~) log P(cm Iz(n), 8), where the variable t~) is one if m is the correct class at exemplar (n) and zero otherwise. This is done via the expectation maximisation (EM) algorithm of Dempster, Laird & Rubin (1977), as described by Jordan & Jacobs (1994). TREE GROWING The standard HME differs from most tree based statistical models in that its architecture is fixed. By relaxing this constraint and allowing the tree to grow, we achieve a greater degree of flexibility in the network. Following the work on CART we start with a simple tree, for instance with two experts and one gate which we train for a small number of cycles. Given this semi-trained network, we then make a set of candidate splits {&} of terminal nodes {z;}. Each split involves replacin~ an expert Zi with a pair of new experts {Zu}j=1 and a gate, as shown in Figure 2. We wish to select eventually only the "best" split S out of these candidate splits. Let us define the best split as being that which maximises the increase in overall log-likelihood due to the split, IlL = L(P+1) L(P) where L(P) is the likelihood at the pth generation of the tree. If we make the constraint that all the parameters of the tree remain fixed apart from the paramL(P) \ \ L(P+I) Figure 2: Making a candidate split of a terminal node. \ \ eters of the new split whenever a candidate split is made, then the maximisation is simplified into a dependency on the increases in the local likelihoods {Li} of the nodes {Zi}. We thus constrain the tree growing process to be localised such that we find the node which gains the most by being split. max M(&) _ max M· = max(Ly*1) - L(P» i i I i I I Constructive Algorithms for Hierarchical Mixtures of Experts 587 Figure 3: Growing the HME. This figure shows the addition of a pair of experts to the partially grown tree. where n m L~+l) = L L t~) log L P(zijlz(n), c;;,zi)P(cmlz(n), zij, wij) n m j This splitting rule is similar in form to the CART splitting criterion which uses maximisation of the entropy of the node split, equivalent to our local increase in lop;-likelihood. TIle final growing algorithm starts with a tree of generation p and firstly fixes the parameters of all non-terminal nodes. All terminal nodes are then split into two experts and a gate. A split is only made if the sum of posterior probabilities En h~n), as described (1), at the node is greater than a small threshold. This prevents splits being made on nodes which have very little data assigned to them. In order to break symmetry, the new experts of a split are initialised by adding small random noise to the original expert parameters. The gate parameters are set to small random weights. For each node i, we then evaluate M; by training the tree using the standard EM method. Since all non-terminal node parameters are fixed the only changes to the log-likelihood are due the new splits. Since the parameters of each split are thus independent of one another, all splits can be trained at once, removing the need to train multiple trees separately. After each split has been evaluated, the best split is chosen. This split is kept and all other splits are discarded. The original tree structure is then recovered except for the additional winning split, as shown in Figure 3. The new tree, of generation p + I is then trained as usual using EM. At present the decision on when to add a new split to the tree is fairly straightforward: a candidate split is made after training the fixed tree for a set number of iterations. An alternative scheme we have investigated is to make a split when the overall log-likelihood of the fixed tree has not increased for a set number of cycles. In addition, splits are rejected if they add too little to the local log-likelihood. Although we have not discussed the issue of over-fitting in this paper, a number of techniques to prevent over-fitting can be used in the HME. The most simple technique, akin to those used in CART, involves growing a large tree and successively removing nodes from the tree until the performance on a cross validation set reaches an optimum. Alternatively the Bayesian techniques of Waterhouse, MacKay & Robinson (1995) could be applied. 588 S. R. WATERHOUSE, A. J. ROBINSON Tree growing simulations This algorithm was used to solve the 8-bit parity classification task. We compared the growing algorithm to a fixed HME with depth of 4 and binary branches. As can be seen in Figures 4(a) and (b), the factorisation enabled by the growing algorithm significantly speeds up computation over the standard fixed structure. The final tree shape obtained is shown in Figure 4(c). We showed in an earlier paper (Waterhouse & Robinson 1994) that the XOR problem may be solved using at least 2 experts and a gate. The 8 bit parity problem is therefore being solved by a series of XOR classifiers, each gated by its parent node, which is an intuitively appealing form with an efficient use of parameters. 8 ,E -200oL----1~0----2~0~--~~~--~4~0--~W Time (a) Evolution of log-likelihood vs. time in CPU seconds. -50 ~-100 "I .§' -150 -2000 1 2 3 4 5 6 Generation (b) Evolution of log-likelihood for (i) vs generations of tree. O'()OI 0.001 (c) Final tree structure obtained from (i), showing utilisation U; of each node where U; = L: P(z;, R;I:c(n») I N, and Ri is the path t~en from the root node to node i . Figure 4: HME GROWING ON THE 8 BIT PARITY PROBLEM;(i) growing HME with 6 generations; (ii) 4 deep binary branching HME (no growing). PATH PRUNING If we consider the HME to be a good model for the data generation process, the case for path pruning becomes clear. In a tree with sufficient depth to model the Constructive Algorithms for Hierarchical Mixtures of Experts 589 underlying sub-processes producing each data point, we would expect the activation of each expert to tend to binary values such that only one expert is selected at each time exemplar. The path pruning scheme is depicted in Figure 5. The pruning scheme utilises the "activation" of each node at each exemplar. The activation is defined as the product of node probabilities along a path from the root node to the current node, lin) = Li log P(zi/Ri, :.:(n»), where Ri is the path taken to node i from the root node. If .l}n) for node l at exemplar n falls below a threshold value, ft, then we ignore the subtree Sl and we backtrack up to the parent node of l. During training this involves not accumulating the statistics of the subtree Sl; during evaluation it involves setting the output of subtree Sl to zero. In addition to this path pruning scheme we can use the activation of the nodes to do more permanent pruning. If the overall utilisation Vi = Ln P(Zi, Rd:.:(n»)IN of a node falls below a small threshold, then a node is pruned completely from the tree. The sister subtrees of the removed node then subsume their parent nodes. This process is used solely to improve computational efficiency in this paper, although conceivably it could be used as a regularisation method, akin to the brain surgery techniques of Cun, Denker & Solla (1990). In such a scheme, however, a more useful measure of node utilisation would be the effective number of parameters (Moody 1992). Path pruning simulations ..... _---_ .. -.. Figure 5: Path pruning in the HME. Figure 6 shows the application of the pruning algorithm to the task of discriminating between two interlocking spirals. With no pruning the solution to the two-spirals takes over 4,000 CPU seconds, whereas with pruning the solution is achieved in 155 CPU seconds. One problem which we encountered when implementing this algorithm was in computing updates for the parameters of the tree in the case of high pruning thresholds. If a node is visited too few times during a training pass, it will sometimes have too little data to form reliable statistics and thus the new parameter values may be unreliable and lead to instability. This is particularly likely when the gates are saturated. To avoid this saturation we use a simplified version of the regularisation scheme described in Waterhouse et al. (1995). CONCLUSIONS We have presented two extensions to the standard HME architecture. By pruning branches either during training or evaluation we may significantly reduce the computational requirements of the HME. By applying tree growing we allow greater flexibility in the HME which results in faster training and more efficient use of parameters. 590 0 -20 "C -40 0 0 ;5 -60 ~ T -80 0> 0 ....J -100 -120 S. R. WATERHOUSE, A. J. ROBINSON (a) ,.,fi (iii " (iv) ,: .. .i .... ':1 ,I ,.,. , , , , I . / ( .'/ ."..:'" ."" ' ,,,.. ~ -,~ '.'' 10 100 1000 Time (5) (b) (c) Figure 6: The effect of pruning on the two spirals classification problem by a 8 deep binary branching hme:(a) Log-likelihood vs. Time (CPU seconds), with log pruning thresholds for experts and gates f: (i) f = -5. 6,(ii) f = -lO,(iii) f = -15,(iv) no pruning, (b) training set for two-spirals task; the two classes are indicated by crosses and circles, (c) Solution to two spirals problem. References Breiman, L., Friedman, J., Olshen, R. & Stone, C. J. (1984), Classification and Regression Trees, Wadswoth and Brooks/Cole. Cun, Y. L., Denker, J. S. & Solla, S. A. (1990), Optimal brain damage, in D. S. Touretzky, ed., 'Advances in Neural Information Processing Systems 2', Morgan Kaufmann, pp. 598-605. Dempster, A. P., Laird, N. M. & Rubin, D. B. (1977), 'Maximum likelihood from incomplete data via the EM algorithm', Journal of the Royal Statistical Society, Series B 39, 1-38. Fahlman, S. E. & Lebiere, C. (1990), The Cascade-Correlation learning architecture, Technical Report CMU-CS-90-100, School of Computer Science, Carnegie Mellon University, Pittsburgh, PA 15213. Jordan, M. I. & Jacobs, R. A. (1994), 'Hierarchical Mixtures of Experts and the EM algorithm', Neural Computation 6, 181-214. Moody, J. E. (1992), The effective number of parameters: An analysis of generalization and regularization in nonlinear learning systems, in J. E. Moody, S. J. Hanson & R. P. Lippmann, eds, 'Advances in Neural Information Processing Systems 4', Morgan Kaufmann, San Mateo, California, pp. 847-854. Waterhouse, S. R. & Robinson, A. J . (1994), Classification using hierarchical mixtures of experts, in 'IEEE Workshop on Neural Networks for Signal Processing', pp. 177-186. Waterhouse, S. R., MacKay, D. J. C. & Robinson, A. J. (1995), Bayesian methods for mixtures of experts, in M. C. M. D. S. Touretzky & M. E. Hasselmo, eds, 'Advances in Neural Information Processing Systems 8', MIT Press. Wolpert, D. H. (1993), Stacked generalization, Technical Report LA-UR-90-3460, The Santa Fe Institute, 1660 Old Pecos Trail, Suite A, Santa Fe, NM, 87501.
|
1995
|
102
|
1,002
|
Discovering Structure in Continuous Variables Using Bayesian Networks Reimar Hofmann and Volker Tresp* Siemens AG, Central Research Otto-Hahn-Ring 6 81730 Munchen, Germany Abstract We study Bayesian networks for continuous variables using nonlinear conditional density estimators. We demonstrate that useful structures can be extracted from a data set in a self-organized way and we present sampling techniques for belief update based on Markov blanket conditional density models. 1 Introduction One of the strongest types of information that can be learned about an unknown process is the discovery of dependencies and -even more important- of independencies. A superior example is medical epidemiology where the goal is to find the causes of a disease and exclude factors which are irrelevant. Whereas complete independence between two variables in a domain might be rare in reality (which would mean that the joint probability density of variables A and B can be factored: p(A, B) = p(A)p(B)), conditional independence is more common and is often a result from true or apparent causality: consider the case that A is the cause of B and B is the cause of C, then p(CIA, B) = p(CIB) and A and C are independent under the condition that B is known. Precisely this notion of cause and effect and the resulting independence between variables is represented explicitly in Bayesian networks. Pearl (1988) has convincingly argued that causal thinking leads to clear knowledge representation in form of conditional probabilities and to efficient local belief propagating rules. Bayesian networks form a complete probabilistic model in the sense that they represent the joint probability distribution of all variables involved. Two of the powerful Reimar.Hofmann@zfe.siemens.de Volker.Tresp@zfe.siemens.de Discovering Structure in Continuous Variables Using Bayesian Networks 501 features of Bayesian networks are that any variable can be predicted from any subset of known other variables and that Bayesian networks make explicit statements about the certainty of the estimate of the state of a variable. Both aspects are particularly important for medical or fault diagnosis systems. More recently, learning of structure and of parameters in Bayesian networks has been addressed allowing for the discovery of structure between variables (Buntine, 1994, Heckerman, 1995). Most of the research on Bayesian networks has focused on systems with discrete variables, linear Gaussian models or combinations of both. Except for linear models, continuous variables pose a problem for Bayesian networks. In Pearl's words (Pearl, 1988): "representing each [continuous] quantity by an estimated magnitude and a range of uncertainty, we quickly produce a computational mess. [Continuous variables] actually impose a computational tyranny of their own." In this paper we present approaches to applying the concept of Bayesian networks towards arbitrary nonlinear relations between continuous variables. Because they are fast learners we use Parzen windows based conditional density estimators for modeling local dependencies. We demonstrate how a parsimonious Bayesian network can be extracted out of a data set using unsupervised self-organized learning. For belief update we use local Markov blanket conditional density models which - in combination with Gibbs sampling- allow relatively efficient sampling from the conditional density of an unknown variable. 2 Bayesian Networks This brief introduction of Bayesian networks follows closely Heckerman, 1995. Considering a joint probability density I p( X) over a set of variables {Xl, ••. , X N} we can decompose using the chain rule of probability N p(x) = IIp(xiIXI, ... ,Xi-I). (1) i=l For each variable Xi, let the parents of Xi denoted by Pi ~ {XI, . .. , Xi- d be a set of variables2 that renders Xi and {x!, ... , Xi-I} independent, that is (2) Note, that Pi does not need to include all elements of {XI, ... , Xi-Il which indicates conditional independence between those variables not included in Pi and Xi given that the variables in Pi are known. The dependencies between the variables are often depicted as directed acyclic3 graphs (DAGs) with directed arcs from the members of Pi (the parents) to Xi (the child). Bayesian networks are a natural description of dependencies between variables if they depict causal relationships between variables. Bayesian networks are commonly used as a representation of the knowledge of domain experts. Experts both define the structure of the Bayesian network and the local conditional probabilities. Recently there has been great 1 For simplicity of notation we will only treat the continuous case. Handling mixtures of continuous and discrete variables does not impose any additional difficulties. 2Usually the smallest set will be used. Note that in Pi is defined with respect to a given ordering of the variables. :li.e. not containing any directed loops. 502 R. HOFMANN. V. TRESP emphasis on learning structure and parameters in Bayesian networks (Heckerman, 1995). Most of previous work concentrated on models with only discrete variables or on linear models of continuous variables where the probability distribution of all continuous given all discrete variables is a multidimensional Gaussian. In this paper we use these ideas in context with continuous variables and nonlinear dependencies. 3 Learning Structure and Parameters in Nonlinear Continuous Bayesian Networks Many of the structures developed in the neural network community can be used to model the conditional density distribution of continuous variables p( Xi IPi). Under the usual signal-plus independent Gaussian noise model a feedforward neural network N N(.) is a conditional density model such that p(Xi IPi) = G(Xi; N N(Pi), 0-2 ), where G(x; c, 0-2 ) is our notation for a normal density centered at c and with variance 0-2• More complex conditional densities can, for example, be modeled by mixtures of experts or by Parzen windows based density estimators which we used in our experiments (Section 5). We will use pM (Xi IP;) for a generic conditional probability model. The joint probability model is then N pM (X) = II pM (xi/Pi). (3) i=l following Equations 1 and 2. Learning Bayesian networks is usually decomposed into the problems of learning structure (that is the arcs in the network) and of learning the conditional density models pM (Xi IPi) given the structure4 . First assume the structure of the network is given. If the data set only contains complete data, we can train conditional density models pM (Xi IPi ) independently of each other since the log-likelihood of the model decomposes conveniently into the individual likelihoods of the models for the conditional probabilities. Next, consider two competing network structures. We are basically faced with the well-known bias-variance dilemma: if we choose a network with too many arcs, we introduce large parameter variance and if we remove too many arcs we introduce bias. Here, the problem is even more complex since we also have the freedom to reverse arcs. In our experiments we evaluate different network structures based on the model likelihood using leave-one-out cross-validation which defines our scoring function for different network structures. More explicitly, the score for network structure S is Score = 10g(p(S)) + Lev, where p(S) is a prior over the network structures and Lev = ~f=llog(pM (xkIS, X - {xk})) is the leave-one-out cross-validation loglikelihood (later referred to as cv-Iog-likelihood). X = {xk}f=l is the set of training samples, and pM (xk IS, X - {xk}) is the probability density of sample Xk given the structure S and all other samples. Each of the terms pM (xk IS, X - {xk}) can be computed from local densities using Equation 3. Even for small networks it is computationally impossible to calculate the score for all possible network structures and the search for the global optimal network structure 4Differing from Heckerman we do not follow a fully Bayesian approach in which priors are defined on parameters and structure; a fully Bayesian approach is elegant if the occurring integrals can be solved in closed form which is not the case for general nonlinear models or if data are incomplete. Discovering Structure in Continuous Variables Using Bayesian Networks 503 is NP-hard. In the Section 5 we describe a heuristic search which is closely related to search strategies commonly used in discrete Bayesian networks (Heckerman, 1995). 4 Prior Models In a Bayesian framework it is useful to provide means for exploiting prior knowledge, typically introducing a bias for simple structures. Biasing models towards simple structures is also useful if the model selection criteria is based on cross-validation, as in our case, because of the variance in this score. In the experiments we added a penalty per arc to the log-likelihood i.e. 10gp(S) ex: -aNA where NA is the number of arcs and the parameter a determines the weight of the penalty. Given more specific knowledge in form of a structure defined by a domain expert we can alternatively penalize the deviation in the arc structure (Heckerman, 1995). Furthermore, prior knowledge can be introduced in form of a set of artificial training data. These can be treated identical to real data and loosely correspond to the concept of a conjugate prior. 5 Experiment In the experiment we used Parzen windows based conditional density estimators to model the conditional densities pM (Xj IPd from Equation 2, i.e. (4) where {xi }f=l is the training set. The Gaussians in the nominator are centered at (x7, Pf) which is the location of the k-th sample in the joint input/output (or parent/child) space and the Gaussians in the denominator are centered at (Pf) which is the location of the k-th sample in the input (or parent) space. For each conditional model, (J"j was optimized using leave-one-out cross validation5• The unsupervised structure optimization procedure starts with a complete Bayesian model corresponding to Equation 1, i.e. a model where there is an arc between any pair of variables6 • Next, we tentatively try all possible arc direction changes, arc removals and arc additions which do not produce directed loops and evaluate the change in score. After evaluating all legal single modifications, we accept the change which improves the score the most. The procedure stops if every arc change decreases the score. This greedy strategy can get stuck in local minima which could in principle be avoided if changes which result in worse performance are also accepted with a nonzero probability 7 (such as in annealing strategies, Heckerman, 1995). Calculating the new score at each step requires only local computation. The removal or addition of an arc corresponds to a simple removal or addition of the corresponding dimension in the Gaussians of the local density model. However, 5Note that if we maintained a global (7 for all density estimators, we would maintain likelihood equivalence which means that each network displaying the same independence model gets the same score on any test set. 6The order of nodes determining the direction of initial arcs is random. 7 In our experiments we treated very small changes in score as if they were exactly zero thus allowing small decreases in score. 504 R. HOFMANN. V. TRESP 15~-----------------------, 100~------~------~------~ 10 - - - --50 ~ ~ "T ~ I -5 -100 -10~----------~----------~ o 50 100 -150~--------------------~ o 5 10 15 Number of Iterations Number of inputs Figure 1: Left: evolution of the cv-log-Iikelihood (dashed) and of the log-likelihood on the test set (continuous) during structure optimization. The curves are averages over 20 runs with different partitions of training and test sets and the likelihoods are normalized with respect to the number of cv- or test-samples, respectively. The penalty per arc was a = 0.1. The dotted line shows the Parzen joint density model commonly used in statistics, i.e. assuming no independencies and using the same width for all Gaussians in all conditional density models. Right: log-likelihood of the local conditional Parzen model for variable 3 (pM (x3IP3)) on the test set (continuous) and the corresponding cv-log-likelihood (dashed) as a function of the number of parents (inputs). crime ra.te 2 percent land zoned for lots a percent nonretail business 4 located on Charles river? 5 nitrogen oxide concentration 6 Average number of rooms 7 percent built before 1940 8 weighted distance to employment center 9 access to radial highways 10 tax rate 11 pupil/teacher ratio 12 percent black 13 percent lower-status population 14 median value of homes Figure 2: Final structure of a run on the full data set. after each such operation the widths of the Gaussians O'i in the affected local models have to be optimized. An arc reversal is simply the execution of an arc removal followed by an arc addition. In our experiment, we used the Boston housing data set, which contains 506 samples. Each sample consists of the housing price and 14 variables which supposedly influence the housing price in a Boston neighborhood (Figure 2). Figure 1 (left) shows an experiment where one third of the samples was reserved as a test set to monitor the process. Since the algorithm never sees the test data the increase in likelihood of the model on the test data is an unbiased estimator for how much the model has improved by the extraction of structure from the data. The large increase in the log-likelihood can be understood by studying Figure 1 (right). Here we picked a single variable (node 3) and formed a density model to predict this variable from the remaining 13 variables. Then we removed input variables in the order of their significance. After the removal of a variable, 0'3 is optimized. Note that the cv-Iog-likelihood increases until only three input variables are left due to the fact Discovering Structure in Continuous Variables Using Bayesian Networks 505 that irrelevant variables or variables which are well represented by the remaining input variables are removed. The log-likelihood of the fully connected initial model is therefore low (Figure 1 left). We did a second set of 15 runs with no test set. The scores of the final structures had a standard deviation of only 0.4. However, comparing the final structures in terms of undirected arcs8 the difference was 18% on average. The structure from one of these runs is depicted in Figure 2 (right). In comparison to the initial complete structure with 91 arcs, only 18 arcs are left and 8 arcs have changed direction. One of the advantages of Bayesian networks is that they can be easily interpreted. The goal of the original Boston housing data experiment was to examine whether the nitrogen oxide concentration (5) influences the housing price (14). Under the structure extracted by the algorithm, 5 and 14 are dependent given all other variables because they have a common child, 13. However, if all variables except 13 are known then they are independent. Another interesting question is what the relevant quantities are for predicting the housing price, i.e. which variables have to be known to render the housing price independent from all other variables. These are the parents, children, and children's parents of variable 14, that is variables 8, 10, 11, 6, 13 and 5. It is well known that in Bayesian networks, different constellations of directions of arcs may induce the same independencies, i.e. that the direction of arcs is not uniquely determined. It can therefore not be expected that the arcs actually reflect the direction of causality. 6 Missing Data and Markov Blanket Conditional Density Model Bayesian networks are typically used in applications where variables might be missing. Given partial information (i. e. the states of a subset of the variables) the goal is to update the beliefs (i. e. the probabilities) of all unknown variables. Whereas there are powerful local update rules for networks of discrete variables without (undirected) loops, the belief update in networks with loops is in general NP-hard. A generally applicable update rule for the unknown variables in networks of discrete or continuous variables is Gibbs sampling. Gibbs sampling can be roughly described as follows: for all variables whose state is known, fix their states to the known values. For all unknown variables choose some initial states. Then pick a variable Xi which is not known and update its value following the probability distribution p(xil{Xl, ... , XN} \ {xd) ex: p(xilPd II p(xjIPj ). (5) x.E1'j Do this repeatedly for all unknown variables. Discard the first samples. Then, the samples which are generated are drawn from the probability distribution of the unknown variables given the known variables. Using these samples it is easy to calculate the expected value of any of the unknown variables, estimate variances, covariances and other statistical measures such as the mutual information between variables. 8 Since the direction of arcs is not unique we used the difference in undirected arcs to compare two structures. We used the number of arcs present in one and only one of the structures normalized with respect to the number of arcs in a fully connected network. 506 R. HOFMANN, V. TRESP Gibbs sampling requires sampling from the univariate probability distribution in Equation 5 which is not straightforward in our model since the conditional density does not have a convenient form. Therefore, sampling techniques such as importance sampling have to be used. In our case they typically produce many rejected samples and are therefore inefficient. An alternative is sampling based on Markov blanket conditional density models. The Markov blanket of Xi, Mi is the smallest set of variables such that P(Xi I{ Xb . .. , XN} \ Xi) = P(Xi IMi) (given a Bayesian network, the Markov blanket of a variable consists of its parents, its children and its children's parents.). The idea is to form a conditional density model pM (xilMd ~ p(xdMd for each variable in the network instead of computing it according to Equation 5. Sampling from this model is simple using conditional Parzen models: the conditional density is a mixture of Gaussians from which we can sample without rejection9 • Markov blanket conditional density models are also interesting if we are only interested in always predicting one particular variable, as in most neural network applications. Assuming that a signal-plus-noise model is a reasonably good model for the conditional density, we can train an ordinary neural network to predict the variable of interest. In addition, we train a model for each input variable predicting it from the remaining variables. In addition to having obtained a model for the complete data case, we can now also handle missing inputs and do backward inference using Gibbs sampling. 7 Conclusions We demonstrated that Bayesian models of local conditional density estimators form promising nonlinear dependency models for continuous variables. The conditional density models can be trained locally if training data are complete. In this paper we focused on the self-organized extraction of structure. Bayesian networks can also serve as a framework for a modular construction of large systems out of smaller conditional density models. The Bayesian framework provides consistent update rules for the probabilities i.e. communication between modules. Finally, consider input pruning or variable selection in neural networks. Note, that our pruning strategy in Figure 1 can be considered a form of variable selection by not only removing variables which are statistically independent of the output variable but also removing variables which are represented well by the remaining variables. This way we obtain more compact models. If input values are missing then the indirect influence of the pruned variables on the output will be recovered by the sampling mechanism. References Buntine, W. (1994). Operations for learning with graphical models. Journal of Artificial Intelligence Research 2: 159-225. Heckerman, D. (1995). A tutorial on learning Bayesian networks. Microsoft Research, TR. MSR-TR-95-06, 1995. Pearl, J. (1988). Probabilistic Reasoning in Intelligent Systems. San Mateo, CA: Morgan Kaufmann. 9There are, however, several open issues concerning consistency between the conditional models.
|
1995
|
103
|
1,003
|
Temporal coding in the sub-millisecond range: Model of barn owl auditory pathway Richard Kempter* Institut fur Theoretische Physik Physik-Department der TU Munchen D-85748 Garching bei Munchen Germany J. Leo van Hemmen Institut fur Theoretische Physik Physik-Department der TU Munchen 0-85748 Garching bei Munchen Germany Wulfram Gerstner Institut fur Theoretische Physik Physik-Department der TU Munchen D-85748 Garching bei Munchen Germany Hermann Wagner Institut fur Zoologie Fakultiit fur Chemie und Biologie D-85748 Garching bei Munchen Germany Abstract Binaural coincidence detection is essential for the localization of external sounds and requires auditory signal processing with high temporal precision. We present an integrate-and-fire model of spike processing in the auditory pathway of the barn owl. It is shown that a temporal precision in the microsecond range can be achieved with neuronal time constants which are at least one magnitude longer. An important feature of our model is an unsupervised Hebbian learning rule which leads to a temporal fine tuning of the neuronal connections. ·email: kempter.wgerst.lvh@physik.tu-muenchen.de Temporal Coding in the Submillisecond Range: Model of Bam Owl Auditory Pathway 125 1 Introduction Owls are able to locate acoustic signals based on extraction of interaural time difference by coincidence detection [1, 2]. The spatial resolution of sound localization found in experiments corresponds to a temporal resolution of auditory signal processing well below one millisecond. It follows that both the firing of spikes and their transmission along the so-called time pathway of the auditory system must occur with high temporal precision. Each neuron in the nucleus magnocellularis, the second processing stage in the ascending auditory pathway, responds to signals in a narrow frequency range. Its spikes are phase locked to the external signal (Fig. 1a) for frequencies up to 8 kHz [3]. Axons from the nucleus magnocellularis project to the nucleus laminaris where signals from the right and left ear converge. Owls use the interaural phase difference for azimuthal sound localization. Since barn owls can locate signals with a precision of one degree of azimuthal angle, the temporal precision of spike encoding and transmission must be at least in the range of some 10 J.lS. This poses at least two severe problems. First, the neural architecture has to be adapted to operating with high temporal precision. Considering the fact that the total delay from the ear to the nucleus magnocellularis is approximately 2-3 ms [4], a temporal precision of some 10 J.lS requires some fine tuning, possibly based on learning. Here we suggest that Hebbian learning is an appropriate mechanism. Second, neurons must operate with the necessary temporal precision. A firing precision of some 10 J.ls seems truly remarkable considering the fact that the membrane time constant is probably in the millisecond range. Nevertheless, it is shown below that neuronal spikes can be transmitted with the required temporal precision. 2 Neuron model We concentrate on a single frequency channel of the auditory pathway and model a neuron of the nucleus magnocellularis. Since synapses are directly located on the soma, the spatial structure of the neuron can be reduced to a single compartment. In order to simplify the dynamics, we take an integrate-and-fire unit. Its membrane potential changes according to d u -u = -- + 1(t) dt TO (1) where 1(t) is some input and TO is the membrane time constant. The neuron fires, if u(t) crosses a threshold {) = 1. This defines a firing time to. After firing u is reset to an initial value uo = O. Since auditory neurons are known to be fast, we assume a membrane time constant of 2 ms. Note that this is shorter than in other areas of the brain, but still a factor of 4 longer than the period of a 2 kHz sound signal. The magnocellular neuron receives input from several presynaptic neurons 1 ~ k ~ J{. Each input spike at time t{ generates a current pulse which decays exponentially with a fast time constant Tr = 0.02 ms. The magnitude of the current pulse depends on the coupling strength h. The total input is t tf 1(t) = L h: exp( --=-.!. ) O(t - t{) k,f Tr (2) where O(x) is the unit step function and the sum runs over all input spikes. 126 R. KEMPTER, W. GERSTNER, J. L. VAN HEMMEN, H. WAGNER a) foE- T~ /\ h b /\ vvt I I I I I I I I I I I o <p 21t b) t t Fig. 1. Principles of phase locking and learning. a) The stimulus consists of a sound wave (top). Spikes of auditory nerve fibers leading to the nucleus magnocellularis are phase-locked to the periodic wave, that is, they occur at a preferred phase in relation to the sound, but with some jitter 0". Three examples of phase-locked spike trains are indicated. b) Before learning (left), many auditory input fibers converge to a neuron of the nucleus magnocellularis. Because of axonal delays which vary between different fibers, spikes arrive incoherently even though they are generated in a phase locked fashion. Due to averaging over several incoherent inputs, the total postsynaptic potential (bottom left) of a magnocellular neuron follows a rather smooth trajectory with no significant temporal structure. After learning (right) most connections have disappeared and only a few strong contacts remain. Input spikes now arrive coherently and the postsynaptic potential exhibits a clear oscillatory structure. Note that firing must occur during the rising phase of the oscillation. Thus output spikes will be phase locked. Temporal Coding in the Submillisecond Range: Model of Bam Owl Auditory Pathway 127 All input signals belong to the same frequency channel with a carrier frequency of 2 kHz (period T = 0.5 ms), but the inputs arise from different presynaptic neurons (1 ~ k ~ K). Their axons have different diameter and length leading to a signal transmission delay ~k which varies between 2 and 3 ms [4]. Note that a delay as small as 0.25 ms shifts the signal by half a period. Each input signal consists of a periodic spike train subject to two types of noise. First, a presynaptic neuron may not fire regularly every period but, on average, every nth period only where n ~ 1/(vT) and v is the mean firing rate of the neuron. For the sake of simplicity, we set n = 1. Second, the spikes may occur slightly too early or too late compared to the mean delay~. Based on experimental results, we assume a typical shift (1 = ±0.05 ms [3]. Specifically we assume in our model that inputs from a presynaptic neuron k arrive with the probability density P( J) __ 1_ ~ [-(t{ -nT- ~k)2l tk . m= L...t exp 2 v2~(1 2(1 n=-OO (3) where ~k is the axonal transmission delay of input k (Fig. 1). 3 Temporal tuning through learning We assume a developmental period of unsupervised learning during which a fine tuning of the temporal characteristics of signal transmission takes place (Fig. Ib). Before learning the magnocellular neuron receives many inputs (K = 50) with weak coupling (Jk = 1). Due to the broad distribution of delays the tptal input (2) has, apart from fluctuations, no temporal structure. After learning, the magnocellular neuron receives input from two or three presynaptic neurons only. The connections to those neurons have become very effective; cf. Fig. 2. a) <f ,c) <f ,30 20 10 0 2.0 30 20 10 0 2.0 2.5 ~[ms) 2.5 ~[ms) 3.0 3.0 b) <f ,d) <f ,30 20 10 0 2.0 30 20 10 0 2.0 2.5 ~[ms] 2.5 ~[ms) 3.0 3.0 Fig. 2. Learning. We plot the number of synaptic contacts (y-axis) for each delay ~ (x-axis). (a) At the beginning, the neuron has contacts to 50 presynaptic neurons with delays 2ms ~ ~ ~ 3ms. (b) and (c) During learning, some presynaptic neurons increase their number of contacts, other contacts disappear. (d) After learning, contacts to three presynaptic neurons with delays 2.25, 2.28, and 2.8 ms remain. The remaining contacts are very strong. 128 R. KEMPfER, W. GERSTNER, J. L. VAN HEMMEN, H. WAGNER The constant h: measures the total coupling strength between a presynaptic neuron k and the postsynaptic neuron. Values of h: larger than one indicate that several synapses have been formed. It has been estimated from anatomical data that a fully developed magnocellular neuron receives inputs from as few as 1-4 presynaptic neurons, but each presynaptic axon shows multiple branching near the postsynaptic soma and makes up to one hundred synaptic contacts on the soma of the magnocellular neuron[5]. The result of our simulation study is consistent with this finding. In our model, learning leads to a final state with a few but highly effective inputs. The remaining inputs all have the same time delay modulo the period T of the stimulus. Thus, learning leads to reduction of the number of input neurons contacts with a nucleus magnocellularis neuron. This is the fine tuning of the neuronal connections necessary for precise temporal coding (see below, section 4). a) 0.2 X -3: 0.0 o b) 1.0 X -w 0.5 0.0 o t:: j 0.0 5 X [ms] 5 X [ms] 0.5 10 10 Fig. 3. (a) Time window of learning W(x). Along the x-axis we plot the time difference between presynaptic and postsynaptic fiing x = t{ tl:. The window function W(x) has a positive and a negative phase. Learning is most effective, if the postsynaptic spike is late by 0.08 ms (inset). (b) Postsynaptic potential {(x). Each input spike evoked a postsynaptic potential which decays with a time constant of 2 ms. Since synapses are located directly at the soma, the rise time is very fast (see inset). Our learning scenario requires that the rise time of {(x) should be approximately equal to the time x where W(x) has its maximum. In our model, temporal tuning is achieved by a variant of Hebbian learning. In standard Hebbian learning, synaptic weights are changed if pre- and postsynaptic activity occurs simultaneously. In the context of temporal coding by spikes, the concept of (simultaneous activity' has to be refined. We assume that a synapse k is Temporal Coding in the Submillisecond Range: Model of Barn Owl Auditory Pathway 129 changed, if a presynaptic spike t{ and a postsynaptic spike to occur within a time window W(t{ -to). More precisely, each pair of presynaptic and postsynaptic spikes changes a synapse Jk by an amount (4) with a prefactor , = 0.2. Depending on the sign of W( x), a contact to a presynaptic neuron is either increased or decreased. A decrease below Jk = 0 is not allowed. In our model, we assume a function W(x) with two phases; cf. Fig. 3. For x ~ 0, the function W(x) is positive. This leads to a strengthening (potentiation) of the contact with a presynaptic neuron k which is active shortly before or after a postsynaptic spike. Synaptic contacts which become active more than 3 ms later than the postsynaptic spike are decreased. Note that the time window spans several cycles of length T. The combination of decrease and increase balances the average effects of potentiation and depression and leads to a normalization of the number and weight of synapses. Learning is stopped after 50.000 cycles of length T. 4 Temporal coding after learning After learning contacts remain to a small number of presynaptic neurons. Their axonal transmission delays coincide or differ by multiples of the period T. Thus the spikes arriving from the few different presynaptic neurons have approximately the same phase and add up to an input signal (2) which retains, apart from fluctuations, the periodicity of the external sound signal (Fig.4a). a) -. 9-+'" CJ) ~ o b) 1t 21t o 1t 21t Fig. 4. (a) Distribution of input phases after learning. The solid line shows the number of instances that an input spike with phase <p has occured (arbitrary units). The input consists of spikes from the three presynaptic neurons which have survived after learning; cf. Fig. 1 d. Due to the different delays, the mean input phase v(lries slightly between the three input channels. The dashed curves show the phase distribution of the individual channels, the solid line is the sum of the three dashed curves. (b) Distribution of output phases after learning. The histogram of output phases is sharply peaked. Comparison of the position of the maxima of the solid curves in (a) and (b) shows that the output is phase locked to the input with a relative delay fl<p which is related to the rise time of the postsynaptic potential. 130 R. KEMPTER, W. GERSTNER, J. L. VAN HEMMEN, H. WAGNER Output spikes of the magnocellular neuron are generated by the integrate-and-fire process (1). In FigAb we show a histogram of the phases of the output spikes. We find that the phases have a narrow distribution around a peak value. Thus the output is phase locked to the external signal. The width of the phase distribution corresponds to a precision of 0.084 phase cycles which equals 42 jlS for a 2 kHz stimulus. Note that the temporal precision of the output has improved compared to the input where we had three channels with slightly different mean phases and a variation of (T = 50jls each. The increase in the precision is due to the average over three uncorrelated input signals. We assume that the same principles are used during the following stages along the auditory pathway. In the nucleus laminaris several hundred signals are combined. This improves the signal-to-noise ratio further and a temporal precision below 10 jlS could be achieved. 5 Discussion We have demonstrated that precise temporal coding in the microsecond range is possible despite neuronal time constants in the millisecond range. Temporal refinement has been achieved through a slow developmental learning rule. It is a correlation based rule with a time window W which spans several milliseconds. Nevertheless learning leads to a fine tuning of the connections supporting temporal coding with a resolution of 42 jlS. The membrane time constant was set to 2 ms. This is nearly two orders of magnitudes longer than the achieved resolution. In our model, there is only one fast time constant which describes the typical duration of a input current pulse evoked by a presynaptic spike. Our value of Tr = 20 jlS corresponds to a rise time of the postsynaptic potential of 100 jls. This seems to be realistic for auditory neurons since synaptic contacts are located directly on the soma of the postsynaptic neuron. The basic results of our model can also be applied to other areas of the brain and can shed new light on some aspects of temporal coding with slow neurons. Acknowledgments: R.K. holds scholarship of the state of Bavaria. W.G. has been supported by the Deutsche Forschungsgemeinschaft (DFG) under grant number He 1729/22. H.W. is a Heisenberg fellow of the DFG. References [1] L. A. Jeffress, J. Compo Physiol. Psychol. 41, 35 (1948). [2] M. Konishi, Trends Neurosci. 9, 163 (1986). [3] C. E. Carr and M. Konishi, J. Neurosci. 10,3227 (1990). [4] W. E. Sullivan and M. Konishi, J. Neurosci. 4,1787 (1984). [5] C. E. Carr and R. E. Boudreau, J. Compo Neurol. 314, 306 (1991).
|
1995
|
104
|
1,004
|
Stable Dynamic Parameter Adaptation Stefan M. Riiger Fachbereich Informatik, Technische Universitat Berlin Sekr. FR 5-9, Franklinstr. 28/29 10587 Berlin, Germany async~cs. tu-berlin.de Abstract A stability criterion for dynamic parameter adaptation is given. In the case of the learning rate of backpropagation, a class of stable algorithms is presented and studied, including a convergence proof. 1 INTRODUCTION All but a few learning algorithms employ one or more parameters that control the quality of learning. Backpropagation has its learning rate and momentum parameter; Boltzmann learning uses a simulated annealing schedule; Kohonen learning a learning rate and a decay parameter; genetic algorithms probabilities, etc. The investigator always has to set the parameters to specific values when trying to solve a certain problem. Traditionally, the metaproblem of adjusting the parameters is solved by relying on a set of well-tested values of other problems or an intensive search for good parameter regions by restarting the experiment with different values. In this situation, a great deal of expertise and/or time for experiment design is required (as well as a huge amount of computing time). 1.1 DYNAMIC PARAMETER ADAPTATION In order to achieve dynamic parameter adaptation, it is necessary to modify the learning algorithm under consideration: evaluate the performance of the parameters in use from time to time, compare them with the performance of nearby values, and (if necessary) change the parameter setting on the fly. This requires that there exist a measure of the quality of a parameter setting, called performance, with the following properties: the performance depends continuously on the parameter set under consideration, and it is possible to evaluate the performance locally, i. e., at a certain point within an inner loop of the algorithm (as opposed to once only at the end of the algorithm). This is what dynamic parameter adaptation is all about. 226 S.M.RUOER Dynamic parameter adaptation has several virtues. It is automatic; and there is no need for an extra schedule to find what parameters suit the problem best. When the notion of what the good values of a parameter set are changes during learning, dynamic parameter adaptation keeps track of these changes. 1.2 EXAMPLE: LEARNING RATE OF BACKPROPAGATION Backpropagation is an algorithm that implements gradient descent in an error function E: IRn ~ llt Given WO E IRn and a fixed '" > 0, the iteration rule is WH1 = wt - ",V E(wt). The learning rate", is a local parameter in the sense that at different stages of the algorithm different learning rates would be optimal. This property and the following theorem make", especially interesting. Trade-off theorem for backpropagation. Let E: JR1l ~ IR be the error function of a neural net with a regular minimum at w· E IRn , i. e., E is expansible into a Taylor series about w· with vanishing gradient V E( w·) and positive definite Hessian matrix H(w·) . Let A denote the largest eigenvalue of H(w·). Then, in general, backpropagation with a fixed learning rate", > 2/ A cannot converge to w· . Proof. Let U be an orthogonal matrix that diagonalizes H(w·), i. e., D := UT H ( w·) U is diagonal. U sing the coordinate transformation x = UT (w - w·) and Taylor expansion, E(w) - E(w·) can be approximated by F(x) := xT Dx/2. Since gradient descent does not refer to the coordinate system, the asymptotic behavior of backpropagation for E near w· is the same as for F near O. In the latter case, backpropagation calculates the weight components x~ = x~(I- Dii",)t at time step t. The diagonal elements Dii are the eigenvalues of H(w·); convergence for all geometric sequences t 1-7 x~ thus requires", < 2/ A. I The trade-off theorem states that, given "', a large class of minima cannot be found, namely, those whose largest eigenvalue of the corresponding Hessian matrix is larger than 2/",. Fewer minima might be overlooked by using a smaller "', but then the algorithm becomes intolerably slow. Dynamic learning-rate adaptation is urgently needed for backpropagation! 2 STABLE DYNAMIC PARAMETER ADAPTATION Transforming the equation for gradient descent, wt+l = wt - ",VE(wt), into a differential equation, one arrives at awt fat = -",V E(wt). Gradient descent with constant step size", can then be viewed as Euler's method for solving the differential equation. One serious drawback of Euler's method is that it is unstable: each finite step leaves the trajectory of a solution without trying to get back to it. Virtually any other differential-equation solver surpasses Euler's method, and there are even some featuring dynamic parameter adaptation [5]. However, in the context of function minimization, this notion of stability ("do not drift away too far from a trajectory") would appear to be too strong. Indeed, differential-equation solvers put much effort into a good estimation of points that are as close as possible to the trajectory under consideration. What is really needed for minimization is asymptotic stability: ensuring that the performance of the parameter set does not decrease at the end of learning. This weaker stability criterion allows for greedy steps in the initial phase of learning. There are several successful examples of dynamic learning-rate adaptation for backpropagation: Newton and quasi-Newton methods [2] as an adaptive ",-tensor; individual learning rates for the weights [3, 8]; conjugate gradient as a one-dimensional ",-estimation [4]; or straightforward ",-adaptation [1, 7]. Stable Dynamic Parameter Adaptation 227 A particularly good example of dynamic parameter adaptation was proposed by Salomon [6, 7]: let ( > 1; at every step t of the backpropagation algorithm test two values for 17, a somewhat smaller one, 17d(, and a somewhat larger one, 17t(; use as 17HI the value with the better performance, i. e., the smaller error: The setting of the new parameter (proves to be uncritical (all values work, especially sensible ones being those between 1.2 and 2.1). This method outperforms many other gradient-based algorithms, but it is nonetheless unstable. b) Figure 1: Unstable Parameter Adaptation The problem arises from a rapidly changing length and direction of the gradient, which can result in a huge leap away from a minimum, although the latter may have been almost reached. Figure 1a shows the niveau lines of a simple quadratic error function E: 1R2 -+ IR along with the weight vectors wo, WI , . .. (bold dots) resulting from the above algorithm. This effect was probably the reason why Salomon suggested using the normalized gradient instead of the gradient, thus getting rid of the changes in the length of the gradient. Although this works much better, Figure 1b shows the instability of this algorithm due to the change in the gradient's direction. There is enough evidence that these algorithms converge for a purely quadratic error function [6, 7]. Why bother with stability? One would like to prove that an algorithm asymptotically finds the minimum, rather than occasionally leaping far away from it and thus leaving the region where the quadratic Hessian term of a globally nonquadratic error function dominates. 3 A CLASS OF STABLE ALGORITHMS In this section, a class of algorithms is derived from the above ones by adding stability. This class provides not only a proof of asymptotic convergence, but also a significant improvement in speed. Let E: IRn -+ IR be an error function of a neural net with random weight vector W O E IRn. Let ( > 1, 170 > 0, 0 < c ~ 1, and 0 < a ~ 1 ~ b. At step t of the algorithm, choose a vector gt restricted only by the conditions gtV E(wt)/Igtllv Ewt I ~ c and that it either holds for all t that 1/1gtl E [a, b) or that it holds for all t that IV E(wt)I/lgtl E [a, b), i. e., the vectors g have a minimal positive projection onto the gradient and either have a uniformly bounded length or are uniformly bounded by the length of the gradient. Note that this is always possible by choosing gt as the gradient or the normalized gradient. Let e: 17 t-t E (wt - 17gt) denote a one-dimensional error function given by E, wt and gt. Repeat (until the gradient vanishes or an upper limit of t or a lower limit Emin 228 S.M.ROOER of E is reached) the iteration WH1 = wt - 'T/tHgt with 'T/* .'T/t(/2 if e(O) < e('T/t() .- 1 + e('T/t() - e(O) 'T/Hl = 'T/t(gt\1 E(wt) (1) 'T/d( if e('T/d() ::; e('T/t() ::; e(O) 'T/t( otherwise. The first case for 'T/Hl is a stabilizing term 'T/*, which definitely decreases the error when the error surface is quadratic, i. e., near a minimum. 'T/* is put into effect when the errOr e(T}t() , which would occur in the next step if'T/t+l = 'T/t( was chosen, exceeds the error e(O) produced by the present weight vector wt . By construction, 'T/* results in a value less than 'T/t(/2 if e('T/t() > e(O); hence, given ( < 2, the learning rate is decreased as expected, no matter what E looks like. Typically, (if the values for ( are not extremely high) the other two cases apply, where 'T/t( and 'T/d ( compete for a lower error. Note that, instead of gradient descent, this class of algorithms proposes a "gt descent," and the vectors gt may differ from the gradient. A particular algorithm is given by a specification of how to choose gt. 4 PROOF OF ASYMPTOTIC CONVERGENCE Asymptotic convergence. Let E: w f-t 2:~=1 AiW; /2 with Ai > O. For all ( > 1, o < c ::; 1, 0 < a ::; 1 ::; b, 'T/o > 0, and WO E IRn, every algorithm from Section :1 produces a sequence t f-t wt that converges to the minimum 0 of E with an at least exponential decay of t f-t E(wt). Proof. This statement follows if a constant q < 1 exists with E(WH1 ) ::; qE(wt) for all t. Then, limt~oo wt = 0, since w f-t ..jE(w) is a norm in IRn. Fix a wt, 'T/t, and a gt according to the premise. Since E is a positive definite quadratic form, e: 'T/ f-t E( wt - 'T/gt) is a one-dimensional quadratic function with a minimum at, say, 'T/*. Note that e(O) = E(wt) and e('T/tH) = E(wt+l). e is completely determined by e(O), e'(O) = -gt\1 E(wt), 'T/te and e('T/t(). Omitting the algebra, it follows that 'T/* can be identified with the stabilizing term of (1). e(O) .A'-~--I qe( 0) -...-...J'----+I (1 - q11)e(0) + q11e('T/*) e"----r-++--+j qee(O) __ ~<-+--+I 11t+~:11· e(O) + (1 11t±~:11· )e('T/*) e( 'T/*) 1--____ ---""' ...... ----A~-_+_--+t e( 'T/tH) o Figure 2: Steps in Estimating a Bound q for the Improvement of E. Stable Dynamic Parameter Adaptation 229 If e(17t() > e(O), by (1) 17t+l will be set to 17·; hence, Wt+l has the smallest possible error e(17·) along the line given by l. Otherwise, the three values 0, 17t!(, and 17t( cannot have the same error e, as e is quadratic; e(17t() or e(17t!() must be less than e(O), and the argument with the better performance is used as 17tH' The sequence t I-t E(wt) is strictly decreasing; hence, a q ~ 1 exists. The rest of the proof shows the existence of a q < 1. Assume there are two constants 0 < qe, qT/ < 1 with E [qT/,2 - qT/] ~ qee(O). Let 17tH ~ 17·; using first the convexity of e, then (2), and (3), one obtains < < < e(17tH -17· 2 • + (1- 17t+l -17·) .) 17. 17 17. 17 17t+l -17· e(O) + (1- 17tH -17· )e(17.) 17· 17· (1 - qT/)e(O) + qf/e(17·) (1- qT/(1 - qe))e(O). (2) (3) Figure 2 shows how the estimations work. The symmetric case 0 < 17tH ~ 17· has the same result E(wt+l) ~ qE(wt) with q := 1 - qT/(1 - qe) < 1. Let ,X < := minPi} and ,X> := max{'xi}. A straightforward estimation for qe yields ,X< qe := 1 - c2 ,X> < 1. Note that 17· depends on wt and gt. A careful analysis of the recursive dependence of 17t+l /17· (wt , gt) on 17t /17·( wt - 1 ,l-l) uncovers an estimation ._ min _2_ ~ ca ~ 17o (,X > 0 ( <) 3/2 < qT/ .{(2 + l' (2 + 1 b'x> , bmax{1, J2'x> E(WO)}} . 5 NON-GRADIENT DIRECTIONS CAN IMPROVE CONVERGENCE • It is well known that the sign-changed gradient of a function is not necessarily the best direction to look for a minimum. The momentum term of a modified backpropagation version uses old gradient directions; Newton or quasi-Newton methods explicitly or implicitly exploit second-order derivatives for a change of direction; another choice of direction is given by conjugate gradient methods [5]. The algorithms from Section 3 allow almost any direction, as long as it is not nearly perpendicular to the gradient. Since they estimate a good step size, these algorithms can be regarded as a sort of "trial-and-error" line search without bothering to find an exact minimum in the given direction, but utilizing any progress made so far. One could incorporate the Polak-Ribiere rule, cttH = \1 E( Wt+l) + a(3ctt, for conjugate directions with dO = \1 E (WO), a = 1, and (\1E(Wt+l) - \1E(wt))\1E(wt+l) (3 = (\1 E(Wt))2 230 S.M. RUOER to propose vectors gt := ett /Iettl for an explicit algorithm from Section 3. As in the conjugate gradient method, one should reset the direction ett after each n (the number of weights) updates to the gradient direction. Another reason for resetting the direction arises when gt does not have the minimal positive projection c onto the normalized gradient. a = 0 sets the descent direction gt to the normalized gradient "V E(wt)/I"V E(wt)lj this algorithm proves to exhibit a behavior very similar to Salomon's algorithm with normalized gradients. The difference lies in the occurrence of some stabilization steps from time to time, which, in general, improve the convergence. Since comparisons of Salomon's algorithm to many other methods have been published [7], this paper confines itself to show that significant improvements are brought about by non-gradient directions, e. g., by Polak-Ribiere directions (a = 1). Table 1: Average Learning Time for Some Problems PROBLEM Emin a = 0 a = 1 (a) 3-2-4 regression 10° 195± 95% 58 ± 70% (b) 3-2-4 approximation 10-4 1070 ± 140% 189± 115% (c) Pure square (n = 76) 10-16 464± 17% 118± 9% (d) Power 1.8 (n = 76) 10-4 486± 29% 84± 23% (e) Power 3.8 (n = 76) 10-16 28 ± 10% 37± 14% (f) 8-3-8 encoder 10-4 1380± 60% 300± 60% Table 1 shows the average number of epochs of two algorithms for some problems. The average was taken over many initial random weight vectors and over values of ( E [1.7,2.1]j the root mean square error of the averaging process is shown as a percentage. Note that, owing to the two test steps for ",t/( and "'t(, one epoch has an overhead of around 50% compared to a corresponding epoch of backpropagation. a f:. 0 helps: it could be chosen by dynamic parameter adaptation. Problems (a) and (b) represent the approximation of a function known only from some example data. A neural net with 3 input, 2 hidden, and 4 output nodes was used to generate the example dataj artificial noise was added for problem (a). The same net with random initial weights was then used to learn an approximation. These problems for feedforward nets are expected to have regular minima. Problem (c) uses a pure square error function E: w rt L:~1 ilwil P /2 with p = 2 and n = 76. Note that conjugate gradient needs exactly n epochs to arrive at the minimum [5]. However, the few additional epochs that are needed by the a = 1 algorithm to reach a fairly small error (here 118 as opposed to 76) must be compared to the overhead of conjugate gradient (one line search per epoch). Powers other than 2, as used in (d) or (e), work well as long as, say, p > 1.5. A power p < 1 will (if n ~ 2) produce a "trap" for the weight vector at a location near a coordinate axis, where, owing to an infinite gradient component, no gradient-based algorithm can escape1 . Problems are expected even for p near 1: the algorithms of Section 3 exploit the fact that the gradient vanishes at a minimum, which in turn is numerically questionable for a power like 1.1. Typical minima, however, employ powers 2,4, ... Even better convergence is expected and found for large powers. IDynamic parameter adaptation as in (1) can cope with the square-root singularity (p = 1/2) in one dimension, because the adaptation rule allows a fast enough decay of the learning rate; the ability to minimize this one-dimensional square-root singularity is somewhat overemphasized in [7]. Stable Dynamic Parameter Adaptation 231 The 8-3-8 encoder (f) was studied, because the error function has global minima at the boundary of the domain (one or more weights with infinite length). These minima, though not covered in Section 4, are quickly found. Indeed, the ability to increase the learning rate geometrically helps these algorithms to approach the boundary in a few steps. 6 CONCLUSIONS It has been shown that implementing asymptotic stability does help in the case of the backpropagation learning rate: the theoretical analysis has been simplified, and the speed of convergence has been improved. Moreover, the presented framework allows descent directions to be chosen flexibly, e. g., by the Polak-Ribiere rule. Future work includes studies of how to apply the stability criterion to other parametric learning problems. References [1] R. Battiti. Accelerated backpropagation learning: Two optimization methods. Complex Systems, 3:331-342, 1989. [2] S. Becker and Y. Ie Cun. Improving the convergence of back-propagation learning with second order methods. In D. Touretzky, G. Hinton, and T. Sejnowski, editors, Proceedings of the 1988 Connectionist Models Summer School, pages 29-37. Morgan Kaufmann, San Mateo, 1989. [3] R. Jacobs. Increased rates of convergence through learning rate adaptation. Neural Networks, 1:295-307, 1988. [4] A. Kramer and A. Sangiovanni-Vincentelli. Efficient parallel learning algorithms for neural networks. In D. Touretzky, editor, Advances in Neural Information Processing Systems 1, pages 40-48. Morgan Kaufmann, San Mateo, 1989. [5] W. H. Press, B. P. Flannery, S. A. Teukolsky, and W. T. Vetterling. Numerical Recipes in C. Cambridge University Press, 1988. [6] R. Salomon. Verbesserung konnektionistischer Lernverfahren, die nach der Gradientenmethode arbeiten. PhD thesis, TU Berlin, October 1991. [7] R. Salomon and J. L. van Hemmen. Accelerating backpropagation through dynamic self-adaptation. Neural Networks, 1996 (in press). [8] F. M. Silva and L. B. Almeida. Speeding up backpropagation. In Proceedings of NSMS - International Symposium on Neural Networks for Sensory and Motor Systems, Amsterdam, 1990. Elsevier.
|
1995
|
105
|
1,005
|
Optimizing Cortical Mappings Geoffrey J. Goodhill The Salk Institute 10010 North Torrey Pines Road La Jolla, CA 92037, USA Steven Finch Human Communication Research Centre University of Edinburgh, 2 Buccleuch Place Edinburgh EH8 9LW, GREAT BRITAIN Terrence J. Sejnowski The Howard Hughes Medical Institute The Salk Institute for Biological Studies 10010 North Torrey Pines Road, La Jolla, CA 92037, USA & Department of Biology, University of California San Diego La Jolla, CA 92037, USA Abstract "Topographic" mappings occur frequently in the brain. A popular approach to understanding the structure of such mappings is to map points representing input features in a space of a few dimensions to points in a 2 dimensional space using some selforganizing algorithm. We argue that a more general approach may be useful, where similarities between features are not constrained to be geometric distances, and the objective function for topographic matching is chosen explicitly rather than being specified implicitly by the self-organizing algorithm. We investigate analytically an example of this more general approach applied to the structure of interdigitated mappings, such as the pattern of ocular dominance columns in primary visual cortex. 1 INTRODUCTION A prevalent feature of mappings in the brain is that they are often "topographic". In the most straightforward case this simply means that neighbouring points on a two-dimensional sheet (e.g. the retina) are mapped to neighbouring points in a more central two-dimensional structure (e.g. the optic tectum). However a more complex case, still often referred to as topographic, is the mapping from an abstract space of features (e.g. position in the visual field, orientation, eye of origin etc) to Optimizing Cortical Mappings 331 the cortex (e.g. layer 4 of VI). In many cortical sensory areas, the preferred sensory stimuli of neighbouring neurons changes slowly, except at discontinuous jumps, suggestive of an optimization principle that attempts to match "similar" features to nearby points in the cortex. In this paper, we (1) discuss what might constitute an appropriate measure of similarity between features, (2) outline an optimization principle for matching the similarity structure of two abstract spaces (i.e. a measure of the degree of topography of a mapping), and (3) use these ideas to analyse the case where two equivalent input variables are mapped onto one target structure, such as the "ocular dominance" mapping from the right and left eyes to VI in the cat and monkey. 2 SIMILARITY MEASURES A much-investigated computational approach to the study of mappings in VI is to consider the input features as pOints in a multidimensional euclidean space [1,5,9]. The input dimensions then consist of e.g. spatial position, orientation, ocular dominance, and so on. Some distribution of points in this space is assumed which attempts, in some sense, to capture the statistics of these features in the visual world. For instance, in [5], distances between points in the space are interpreted as a decreasing function of the degree to which the corresponding features are correlated over an ensemble of images. Some self-organizing algorithm is then applied which produces a mapping from the high-dimensional feature space to a two-dimensional sheet representing the cortex, such that nearby points in the feature space map to nearby points in the two-dimensional sheet.l However, such approaches assume that the dissimilarity structure of the input features is well-captured by euclidean distances in a geometric space. There is no particular reason why this should be true. For instance, such a representation implies that the dissimilarity between features can become arbitrarily large, an unlikely scenario. In addition, it is difficult to capture higher-order relationships in such a representation, such as that two oriented line-segment detectors will be more correlated if the line segments are co-linear than if they are not. We propose instead that, for a set of features, one could construct directly from the statistics of natural stimuli a feature matrix representing similarities or dissimilarities, without regard to whether the resulting relationships can be conveniently captured by distances in a euclidean feature space. There are many ways this could be done; one example is given below. Such a similarity matrix for features can then be optimally matched (in some sense) to a similarity matrix for positions in the output space. A disadvantage from a computational point of view of this generalized approach is that the self-organizing algorithms of e.g. [6,2] can no longer be applied, and possibly less efficient optimization techniques are required. However, an advantage of this is that one may now explore the consequences of optimizing a whole range of objective functions for quantifying the quality of the mapping, rather than having to accept those given explicitly or implicitly by the particular self-organizing algorithm. lWe mean this in a rather loose sense, and wish to include here the principles of mapping nearby points in the sheet to nearby points in the feature space, mapping distant points in the feature space to distant points in the sheet, and so on. 332 G. J. GOODHILL. S. FINCH. T.J. SEJNOWSKI Vout M Figure 1: The mapping framework. 3 OPTIMIZATION PRINCIPLES We now outline a general framework for measuring to what degree a mapping matches the structure of one similarity matrix to that of another. It is assumed that input and output matrices are of the same (finite) dimension, and that the mapping is bijective. Consider an input space Yin and an output space Vout, each of which contains N points. Let M be the mapping from points in Yin to points in Vout (see figure 1). We use the word "space" in a general sense: either or both of Yin and Vout may not have a geometric interpretation. Assume that for each space there is a symmetric "similarity" function which, for any given pair of points in the space, specifies how similar (or dissimilar) they are. Call these functions F for Yin and G for Vout. Then we define a cost functional C as follows N C = L L F(i,j)G(M(i), MO)), (1) i=1 i<i where i and j label pOints in ViT\J and M(i) and M(j) are their respective images in Vout. The sum is over all possible pairs of points in Yin. Since M is a bijection it is invertible, and C can equivalently be written N C = LL F(M-1(i),M-1(j))G(i,j), (2) i=1 i<i where now i and j label points in Vout! and M - I is the inverse map. A good (i.e. highly topographic) mapping is one with a high value of C. However, if one of F or G were given as a dissimilarity function (i.e. increasing with decreasing similarity) then a good mapping would be one with a low value of C. How F and G are defined is problem-specific. C has a number of important properties that help to justify its adoption as a measure of the degree of topography of a mapping (for more details see [3]). For instance, it can be shown that if a mapping that preserves ordering relationships between two similarity matrices exists, then maximizing C will find it. Such maps are homeomorphisms. However not all homeomorphisms have this propert}j so we refer to such "perfect" maps as "topographic homeomorphisms". Several previously defined optimization principles, such as minimum path and minimum Optimizing Cortical Mappings 333 wiring [1], are special cases of C. It is also closely related (under the assumptions above) to Luttrell's minimum distortion measure [7], if F is euclidean distance in a geometric input space, and G gives the noise process in the output space. 4 INTERDIGITATED MAPPINGS As a particular application of the principles discussed so far, we consider the case where the similarity structure of Yin can be expressed in matrix form as where Qs and Qc are of dimension Nil. This means that Yin consists of two halves, each with the same internal similarity structure, and an in general different similarity structure between the two halves. The question is how best to match this dual similarity structure to a single similarity structure in Vout. This is of mathematical interest since it is one of the simplest cases of a mismatch between the similarity structures of V in and Vout! and of biological interest since it abstractly represents the case of input from two equivalent sets of receptors coming together in a single cortical sheet, e.g. ocular dominance columns in primary visual cortex (see e.g. [8, 5]). For simplicity we consider only the case of two one-dimensional retinae mapping to a one-dimensional cortex. The feature space approach to the problem presented in [5] says that the dissimilarities in Yin are given by squared euclidean distances between points arranged in two parallel rows in a two-dimensional space. That is, { I· '12 . . l.- J F(l., J) = Ii _ j _ NIll2 + k2 : i, j in same half of Yin : i, j in different halves of Yin (3) assuming that indices 1 ... Nil give points in one half and indices Nil + 1 ... N give pOints in the other half. G (i, j) is given by G (. .) _ {1 : i, j neighbouring l., J 0 : otherwise (4) It can be shown that the globally optimal mapping (i.e. minimum of C) when k > 1 is to keep the two halves of V in entirely separate in Vout [5]. However, there is also a local minimum for an interdigitated (or "striped") map, where the interdigitations have width n = lk. By varying the value of k it is thus possible to smoothly vary the periodicity of the locally optimal striped map. Such behavior predicted the outcome of a recent biological experiment [4]. For k < 1 the globally optimal map is stripes of width n = 1. However, in principle many alternative ways of measuring the similarity in Yin are possible. One obvious idea is to assume that similarity is given directly by the degree of correlation between points within and between the two eyes. A simple assumption about the form of these correlations is that they are a gaussian function of physical distance between the receptors (as in [8]). That is, { I· '12 . . e- ott-) F(l.,J)= ce-f3li-i-N/211 i, j in same half of Yin i, j in different halves of Yin (5) with c < 1. We assume for ease of analysis that G is still as given in equation 4. This directly implements an intuitive notion put forward to account for the interdigitation of the ocular dominance mapping [4]: that the cortex tries to represent 334 G. J. GOODHILL, S. FINCH, TJ. SEJNOWSKI similar inputs close together, that similarity is given by the degree of correlation between the activities of points (cells), and additionally that natural visual scenes impose a correlational structure of the same qualitative form as equation 5. We now calculate C analytically for various mappings (c.f. [5]), and compare the cost of a map that keeps the two halves of Yin entirely separate in Vout to those which interdigitate the two halves of Yin with some regular periodicity. The map of the first type we consider will be refered to as the "up and down" map: moving from one end of Vout to the other implies moving entirely through one half of ViT\l then back in the opposite direction through the other half. For this map, the cost Cud is given by Cud = 2(N - l)e- ct + c. (6) For an interdigitated (striped) map where the stripes are of width n ~ 2: Cs(n) = N [2 (1 - ~) e- ct + ~ (e-~f(n) + e-~g(n))] (7) where for n even f(n) = g(n) = (n"22)2 and for n odd f(n) = (n"2I)2, g(n) = (n"23 ) 2. To characterize this system we now analyze how the n for which C s ( n) has a local maximum varies with c, a., 13, and when this local maximum is also a global maximum. Setting dCci£n) = 0 does not yield analytically tractable expressions (unlike [5]). However, more direct methods can be used: there is a local maximum atnifCs(n-1) < Cs(n) > Cs(n+ 1). Using equation 7we derive conditions on C for this to be true. For n odd, we obtain the condition CI < C < C2 where CI = C2; that is, there are no local maxima at odd values of n. For n even, we also obtain CI < C < C2 where now 2e- ct CI = n-4 2 n-2 2 ne-~(-z) - (n 2)e-~(-z) and c2(n) = CI (n + 2). CI (n) and c2(n) are plotted in figure 2, from which one can see the ranges of C for which particular n are local maxima. As 13 increases, maxima for larger values of n become apparent, but the range of c for which they exist becomes rather small. It can be shown that Cud is always the global maximum, except when e- ct > c, when n = 2 is globally optimal. As C decreases the optimal stripe width gets wider, analogously to k increasing in the dissimilarities given by equation 3. When 13 is such that there is no local maximum the only optimum is stripes as wide as possible. This fits with the intuitive idea that if corresponding points in the two halves of Yin (Le. Ii - j I = N/2) are sufficiently similar then it is favorable to interdigitate the two halves in VoutJ otherwise the two halves are kept completely separate. The qualitative behavior here is similar to that for equation 3. n = 2 is a global optimum for large c (small k), then as C decreases (k increases) n = 2 first becomes a local optimum, then the position of the local optimum shifts to larger n. However, ~n important difference is that in equation 3 the dissimilarities increase without limit with distance, whereas in equation 5 the similarities tend to zero with distance. Thus for equation 5 the extra cost of stripes one unit wider rapidly becomes negligible, whereas for equation 3 this extra cost keeps on increasing by ever larger amounts. As n -+ 00, Cud'" Cs(n) for the similarities defined by equation 5 (i.e. there is the same cost for traversing the two blocks in the same direction as in the opposite direction), whereas for the dissimilarities defined by equation 3 there is a quite different cost in these two cases. That F and G should tend to a bounded value as i and j become ever more distant neighbors seems biologically more plausible than that they should be potentially unbounded. Optimizing Cortical Mappings (a) '" 1.0 "" ~ 0.9 <.> 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0.0 0 2 · · • · • · · o D cl c:2 4 6 8 10 12 14 n (b) '" 1.0 .t< ~ 0.9 <.> 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0.0 335 , • , , . , . • , , , • . . . ...... ,' cl c:2 0 n Figure 2: The ranges of c for which particular n are local maxima. (a) oc = f3 = 0.25. (b) oc = 0.25, i3 = 0.1. When the Cl (dashed) line is below the c, (solid) line no local maxima exist. For each (even) value of n to the left of the crossing point, the vertical range between the two lines gives the values of c for which that n is a local maximum. Below the solid line and to the right of the crossing point the only maximum is stripes as wide as possible. Issues such as those we have addressed regarding the transition from "striped" to "blocked" solutions for combining two sets of inputs distinguished by their intraand inter-population similarity structure may be relevant to understanding the spatial representation of functional attributes across cortex. The results suggest the hypothesis that two variables are interdigitated in the same area rather than being represented separately in two distinct areas if the inter-population similarity is sufficiently high. An interesting point is that the striped solutions are often only local optima. It is possible that in reality developmental constraints (e.g. a chemically defined bias towards overlaying the two projections) impose a bias towards finding a striped rather than blocked solution, even though the latter may be the global optimum. 5 DISCUSSION We have argued that, in order to understand the structure of mappings in the brain, it could be useful to examine more general measures of similarity and of topographic matching than those implied by standard feature space models. The consequences of one particular alternative set of choices has been examined for the case of an interdigitated map of two variables. Many alternative objective functions for topographic matching are of course possible; this topic is reviewed in [3]. Two issues we have not discussed are the most appropriate way to define the features of interest, and the most appropriate measures of similarity between features (see [10] for an interesting discussion). A next step is to apply these methods to more complex structures in VI than just the ocular dominance map. By examining more of the space of possibilities than that occupied by the current feature space models, we hope to understand more about the optimization strategies that might be being pursued by the cortex. Feature space models may still tum out to be more or less the right answer; however even if this is true, our approach will at least give a deeper level of understanding why. 336 G. 1. GOODHILL, S. FINCH, T.l. SEINOWSKI Acknowledgements We thank Gary Blasdel, Peter Dayan and Paul Viola for stimulating discussions. References [1] Durbin, R. & Mitchison, G. (1990). A dimension reduction framework for understanding cortical maps. Nature, 343, 644-647. [2] Durbin, R. & Willshaw, D.J. (1987). An analogue approach to the travelling salesman problem using an elastic net method. Nature, 326,689-691. [3] Goodhill, G. J., Finch, S. & Sejnowski, T. J. (1995). Quantifying neighbourhood preservation in topographic mappings. Institute for Neural Computation Technical Report Series, No. INC-9505, November 1995. Available from ftp:/ / salk.edu/pub / geoff/ goodhillJinch_sejnowski_tech95.ps.Z or http://cnl.salk.edu/ ""geoff. [4] Goodhill, G.J. & Lowel, S. (1995). Theory meets experiment: correlated neural activity helps determine ocular dominance column periodicity. Trends in Neurosciences, 18,437-439. [5] Goodhill, G.J. & Willshaw, D.J. (1990). Application of the elastic net algorithm to the formation of ocular dominance stripes. Network, 1, 41-59. [6] Kohonen, T. (1982). Self-organized formation of topologically correct feature maps. Bioi. Cybern., 43, 59-69. [7] Luttrell, S.P. (1990). Derivation of a class of training algorithms. IEEE Trans. Neural Networks, 1,229-232. [8] Miller, KD., Keller, J.B. & Stryker, M.P. (1989). Ocular dominance column development: Analysis and simulation. Science, 245, 605-615. [9] Obermayer, K, Blasdel, G.G. & Schulten, K (1992). Statistical-mechanical analysis of self-organization and pattern formation during the development of visual maps. Phys. Rev. A, 45, 7568-7589. [10] Weiss, Y. & Edelman, S. (1995). Representation of similarity as a goal of early sensory coding. Network, 6, 19-41.
|
1995
|
106
|
1,006
|
Experiments with Neural Networks for Real Time Implementation of Control P. K. Campbell, M. Dale, H. L. Ferra and A. Kowalczyk Telstra Research Laboratories 770 Blackburn Road Clayton, Vic. 3168, Australia {p.campbell, m.dale, h.ferra, a.kowalczyk}@trl.oz.au Abstract This paper describes a neural network based controller for allocating capacity in a telecommunications network. This system was proposed in order to overcome a "real time" response constraint. Two basic architectures are evaluated: 1) a feedforward network-heuristic and; 2) a feedforward network-recurrent network. These architectures are compared against a linear programming (LP) optimiser as a benchmark. This LP optimiser was also used as a teacher to label the data samples for the feedforward neural network training algorithm. It is found that the systems are able to provide a traffic throughput of 99% and 95%, respectively, of the throughput obtained by the linear programming solution. Once trained, the neural network based solutions are found in a fraction of the time required by the LP optimiser. 1 Introduction Among the many virtues of neural networks are their efficiency, in terms of both execution time and required memory for storing a structure, and their practical ability to approximate complex functions. A typical drawback is the usually "data hungry" training algorithm. However, if training data can be computer generated off line, then this problem may be overcome. In many applications the algorithm used to generate the solution may be impractical to implement in real time. In such cases a neural network substitute can become crucial for the feasibility of the project. This paper presents preliminary results for a non-linear optimization problem using a neural network. The application in question is that of capacity allocation in an optical communications network. The work in this area is continuing and so far we have only explored a few possibilities. 2 Application: Bandwidth Allocation in SDH Networks Synchronous Digital Hierarchy (SDH) is a new standard for digital transmission over optical fibres [3] adopted for Australia and Europe equivalent to the SONET (Synchronous Optical NETwork) standard in North America. The architecture of the particular SDH network researched in this paper is shown in Figure 1 (a). 1) Nodes at the periphery of the SDH network are switches that handle individual calls. 974 P. CAMPBELL, M. DALE, H. L. FERRA, A. KOWALCZYK 2) Each switch concentrates traffic for another switch into a number of streams. 3) Each stream is then transferred to a Digital Cross-Connect (DXC) for switching and transmission to its destination by allocating to it one of several alternative virtual paths. The task at hand is the dynamic allocation of capacities to these virtual paths in order to maximize SDH network throughput. This is a non-linear optimization task since the virtual path capacities and the constraints, i.e. the physical limit on capacity of links between DXC's, are quantized, and the objective function (Erlang blocking) depends in a highly non-linear fashion on the allocated capacities and demands. Such tasks can be solved 'optimally' with the use of classical linear programming techniques [5], but such an approach is time-consuming - for large SDH networks the task could even require hours to complete. One of the major features of an SDH network is that it can be remotely reconfigured using software controls. Reconfiguration of the SDH network can become necessary when traffic demands vary, or when failures occur in the DXC's or the links connecting them. Reconfiguration in the case of failure must be extremely fast, with a need for restoration times under 60 ms [1]. o DXC (Digital ® Switch Cross-Connect) Figure 1 (b) link offered capacities traffic output: path capacities synaptic weights (22302) hidden units: 'AND' gates (l10) thresholds (738,67 used) input (a) Example of an Inter-City SDH/SONET Network Topology used in experiments. (b) Example of an architecture of the mask perceptron generated in experiments. In our particular case, there are three virtual paths allocated between any pair of switches, each using a different set of links between DXC's of the SDH network. Calls from one switch to another can be sent along any of the virtual paths, leading to 126 paths in total (7 switches to 6 other switches, each with 3 paths). The path capacities are normally set to give a predefined throughput. This is known as the "steady state". If links in the SDH network become partially damaged or completely cut, the operation of the SDH network moves away from the steady state and the path capacities must be reconfigured to satisfy the traffic demands subject to the following constraints: (i) Capacities have integer values (between 0 and 64 with each unit corresponding to a 2 Mb/s stream, or 30 Erlangs), (ii) The total capacity of all virtual paths through anyone link of the SDH network Experiments with Neural Networks for Real Time Implementation of Control 975 cannot exceed the physical capacity of that link. The neural network training data consisted of 13 link capacities and 42 traffic demand values, representing situations in which the operation of one or more links is degraded (completely or partially). The output data consisted of 126 integer values representing the difference between the steady state path capacities and the final allocated path capacities. 3 Previous Work The problem of optimal SDH network reconfiguration has been researched already. In particular Gopal et. al. proposed a heuristic greedy search algorithm [4] to solve this nonlinear integer programming problem. Herzberg in [5] reformulated this non-linear integer optimization problem as a linear programming (LP) task, Herzberg and Bye in [6] investigated application of a simplex algorithm to solve the LP problem, whilst Bye [2] considered an application of a Hopfield neural network for this task, and finally Leckie [8] used another set of AI inspired heuristics to solve the optimization task. All of these approaches have practical deficiencies; the linear programming is slow, while the heuristic approaches are relatively inaccurate and the Hopfield neural network method (simulated on a serial computer) suffers from both problems. In a previous paper Campbell et al. [10] investigated application of a mask perceptron to the problem of reconfiguration for a "toy" SDH network. The work presented here expands on the work in that paper, with the idea of using a second stage mask perceptron in a recurrent mode to reduce link violationslunderutilizations. 4 The Neural Controller Architecture Instead of using the neural network to solve the optimization task, e.g. as a substitute for the simplex algorithm, it is taught to replicate the optimal LP solution provided by it. We decided to use a two stage approach in our experiments. For the first stage we developed a feedforward network able to produce an approximate solution. More precisely, we used a collection of 2000 random examples for which the linear programming solution of capacity allocations had been pre-computed to develop a feedforward neural network able to approximate these solutions. Then, for a new example, such an "approximate" neural network solution was rounded to the nearest integer, to satisfy constraint (i), and used to seed the second stage providing refinement and enforcement of constraint (ii). For the second stage experiments we initially used a heuristic module based on the Gopal et al. approach [4]. The heuristic firstly reduces the capacities assigned to all paths which cause a physical capacity violation on any links, then subsequently increases the capacities assigned to paths across links which are being under-utilized. We also investigated an approach for the second stage which uses another feedforward neural network. The teaching signal for the second stage neural network is the difference between the outputs from the first stage neural network alone and the combined first stage neural networkiheuristic solution. This time the input data consisted of 13 link usage values (either a link violation or underutilization) and 42 values representing the amount of traffic lost per path for the current capacity allocations. The second stage neural network had 126 outputs representing the correction to the first stage neural network's outputs. The second stage neural network is run in a recurrent mode, adjusting by small steps the currently allocated link capacities, thereby attempting to iteratively move closer to the combined neural-heuristic solution by removing the link violations and under-utilizations left behind by the first stage network. The setup used during simulation is shown in Figure 2. For each particular instance tested the network was initialised with the solution from the first stage neural network. The offered traffic (demand) and the available maximum link capacities were used to determine the extent of any link violations or underutilizations as well as the amount of lost traffic (demand satisfaction). This data formed the initial input to the second stage network. The outputs of the neural network were then used to check the quality of the 976 P. CAMPBELL, M. DALE, H. L. FERRA, A. KOWALCZYK solution, and iteration continued until either no link violations occurred or a preset maximum number of iterations had been performed. offered traffic link capacities computation of constraint -demand satisfaction [ ........ ~ ........ -..... -----~(+) ! solution (t-l) ! I ! initialization: solution (0) from stage 1 correction (t) demand satisfaction (t-l 42 inputs link capacities violation!underutilization (t-l) 13 inputs Figure 2. Recurrent Network used for second stage experiments. solution (t) When computing the constraint satisfaction the outputs of the neural network where combined and rounded to give integer link violations/under-utilizations. This means that in many cases small corrections made by the network are discarded and no further improvement is possible. In order to overcome this we introduced a scheme whereby errors (link violations/under-utilizations) are occasionally amplified to allow the network a chance of removing them. This scheme works as follows: 1) an instance is iterated until it has either no link violations or until 10 iterations have been performed; 2) if any link violations are still present then the size of the errors are multiplied by an amplification factor (> 1); 3) a further maximum of 10 iterations are performed; 4) if subsequently link violations persist then the amplification factor is increased; the procedure repeats until either all link violations are removed or the amplification factor reaches some fixed value. S Description of Neural Networks Generated The first stage feedforward neural network is a mask perceptron [7], c.f. Figure 1 (b). Each input is passed through a number of arbitrarily chosen binary threshold units. There were a total of 738 thresholds for the 55 inputs. The task for the mask perceptron training algorithm [7] is to select a set of useful thresholds and hidden units out of thousands of possibilities and then to set weights to minimize the mean-square-error on the training set. The mask perceptron training algorithm automatically selected 67 of these units for direct connection to the output units and a further 110 hidden units ("AND" gates) whose Experiments with Neural Networks for Real Time Implementation of Control 977 outputs are again connected to the neural network outputs, giving 22,302 connections in all. Such neural networks are very rapid to simulate since the only operations required are comparison and additions. For the recurrent network used in the second stage we also used a mask perceptron. The training algori thIn used for the recurrent network was the same as for the first stage, in particular note that no gradual adaptation was employed. The inputs to the network are passed through 589 arbitrarily chosen binary threshold units. Of these 35 were selected by the training algorithm for direct connection to the output units via 4410 weighted links. 6 Results The results are presented in Table 1 and Figure 3. The values in the table represent the traffic throughput of the SDH network, for the respective methods, as a percentage of the throughput determined by the LP solution. Both the neural networks were trained using 2000 instances and tested against a different set of 2000 instances. However for the recurrent network approximately 20% of these cases still had link violations after simulation so the values in Table 1 are for the 80% of valid solutions obtained from either the training or test set. Solution type Training Test Feedforward Net/Heuristic 99.08% 98.90%, Feedforward Net/Recurrent Net 94.93% (*) 94.76%(*) Gopal-S 96.38% 96.20% Gopal-O 85.63% 85.43% (*) these numbers are for the 1635 training and 1608 test instances (out of 2000) for which the recurrent network achieved a solution with no link violations after simulation as described in Section 3. Table 1. Efficiency of solutions measured by average fraction of the 'optimal' throughput of the LP solution As a comparison we implemented two solely heuristic algorithms. We refer to these as Gopal-S and Gopal-O. Both employ the same scheme described earlier for the Gopal et al. heuristic. The difference between the two is that Gopal-S uses the steady state solution as an initial starting point to determine virtual path capacities for a degraded network, whereas Gopal-O starts from a point where all path capacities are initially set to zero. Referring to Figure 3, link capacity ratio denotes the total link capacity of the degraded SDH network relative to the total link capacity of the steady state SDH network. A low value of link capacity ratio indicates a heavily degraded network. The traffic throughput ratio denotes the ratio between the throughput obtained by the method in question, and the throughput of the steady state solution. Each dot in the graphs in Figure 3 represents one of the 2000 test set cases. It is clear from the figure that the neural network/heuristic approach is able to find better solutions for heavily degraded networks than each of the other approaches. Overall the clustering of dots for the neural network/heuristic combination is tighter (in the y-direction) and closer to 1.00 than for any of the other methods. The results for the recurrent network are very encouraging being qUalitatively quite close to those for the Gopal-S algorithm. All experiments were run on a SPARCStation 20. The neural network training took a few minutes. During simulation the neural network took an average of 9 ms per test case with a further 36.5 ms for the heuristic, for a total of 45.5 ms. On average the Gopal-S algorithm required 55.3 ms and the Gopal-O algorithm required 43.7 ms per test case. The recurrent network solution required an average of 55.9 ms per test case. The optimal solutions calculated using the linear programming algorithm took between 2 and 60 seconds per case on a SPARCStation 10. 978 P. CAMPBELL, M. DALE, H. L. FERRA, A. KOWALCZYK Neural Network/Heuristic Recurrent Neural Network 1.00 .2 ~ 0.95 8. 0.90 .r: 0> is 0.85 .c t.!.! 0.80 ~ ~ 0.75 0.70 0.50 0.60 0.10 0.80 0.90 1.00 link Capacity Ratio 1.00 .2 ra 0.95 cr ~ 0.90 .r: 0> 6 0.85 .c t-,g 0.80 ~ ~ 0.75 Gopal-S ······:····:· · :,,~i~ffI~ .-. -,. " " - -' - "~':' ........ ~ ...... -... --- -... . 0.70 0.50 0.60 0.70 0.80 0.90 1.00 link Capacity Ratio •• , _ ._0 •• _ • •• • • • •• : • ••• :.' ••• :.~' •• : • •••• :. '0"" _ •• • •• _ •••••••• 0.70 0.50 0.60 0.70 0.80 0.90 1.00 Link Capacity Ratio 1.00 .2 r.; 0.95 cr ~ 0.90 .r: 0> 5 0.85 .c t. ~ 0.80 ~ ~ 0.75 Gopal-O 0.70 0.50 0.60 0.70 0.80 0.90 100 Link Capacily Ratio Figure 3. Experimental results for the Inter-City SDH network (Fig. 1) on the independent test set of 2000 random cases. On the x axis we have the ratio between the total link capacity of the degraded SDH network and the steady state SDH network. On the y axis we have the ratio between the throughput obtained by the method in question, and the throughput of the steady state solution. Fig 3. (a) shows results for the neural network combined with the heuristic second stage. Fig 3. (b) shows results for the recurrent neural network second stage. Fig 3. (c) shows results for the heuristic only, initialised by the steady state (Gopal-S) and Fig 3. (d) has the results for the heuristic initialised by zero (Gopal-O). 7 Discussion and Conclusions The combined neural network/heuristic approach performs very well across the whole range of degrees of SDH network degradation tested. The results obtained in this paper are consistent with those found in [10]. The average accuracy of -99% and fast solution generation times « ffJ ms) highlight this approach as a possible candidate for implementation in a real system, especially when one considers the easily achievable speed increase available from parallelizing the neural network. The mask perceptron used in these experiments is well suited for simulation on a DSP (or other hardware): the operations required are only comparisons, calculation of logical "AND" and the summation of synaptic weights (no multiplications or any non-linear transfonnations are required). The interesting thing to note is the relatively good perfonnance of the recurrent network, namely that it is able to handle over 80% of cases achieving very good perfonnance when compared against the neural network/heuristic solution (95% of the quality of the teacher). One thing to bear in mind is that the heuristic approach is highly tuned to producing a solution which satisfies the constraints, changing the capacity of one link at a time until the desired goal is achieved. On the other hand the recurrent network is generic and does not target the constraints in such a specific manner, making quite crude global changes in Experiments with Neural Networks for Real Time Implementation of Control 979 one hit, and yet is still able to achieve a reasonable level of performance. While the speed for the recurrent network was lower on average than for the heuristic solution in our experiments, this is not a major problem since many improvements are still possible and the results reported here are only preliminary, but serve to show what is possible. It is planned to continue the SOH network experiment in the future; with more investigation on the recurrent network for the second stage and also more complex SDH architectures. Acknowledgments The research and development reported here has the active support of various sections and individuals within the Telstra Research Laboratories (TRL), especially Dr. C. Leckie, Mr. P. Sember, Dr. M. Herzberg, Mr. A. Herschtal and Dr. L. Campbell. The permission of the Managing Director, Research and Information Technology, Telstra, to publish this paper is acknowledged. The research and development reported here has the active support of various sections and individuals within the Telstra Research Laboratories (TRL), especially Dr. C. Leckie and Mr. P. Sember who were responsible for the creation and trialling of the programs designed to produce the testing and training data. The SOH application was possible due to co-operation of a number of our colleagues in TRL, in particular Dr. L. Campbell (who suggested this particular application), Dr. M. Herzberg and Mr. A. Herschtal. The permission of the Managing Director, Research and Information Technology, Telstra, to publish this paper is acknowledged. References [1] E. Booker, Cross-connect at a Crossroads, Telephony, Vol. 215, 1988, pp. 63-65. [2] S. Bye, A Connectionist Approach to SDH Bandwidth Management, Proceedings of the 19th International Conference on Artificial Neural Networks (ICANN-93), Brighton Conference Centre, UK, 1993, pp. 286-290. [3] R. Gillan, Advanced Network Architectures Exploiting the Synchronous Digital Hierarchy, Telecommunications Journal of Australia 39, 1989, pp. 39-42. [4] G. Gopal, C. Kim and A. Weinrib, Algorithms for Reconfigurable Networks, Proceedings of the 13th International Teletraffic Congress (ITC-13), Copenhagen, Denmark, 1991, pp. 341-347. [5] M. Herzberg, Network Bandwidth Management - A New Direction in Network Management, Proceedings of the 6th Australian Teletraffic Research Seminar, Wollongong, Australia, pp. 218-225. [6] M. Herzberg and S. Bye, Bandwidth Management in Reconfigurable Networks, Australian Telecommunications Research 27, 1993, pp 57-70. [7] A. Kowalczyk and H.L. Ferra, Developing Higher Order Networks with Empirically Selected Units, IEEE Transactions on Neural Networks, pp. 698-711, 1994. [8] C. Leckie, A Connectionist Approach to Telecommunication Network Optimisation, in Complex Systems: Mechanism of Adaptation, R.J. Stonier and X.H. Yu, eds., lOS Press, Amsterdam, 1994. [9] M. Schwartz, Telecommunications Networks, Addison-Wesley, Readings, Massachusetts, 1987. [10] p. Campbell, H.L. Ferra, A. Kowalczyk, C. Leckie and P. Sember, Neural Networks in Real Time Decision Making, Proceedings of the International Workshop on Applications of Neural Networks to Telecommunications 2 (IWANNT-95), Ed. J Alspector et. al. Lawrence Erlbaum Associates, New Jersey, 1995, pp. 273-280.
|
1995
|
107
|
1,007
|
Cholinergic suppression of transmission may allow combined associative memory function and self-organization in the neocortex. Michael E. Hasselmo and Milos Cekic Department of Psychology and Program in Neurosciences, Harvard University, 33 Kirkland St., Cambridge, MA 02138 hasselmo@katIa.harvard.edu Abstract Selective suppression of transmission at feedback synapses during learning is proposed as a mechanism for combining associative feedback with self-organization of feed forward synapses. Experimental data demonstrates cholinergic suppression of synaptic transmission in layer I (feedback synapses), and a lack of suppression in layer IV (feedforward synapses). A network with this feature uses local rules to learn mappings which are not linearly separable. During learning, sensory stimuli and desired response are simultaneously presented as input. Feedforward connections form self-organized representations of input, while suppressed feedback connections learn the transpose of feedforward connectivity. During recall, suppression is removed, sensory input activates the self-organized representation, and activity generates the learned response. 1 INTRODUCTION The synaptic connections in most models of the cortex can be defined as either associative or self-organizing on the basis of a single feature: the relative infl uence of modifiable synapses on post-synaptic activity during learning (figure 1). In associative memories, postsynaptic activity during learning is determined by nonmodifiable afferent input connections, with no change in the storage due to synaptic transmission at modifiable synapses (Anderson, 1983; McNaughton and Morris, 1987). In self-organization, post-synaptic activity is predominantly influenced by the modifiable synapses, such that modification of synapses influences subsequent learning (Von der Malsburg, 1973; Miller et al., 1990). Models of cortical function must combine the capacity to form new representations and store associations between these representations. Networks combining self-organization and associative memory function can learn complex mapping functions with more biologically plausible learning rules (Hecht-Nielsen, 1987; Carpenter et al., 1991; Dayan et at., 132 M. E. HASSELMO, M. CEKIC 1995), but must control the influence of feedback associative connections on self-organization. Some networks use special activation dynamics which prevent feedback from influencing activity unless it coincides with feedforward activity (Carpenter et al., 1991). A new network alternately shuts off feedforward and feedback synaptic transmission (Dayan et al., 1995). A. c. Self-organizing Afferent Self-organizing feedforward Associative feedback Figure 1 - Defining characteristics of self-organization and associative memory. A. At self-organizing synapses, post-synaptic activity during learning depends predominantly upon transmission at the modifiable synapses. B. At synapses mediating associative memory function, post-synaptic activity during learning does not depend primarily on the modifiable synapses, but is predominantly influenced by separate afferent input. C. Selforganization and associative memory function can be combined if associative feedback synapses are selectively suppressed during learning but not recall. Here we present a model using selective suppression of feedback synaptic transmission during learning to allow simultaneous self-organization and association between two regions. Previous experiments show that the neuromodulator acetylcholine selectively suppresses synaptic transmission within the olfactory cortex (Hasselmo and Bower, 1992; 1993) and hippocampus (Hasselmo and Schnell, 1994). If the model is valid for neocortical structures, cholinergic suppression should be stronger for feedback but not feedforward synapses. Here we review experimental data (Hasselmo and Cekic, 1996) comparing cholinergic suppression of synaptic transmission in layers with predominantly feedforward or feedback synapses. 2. BRAIN SLICE PHYSIOLOGY As shown in Figure 2, we utilized brain slice preparations of the rat somatosensory neocortex to investigate whether cholinergic suppression of synaptic transmission is selective for feedback but not feedforward synaptic connections. This was possible because feedforward and feedback connections show different patterns of termination in neocortex. As shown in Figure 2, Layer I contains primarily feedback synapses from other cortical regions (Cauller and Connors, 1994), whereas layer IV contains primarily afferent synapses from the thalamus and feedforward synapses from more primary neocortical structures (Van Essen and Maunsell, 1983). Using previously developed techniques (Cauller and Connors, 1994; Li and Cauller, 1995) for testing of the predominantly feedback connections in layer I, we stimulated layer I and recorded in layer I (a cut prevented spread of Cholinergic Suppression of Transmission in the Neocortex 133 activity from layers II and III). For testing the predominantly feedforward connections terminating in layer IV, we elicited synaptic potentials by stimulating the white matter deep to layer VI and recorded in layer IV. We tested suppression by measuring the change in height of synaptic potentials during perfusion of the cholinergic agonist carbachol at lOOJ,1M. Figure 3 shows that perfusion of carbachol caused much stronger suppression of synaptic transmission in layer I as compared to layer IV (Hasselmo and Cekic, 1996), suggesting that cholinergic suppression of transmission is selective for feedback synapses and not for feedforward synapses. I White matter / stimulation 1/ Layer IV recording I ~~ I ll-ill ll-ill IV .1 Foedback IV V-VI V-VI Region 1 Region 2 Figure 2. A. Brain slice preparation of somatosensory cortex showing location of stimulation and recording electrodes for testing suppression of synaptic transmission in layer I and in layer IV. Experiment based on procedures developed by Cauller (Cauller and Connors, 1994; Li and Cauller, 1995). B. Anatomical pattern of feedforward and feedback connectivity within cortical structures (based on Van Essen and Maunsell, 1983). Feedforward -layer IV Control Carbachol (1 OOJlM) Wash I~ -0 5ms Feedback - layer I '!oi' Control Carbachol (1 OOJlM) Wash Figure 3 - Suppression of transmission in somatosensory neocortex. Top: Synaptic potentials recorded in layer IV (where feedforward and afferent synapses predominate) show little effect of l00J.tM carbachol. Bottom: Synaptic potentials recorded in layer I (where feedback synapses predominate) show suppression in the presence of lOOJ,1M carbachol. 134 M. E. HASSELMO, M. CEKIC 3. COMPUTATIONAL MODELING These experimental results supported the use of selective suppression in a computational model (Hasselmo and Cekic, 1996) with self-organization in its feedforward synaptic connections and associative memory function in its feedback synaptic connections (Figs 1 and 4). The proposed network uses local, Hebb-type learning rules supported by evidence on the physiology of long-tenn potentiation in the hippocampus (Gustafsson and Wigstrom, 1986). The learning rule for each set of connections in the network takes the fonn: tlWS:'Y) = 11 (a?) - 9(Y» g (ar» Where W(x. Y) designates the connections from region x to region y, 9 is the threshold of synaptic modification in region y, 11 is the rate of modification, and the output function is g(a;.(x~ = [tanh(~(x) - J.1(x~]+ where []+ represents the constraint to positive values only. Feedforward connections (Wi/x<y» have self-organizing properties, while feedback connections (Wir>=Y~ have associative memory properties. This difference depends entirely upon the selective suppression of feedback synapses during learning, which is implemented in the activation rule in the form (I-c). For the entire network, the activation rule takes the fonn: M II(X) N II(X) II(Y) a?) = A?) + 2, 2, Wi~<Y) g (a~x» + 2, 2, (1- c) Wi~~Y) g (a~x» - 2, Hi~) (g (af») x=lk=l x=lk=l k=l where a;.(y) represents the activity of each of the n(y) neurons in region y, ~ (x) is the activity of each of the n(x) neurons in other regions x, M is the total number of regions providing feedforward input, N is the total number of regions providing feedback input, Aj(y) is the input pattern to region y, H(Y) represents the inhibition between neurons in region y, and (1 - c) represents the suppression of synaptic transmission. During learning, c takes a value between 0 and 1. During recall, suppression is removed, c = O. In this network, synapses (W) between regions only take positive values, reflecting the fact that long-range connections between cortical regions consist of excitatory synapses arising from pyramidal cells. Thus, inhibition mediated by the local inhibitory interneurons within a region is represented by a separate inhibitory connectivity matrix H. After each step of learning, the total weight of synaptic connections is nonnalized pre-synaptically for each neuron j in each region: ~-------------Wij (t+l) = [Wij(t) + l1W;j(t)]I( .i [Wij(t) +l1Wij (t)] 2) 1= 1 Synaptic weights are then normalized post-synaptically for each neuron i in each region (replacing i with j in the sum in the denominator in equation 3). This nonnalization of synaptic strength represents slower cellular mechanisms which redistribute pre and postsynaptic resources for maintaining synapses depending upon local influences. In these simulations, both the sensory input stimuli and the desired output response to be learned are presented as afferent input to the neurons in region 1. Most networks using error-based learning rules consist of feedforward architectures with separate layers of input and output units. One can imagine this network as an auto-encoder network folded back on itself, with both input and output units in region 1, and hidden units in region 2. Cholinergic Suppression of Transmission in the Neocortex 135 As an example of its functional properties, the network presented here was trained on the XOR problem. The XOR problem has previously been used as an example of the capability of error based training schemes for solving problems which are not linearly separable. The specific characteristics of the network and patterns used for this simulation are shown in figure 4. The two logical states of each component of the XOR problem are represented by two separate units (designated on or off in figures 4 and 5), ensuring that activation of the network is equal for each input condition. The problem has the appearance of two XOR problems with inverse logical states being solved simultaneously. As shown in figure 4, the input and desired output of the network are presented simultaneously during learning to region 1. The six neurons in region 1 project along feedforward connections to four neurons in region 2, the hidden units of the network. These four neurons project along feedback connections to the six neurons in region 1. All connections take random initial weights. During learning, the feedforward connections undergo self-organization which ultimately causes the hidden units to become feature detectors responding to each of the four patterns of input to region 1. Thus, the rows of the feedforward synaptic connectivity matrix gradually take the form of the individual input patterns. STIMULUS RESPONSE on off on off yes no 1. oeeo eo 2. oeoe oe Afferent eooe eo input 3. 4. 9 9 9 "'-Region 1 Figure 4 - Network for learning the XOR problem, with 6 units in region 1 and 4 units in region 2. Four different patterns of afferent input are presented successively to region 1. The input stimuli of the XOR problem are represented by the four units on the left, and the desired output designation of XOR or not-XOR is represented by the two units on the right. The XOR problem has four basic states: on-off and off-on on the input is categorized by yes on the output, while on-on and off-off on the input is categorized by no on the output. Modulation is applied during learning in the form of selective suppression of synaptic transmission along feedback connections (this suppression need not be complete), giving these connections associative memory function. Hebbian synaptic modification causes these connections to link each of the feature detecting hidden units in region 2 with the cells in region 1 activated by the pattern to which the hidden unit responds. Gradually, the feedback synaptic connectivity matrix becomes the transpose of the feedforward connectivity matrix. (parameters used in simulation: Aj(l) = 0 or I, h = 2.0, q(l) = 0.5, q(2) = 0.6, (1) = 0.2, (2) = 0.5, c = 1.0 and Hik(2) = 0.6). Function was similar and convergence was obtained more rapidly with c = 0.5. Feedback synaptic transmission prevented con136 M. E. HASSELMO. M. CEKIC vergence during learning when c = 0.367). During recall, modulation of synaptic transmission is removed, and the various input stimuli of the XOR problem are presented to region 1 without the corresponding output pattern. Activity spreads along the self-organized feedforward connections to activate the specific hidden layer unit responding to that pattern. Activity then spreads back along feedback connections from that particular unit to activate the desired output units. The activity in the two regions settles into a final pattern of recall. Figure 5 shows the settled recall of the network at different stages of learning. It can be seen that the network initially may show little recall activity, or erroneous recall activity, but after several cycles of learning, the network settles into the proper response to each of the XOR problem states. Convergence during learning and recall have been obtained with other problems, including recognition of whether on units were on the left or right, symmetry of on units, and number of on units. In addition, larger scale problems involving multiple feedforward and feedback layers have been shown to converge. 1. 2. 3. 4. ~ on on no off off no off on yes on off yes -- --- - --- §L 3 -- -R ----- - ------- - - -- -· · . -----. · · :-=.:: - · - ---- - - · - -· · . 11----. · · _ .. ---- · - .-- . - - · - -· · --- --· · - · - -- - - · - --· - ----- · · -- ---- · - --- - - · - -. · - ----- ::=:: - -- --- - - · -- -. · - 11---- 1 - - =-- - - -- -. · - ---- ----- - - .- - - .: ••• - 11:11:: :11-== - - .-- - -. ----- - - .- - - - -- · - ----- - -- -- - -- · - ----- ----- - - -.- -- - -- · - ----- · ----- - --.--- - -- - - --.- • • > ----- - - ---- • .: - .~.: - -:==~: - - ---0 - ---- - -- - -~. =-=: --:11 := -- - -- ----- • .: - • • < -. -• • - .... -< I. - - - - =: - - - - --- - --• - - ---- - -- -- -- -- --- - - =: --- -- - - --- - - -- -- -• • - - • .: - • • •• -- -- --- - - =: --- :=. -- ---- --- -• • -- • .: - • • •• -- - = -- - - -- --- == - -- - - -- --- -- --- - - -- -= -- -- --- - - -- -- - • :. - - • .: - • • •• ---= -- - - =: --= :== - .- - - --- -.- -- - -- --- 26 -- - - --- - - -- --- en :;0 I ~. 0 t 3 ~ Region 1 Region 2 Figure 5 - Output neuronal activity in the network shown at different learning steps. The four input patterns are shown at top. Below these are degraded patterns presented during recall, missing the response components of the input pattern. The output of the 6 region 1 units and the 4 region 2 units are shown at each stage of learning. As learning progresses, gradually one region 2 unit starts to respond selectively to each input pattern, and the correct output unit becomes active in response to the degraded input. Note that as learning progresses the response to pattern 4 changes gradually from incorrect (yes) to correct (no). Cholinergic Suppression of Transmission in the Neocortex 137 References Anderson, 1.A. (1983) Cognitive and psychological computation with neural models. IEEE Trans. Systems, Man, Cybem. SMC-13,799-815. Carpenter, G.A., Grossberg, S. and Reynolds, 1.H. (1991) ARTMAP: Supervised realtime learning and classification of nonstationary data by a self-organizing neural network. Neural Networks 4: 565-588. Cauller, LJ. and Connors, B.W. (1994) Synaptic physiology of horizontal afferents to layer I in slices of rat SI neocortex. 1. Neurosci. 14: 751-762. Dayan, P., Hinton, G.E., Neal, RM. and Zemel, RS. (1995) The Helmholtz machine. Neural Computation. Gustafsson, B. and Wigstrom, H. (1988) Physiological mechanisms underlying long-term potentiation. Trends Neurosci. 11: 156-162. Hasselmo, M.E. (1993) Acetylcholine and learning in a cortical associative memory. Neural Computation. 5(1}: 32-44. Hasselmo M.E. and Bower 1.M. (1992) Cholinergic suppression specific to intrinsic not afferent fiber synapses in rat piriform (olfactory) cortex. 1. Neurophysiol. 67: 1222-1229. Hasselmo, M.E. and Bower, 1.M. (1993) Acetylcholine and Memory. Trends Neurosci. 26: 218-222. Hasselmo, M.E. and Cekic, M. (1996) Suppression of synaptic transmission may allow combination of associative feedback and self-organizing feed forward connections in the neocortex. Behav. Brain Res. in press. Hasselmo M.E., Anderson B.P. and Bower 1.M. (1992) Cholinergic modulation of cortical associative memory function. 1. Neurophysiol. 67: 1230-1246. Hasselmo M.E. and Schnell, E. (1994) Laminar selectivity of the cholinergic suppression of synaptic transmission in rat hippocampal region CAl: Computational modeling and brain slice physiology. 1. Neurosci. 15: 3898-3914. Hecht-Nielsen, R (1987) Counterpropagation networks. Applied Optics 26: 4979-4984. Li, H. and Cauller, L.l. (1995) Acetylcholine modulation of excitatory synaptic inputs from layer I to the superficial layers of rat somatosensory neocortex in vitro. Soc. Neurosci. Abstr. 21: 68. Linsker, R (1988) Self-organization in a perceptual network. Computer 21: 105-117. McNaughton B.L. and Morris RG.M. (1987) Hippocampal synaptic enhancement and information storage within a distributed memory system. Trends in Neurosci. 10:408-415. Miller, K.D., Keller, 1.B. and Stryker, M.P. (1989) Ocular dominance column development Analysis and simulation. Science 245: 605-615. van Essen, D.C. and Maunsell, 1.H.R. (1983) Heirarchical organization and functional streams in the visual cortex. Trends Neurosci. 6: 370-375. von der Malsburg, C. (1973) Self-organization of orientation sensitive cells in the striate cortex. Kybemetik 14: 85-100.
|
1995
|
108
|
1,008
|
Dynamics of Attention as Near Saddle-Node Bifurcation Behavior Hiroyuki Nakahara" General Systems Studies U ni versi ty of Tokyo 3-8-1 Komaba, Meguro Tokyo 153, Japan nakahara@vermeer.c.u-tokyo.ac.jp Kenji Doya ATR Human Information Processing Research Laboratories 2-2 Hikaridai, Seika, Soraku Kyoto 619-02, Japan doya@hip.atr.co.jp Abstract In consideration of attention as a means for goal-directed behavior in non-stationary environments, we argue that the dynamics of attention should satisfy two opposing demands: long-term maintenance and quick transition. These two characteristics are contradictory within the linear domain. We propose the near saddlenode bifurcation behavior of a sigmoidal unit with self-connection as a candidate of dynamical mechanism that satisfies both of these demands. We further show in simulations of the 'bug-eat-food' tasks that the near saddle-node bifurcation behavior of recurrent networks can emerge as a functional property for survival in nonstationary environments. 1 INTRODUCTION Most studies of attention have focused on the selection process of incoming sensory cues (Posner et al., 1980; Koch et al., 1985; Desimone et al., 1995). Emphasis was placed on the phenomena of causing different percepts for the same sensory stimuli. However, the selection of sensory input itself is not the final goal of attention. We consider attention as a means for goal-directed behavior and survival of the animal. In this view, dynamical properties of attention are crucial. While attention has to be maintained long enough to enable robust response to sensory input, it also has to be shifted quickly to a novel cue that is potentially important. Long-term maintenance and quick transition are critical requirements for attention dynamics. ·currently at Dept. of Cognitive Science and Institute for Neural Computation, U. C. San Diego, La Jolla CA 92093-0515. hnakahar@cogsci.ucsd.edu Dynamics of Attention as Near Saddle-node Bifurcation Behavior 39 We investigate a possible neural mechanism that enables those dynamical characteristics of attention. First, we analyze the dynamics of a network of sigmoidal units with self-connections. We show that both long-term maintenance and quick transition can be achieved when the system parameters are near a "saddle-node bifurcation" point. Then, we test if such a dynamical mechanism can actually be helpful for an autonomously behaving agent in simulations of a 'bug-eat-food' task. The result indicates that near saddle-node bifurcation behavior can emerge in the course of evolution for survival in non-stationary environments. 2 NEAR SADDLE-NODE BIFURCATION BEHAVIOR When a pulse-like input is given to a linear system, the rising and falling phases of the response have the same time constants. This means that long-term maintenance and quick transition cannot be simultaneously achieved by linear dynamics. Therefore, it is essential to consider a nonlinear dynamical mechanism to achieve these two demands. 2.1 DYNAMICS OF A SELF-RECURRENT UNIT First, we consider the dynamics of a single sigmoidal unit with the self-connection weight a and the bias b. y(t + 1) F(x) F(ay(t) + b) , 1 1 + exp( -x)' (1) (2) The parameters (a, b) determine the qualitative behavior of the system such as the number of fixed points and their stabilities. As we change the parameters, the qualitative behavior of the system may suddenly change. This is referred to as "bifurcation" (Guckenheimer, et al., 1983). A typical example is a "saddle-node bifurcation" in which a pair of fixed points, one stable and one unstable, emerges. In our system, this occurs when the state transition curve y(t + 1) = F(ay(t) + b) is tangent to y(t + 1) = y(t). Let y* be this point of tangency. We have the following condi tion for saddle-node bifurcation. F(ay* + b) dF(ay + b) I dy y=y. y* 1 These equations can be solved, by noting F'(x) = F(x)(l- F(x)), as a b = 1 y* (1 - y*) 1 F-1(y*) - ay* = F-l(y*) - -I- y* (3) ( 4) (5) (6) By changing the fixed point value y* between a and 1, we can plot a curve in the parameter space (a, b) on which saddle-node bifurcation occurs, as shown in Figure 1 (left). A pair of a saddle point and a stable fixed point emerges or disappears when the parameters pass across the cusp like curve (cases 2 and 4) . The system has only one stable fixed point when the parameters are outside the cusp (case 1) and three fixed points inside the cusp (case 3). 40 H.NAKAHARA,K.DOYA y ( t+ l ) CASE 1 y(t +1 J CASE 2 " b Bifurcation Diagr am C. 8f::.x.d Pt •. , ~' 1 ~ 8:tixed pts,' 06, ", 0 61 ,,' 041 " 0 "i " O~/ ../ 0'1/ H 0.20. 40.60 8 ly(tJ ""0~ 1O:".,,,,0 ';;;-0;;-; 8 lYlt) ., -'0 Y'~tL · " CASE 3 Y',t." CASE • . ' . o. a: ,,/ 0 B " 't1x.d p t ., "' 3 f lxed p ts " . 2 (I 6 ' 0 6 ' , . OJ' '' 0 " ,,' , . o ,/ 0 2: ' '0 20 40 60 8 1 y( t ) 0 2(' 4~ 60 8: 1 y (tJ - 15 - lO Figure 1: Bifurcation Diagram of a Self-Recurrent Unit. Left: the curve in the parameter space (a , b) on which saddle-node bifurcation is seen. Right: state transition diagrams for four different cases. y ( t · l) o'~=Lil l. 1111 b = - 7. 9 0.6 I 0. 4 " 0 .2 I a ... " y et) 0.20. 40 . 60. 81 y et ) o .~ o. o. o. o 5 10 15 2(fi me (t) Y1d (t:l)ll.ll11 b = ~9 o. I O. I O. " I O. I , o yet) o . 20 . 4 0 60. 8 1 yet) dL o. & 0 . 41 0.21 o 5 10 l S 2a:'irne l t ) Figure 2: Temporal Responses of Self-Recurrent Units. Left: near saddle-node bifurcation. Right: far from bifurcation. An interesting behavior can be seen when the parameters are just outside the cusp, as shown in Figure 2 (left). The system has only one fixed point near Y = 0, but once the unit is activated (y ~ 1), it stays "on" for many time steps and then goes back to the fixed point quickly. Such a mechanism may be useful in satisfying the requirements of attention dynamics: long-term maintenance and quick transition. 2.2 NETWORK OF SELF-RECURRENT UNITS Next, we consider the dynamics of a network of the above self-recurrent units. Yi(t + 1) = F[aYi(t) + b + L CijYj(t) + diUi(t)], (7) j,jti where a is the self connection weight , b is the bias, Cij is the cross connection weight, and di is the input connection weight, and Ui(t) is the external input. The effect of lateral and external inputs is equivalent to the change in the bias, which slides the sigmoid curve horizontally without changing the slope. For example, one parameter set of the bifurcation at y* = 0.9 is a = 11.11 and b ~ -7.80. Let b = -7.90 so that the unit has a near saddle-node bifurcation behavior when there is no lateral or external inputs. For a fixed a = 11.11, as we increase b, the qualitative behavior of the system appears as case 3 in Figure 1, and Dynamics of Attention as Near Saddle-node Bifurcation Behavior 41 Sensory Inputs Actions Network Structure , , 'olr lood " - '--,~ ,- \, :/~ - - - . r-.~ "==:-:"~'on:'oocl ~~ ....... ~~ .... ... ..... Creature ... Creature Inpol. IJrI .111'2 Figure 3: A Creature's Sensory Inputs(Left), Motor System(Center) and Network Architecture(Right) then, it changes again at b:::::: -3.31, where the fixed point at Y = 0.1, or another bifurcation point , appears as case 4 in Figure L Therefore, ifthe input sum is large enough, i.e. Lj ,j;Ci CijYj + diuj > -3.31- (-7.90) :::::: 4.59, the lower fixed point at Y = 0.1 disappears and the state jumps up to the upper fixed point near Y = 1, quickly turning the unit "on". If the lateral connections are set properly, this can in turn suppress the activation of other units. Once the external input goes away, as we see in Figure 2 (left), the state stays "on" for a long time until it returns to the fixed point near Y = O. 3 EVOLUTION OF NEAR BIFURCATION DYNAMICS In the above section, we have theoretically shown the potential usefulness of near saddle-node bifurcation behavior for satisfying demands for attention dynamics. We further hypothesize that such behavior is indeed useful in animal behaviors and can be found in the course of learning and evolution of the neural system. To test our hypothesis, we simulated a 'bug-eat-food' task. Our purpose in t.his simulation was to see whether the attention dynamics discussed in the previous section would help obtain better performance in a non-stationary environment. Vve used evolutionary programming (Fogel et aI, 1990) to optimize the performance of recurrent networks and feedforward networks. 3.1 THE BUG AND THE WORLD In our simulation, a simple creature traveled around a non-stationary environment. In the world, there were a certain number of food items. Each item was fixed at a certain place in the world but appeared or disappeared in a stochastic fashion, as determined by a two-state Markov system. In order to survive, A creature looked for food by traveling the world . The amount of food a creature found in a certain time period was the measure of its performance. A creature had five sensory inputs, each of which detected food in the sector of 45 degrees (Figure 3, right). Its output level was given by L J' .l.., where Tj ,"vas the r J distance to the j-th food item within the sector. Note that the format of the input contained information about distance and also that the creature could only receive the amount of the input but could not distinguish each food from others. For the sake of simplicity, we assumed that the creature lived in a grid-like world. On each time step, it took one of three motor commands: L: turn left (45 degrees), 42 H. NAKAHARA, K. DOYA Density of Food 0.05 0.10 Markov Transition Matrix .5 .5 .8 .8 .5 .5 .8 .8 of each food .5 .5 .2 .2 .5 .5 .2 .2 Random Walk 7.0 6.9 13.8 13.9 Nearest Visible 42.7 18.6 65.3 32.4 FeedForward 58.6 37.3 84.8 60.0 Recurrent 65.7 43.6 94.0 66.1 Nearest Visible/Invisible 97.7 97.1 129.1 128.8 Table 1: Performances of the Recurrent Network and Other Strategies. C: step forward, and R: turn right (Figure 3, center). Simulations were run with different Markov transition matrices of food appearance and with different food densities. A creature got the food when it reached the food, whether it was visible or invisible. When a creature ate a food item, a new food item was placed randomly. The size of the world was 10x10 and both ends were connected as a torus. A creature was composed of two layers: visual layer and motor layer (Figure 3, left). There were five units1 in visual layer, one for each sensory input, and their dynamics were given by Equation (7). The self-connection a, the bias b and the input weight di were the same for all units. There were three units in motor layer, each coding one of three motor commands, and their state was given by ek + L: fkiYi(t), exp(xk(t)) L:/ exp(x/(t)) ' (8) (9) where ek was the bias and fki was the feedforward connection weight. 2 One of the three motor commands (L,C,R) was chosen stochastically with the probability Pk (k=L,C,R). The activation pattern in visual layer was shifted when the creature made a turn, which should give proper mapping between the sensory input and the working memory. 3.2 EVOLUTIONARY PROGRAMMING Each recurrent network was characterized by the parameters (a,b,Cij,di,ek,lkd, some of which were symmetrically shared, e.g. C12 = C21. For comparison, we also tested feedforward networks where recurrent connections were removed, i.e. a = Cij = O. A population of 60 creatures was tested on each generation. The initial population was generated with random parameters. Each of the top twenty scoring creatures produced three offspring; one identical copy of the parameters of the parent's and two copies of these parameters with a Gaussian fluctuation. In this paper, we report the result after 60 generations. 3.3 PERFORMANCE 1 We denote each unit in visual layer by Ul, U2, U3, U4, Us from the left to the right for the later convenience 2In this simulation reported here, we set ek = O. Dynamics of Attention as Near Saddle-node Bifurcation Behavior 43 -, -, -, , . . " .... : .... -, , -, , - 7 , _7 , - 10 - 12 . 5 -L25 a b "Transition matrix = ( : ~ .5 ) .5 bTransition matrix = ( :~ :~ ) Figure 4: The Convergence of the Parameter of (a , b) by Evolutionary Programming Plotted in the Bifurcation Diagram. The food density is 0.10 in both examples above. Table 1 shows the average of food found after 60 generations. As a reference of performance level, we also measured the performances of three other simple algorithms: 1) random walk: one of the three motor commands is taken randomly with equal probability. 2) nearest visible: move toward the nearest food visible at the time within the creature's field of view of (U2, U3, U4). 3) nearest visible/invisible: move toward the nearest food within the view of (U2, U3, U4) no matter if it is visible or not, which gives an upper bound of performance. The performance of recurrent network is better than that of feedforward network and 'nearest visible'. This suggests that the ability of recurrent network to remember the past is advantageous. The performance of feedforward network is better than that of 'nearest visible'. One reason is that feedforward network could cover a broader area to receive inputs than 'nearest visible'. In addition, two factors, the average time in which a creature reaches the food and the average time in which the food disappears, may influence the performance of feedforward network and 'nearest visible'. Feedforward network could optimize its output to adapt two factors with its broader view in evolution while 'nearest visible' did not have such adaptability. It should be noted that both of 'nearest visible/invisible' and 'nearest visible' explicitly assumed the higher-order sensory processing: distinguishing each food item from the others and measuring the distance between each food and its body. Since its performance is so different regardless of its higher-order sensory processing, it implies the importance of remembering the past. We can regard recurrent network as compromising two characteristics, remembering the past as 'nearest visible/invisible' did and optimizing the sensitivity as feedforward network did, although recurrent network did not have a perfect memory as 'nearest visible/invisible'. 3.4 CONVERGENCE TO NEAR-BIFURCATION REGIME We plotted the histogram of the performance in each generation and the history of the performance of a top-scoring creature over generations. Though they are not shown here, the performance was almost optimal after 60 generations. Figure 4 shows that two examples of a graph in which we plotted the parameter 44 H. NAKAHARA, K. DOYA set (a , b) of top twenty scoring creatures in the 60th generation in the bifurcation diagram. In the left graph, we can see the parameter set has converged to a regime that gives a near saddle-node bifurcation behavior. On the other hand, in the right graph, the parameter set has converged into the inside of cusp. It is interesting to note that the area inside of the cusp gives bistable dynamics. Hence, if the input is higher than a repelling point, it goes up and if the input is lower, it goes down. The reason of the convergence to that area is because of the difference of the world setting, that is, a Markov transition matrix. Since food would disappear more quickly and stay invisible longer in the setting of the right graph, it should be beneficial for a creature to remember the direction of higher inputs longer. In most of cases reported in Table 1, we obtained the convergence into our predicted regime and/or the inside of the cusp. 4 DISCUSSION Near saddle-node bifurcation behavior can have the long-term maintenance and quick transition, which characterize attention dynamics. A recurrent network has better performance than memoryless systems for tasks in our simulated nonstationary environment. Clearly, near saddle-node bifurcation behavior helped a creature's survival and in fact, creatures actually evolved to our expected parameter regime. However, we also obtained the convergence into another unexpected regime which gives bistable dynamics. How the bistable dynamics are used remains to be investigated. Acknowledgments H.N. is grateful to Ed Hutchins for his generous support, to John Batali and David Fogel for their advice on the implementation of evolutionary programming and to David Rogers for his comments on the manuscript of this paper. References R. Desimone, E. K. Miller, L. Chelazzi, & A. Lueschow. (1995) Multiple Memory Systems in the Visual Cortex. In M. Gazzaniga (ed.), The Cognitive Neurosciences, 475-486. MIT Press. D. B. Fogel, L. J. Fogel, & V. W. Porto. (1990) Evolving Neural Networks. Biological cybernetics 63:487-493. J. Guckenheimer & P. Homes. (1983) Nonlinear Oscillations, Dynamical Systems, and Bifurcation of Vector Fields C. Koch & S. Ullman. (1985) Shifts in selective visual attention:towards the underlying neural circuitry. Human Neurobiology 4:219-227. M. Posner, C .. R .R. Snyder, & B. J. Davidson. (1980) Attention and the detection of signals. Journal of Experimental Psychology: General 109:160-174
|
1995
|
109
|
1,009
|
Learning with ensembles: How over-fitting can be useful Peter Sollich Department of Physics University of Edinburgh, U.K. P.SollichGed.ac.uk Anders Krogh'" NORDITA, Blegdamsvej 17 2100 Copenhagen, Denmark kroghGsanger.ac.uk Abstract We study the characteristics of learning with ensembles. Solving exactly the simple model of an ensemble of linear students, we find surprisingly rich behaviour. For learning in large ensembles, it is advantageous to use under-regularized students, which actually over-fit the training data. Globally optimal performance can be obtained by choosing the training set sizes of the students appropriately. For smaller ensembles, optimization of the ensemble weights can yield significant improvements in ensemble generalization performance, in particular if the individual students are subject to noise in the training process. Choosing students with a wide range of regularization parameters makes this improvement robust against changes in the unknown level of noise in the training data. 1 INTRODUCTION An ensemble is a collection of a (finite) number of neural networks or other types of predictors that are trained for the same task. A combination of many different predictors can often improve predictions, and in statistics this idea has been investigated extensively, see e.g. [1, 2, 3]. In the neural networks community, ensembles of neural networks have been investigated by several groups, see for instance [4, 5, 6, 7]. Usually the networks in the ensemble are trained independently and then their predictions are combined. In this paper we study an ensemble of linear networks trained on different but overlapping training sets. The limit in which all the networks are trained on the full data set and the one where all the data sets are different has been treated in [8] . In this paper we treat the case of intermediate training set sizes and overlaps ·Present address: The Sanger Centre, Hinxton, Cambs CBIO IRQ, UK. Learning with Ensembles: How Overfitting Can Be Useful 191 exactly, yielding novel insights into ensemble learning. Our analysis also allows us to study the effect of regularization and of having different predictors in an ensemble. 2 GENERAL FEATURES OF ENSEMBLE LEARNING We consider the task of approximating a target function fo from RN to R. It will be assumed that we can only obtain noisy samples of the function, and the (now stochastic) target function will be denoted y(x). The inputs x are taken to be drawn from some distribution P(x). Assume now that an ensemble of K independent predictors fk(X) of y(x) is available. A weighted ensemble average is denoted by a bar, like lex) = L,wkfk(X), (1) k which is the final output of the ensemble. One can think of the weight Wk as the belief in predictor k and we therefore constrain the weights to be positive and sum to one. For an input x we define the error of the ensemble c(x), the error of the kth predictor ck(X), and its ambiguity ak(x) c(x) (y(x) -lex)? (2) ck(X) (y(x) - fk(X)? (3) (fk(X) -1(x»2. (4) The ensemble error can be written as c(x) = lex) - a(x) [7], where lex) = L,k Wkck(X) is the average error over the individual predictors and a(x) = L,k Wkak(X) is the average of their ambiguities, which is the variance of the output over the ensemble. By averaging over the input distribution P(x) (and implicitly over the target outputs y(x», one obtains the ensemble generalization error (5) where c(x) averaged over P(x) is simply denoted c, and similarly for land a. The first term on the right is the weighted average of the generalization errors of the individual predictors, and the second is the weighted average of the ambiguities, which we refer to as the ensemble ambiguity. An important feature of equation (5) is that it separates the generalization error into a term that depends on the generalization errors of the individual students and another term that contains all correlations between the students. The latter can be estimated entirely from unlabeled data, i. e., without any knowledge of the target function to be approximated. The relation (5) also shows that the more the predictors differ, the lower the error will be, provided the individual errors remain constant. In this paper we assume that the predictors are trained on a sample of p examples of the target function, (xt',yt'), where yt' = fo(xt') + TJt' and TJt' is some additive noise (Jl. = 1, ... ,p). The predictors, to which we refer as students in this context because they learn the target function from the training examples, need not be trained on all the available data. In fact, since training on different data sets will generally increase the ambiguity, it is possible that training on subsets of the data will improve generalization. An additional advantage is that, by holding out for each student a different part of the total data set for the purpose of testing, one can use the whole data set for training the ensemble while still getting an unbiased estimate of the ensemble generalization error. Denoting this estimate by f, one has (6) where Ctest = L,k WkCtest,k is the average of the students' test errors. As already pointed out, the estimate ft of the ensemble ambiguity can be found from unlabeled data. 192 P. SOLLICH, A. KROGH So far, we have not mentioned how to find the weights Wk. Often uniform weights are used, but optimization of the weights in some way is tempting. In [5, 6] the training set was used to perform the optimization, i.e., the weights were chosen to minimize the ensemble training error. This can easily lead to over-fitting, and in [7] it was suggested to minimize the estimated generalization error (6) instead. If this is done, the estimate (6) acquires a bias; intuitively, however, we expect this effect to be small for large ensembles. 3 ENSEMBLES OF LINEAR STUDENTS In preparation for our analysis of learning with ensembles of linear students we now briefly review the case of a single linear student, sometimes referred to as 'linear perceptron learning'. A linear student implements the input-output mapping 1 T J(x) = ..JNw x parameterized in terms of an N-dimensional parameter vector w with real components; the scaling factor 1/..JN is introduced here for convenience, and . .. T denotes the transpose of a vector. The student parameter vector w should not be confused with the ensemble weights Wk. The most common method for training such a linear student (or parametric inference models in general) is minimization of the sum-of-squares training error E = L:(y/J - J(x/J))2 + Aw2 /J where J.L = 1, ... ,p numbers the training examples. To prevent the student from fitting noise in the training data, a weight decay term Aw2 has been added. The size of the weight decay parameter A determines how strongly large parameter vectors are penalized; large A corresponds to a stronger regularization of the student. For a linear student, the global minimum of E can easily be found. However, in practical applications using non-linear networks, this is generally not true, and training can be thought of as a stochastic process yielding a different solution each time. We crudely model this by considering white noise added to gradient descent updates of the parameter vector w. This yields a limiting distribution of parameter vectors P(w) ex: exp(-E/2T), where the 'temperature' T measures the amount of noise in the training process. We focus our analysis on the 'thermodynamic limit' N -t 00 at constant normalized number of training examples, ex = p/ N. In this limit, quantities such as the training or generalization error become self-averaging, i.e., their averages over all training sets become identical to their typical values for a particular training set. Assume now that the training inputs x/J are chosen randomly and independently from a Gaussian distribution P(x) ex: exp( ~x2), and that training outputs are generated by a linear target function corrupted by additive noise, i.e., y/J = w'f x/J /..IN + 1]/J, where the 1]/J are zero mean noise variables with variance u2 • Fixing the length of the parameter vector of the target function to w~ = N for simplicity, the generalization error of a linear student with weight decay A and learning noise T becomes [9] 8G (; = (u2 + T)G + A(U2 - A) 8A . (7) On the r.h.s. of this equation we have dropped the term arising from the noise on the target function alone, which is simply u2 , and we shall follow this convention throughout. The 'response function' Gis [10, 11] G = G(ex, A) = (1 - ex - A + )(1 - ex - A)2 + 4A)/2A. (8) Learning with Ensembles: How Overfitting Can Be Useful 193 For zero training noise, T = 0, and for any a, the generalization error (7} is minimized when the weight decay is set to A = (T2j its value is then (T2G(a, (T2), which is the minimum achievable generalization error [9]. 3.1 ENSEMBLE GENERALIZATION ERROR We now consider an ensemble of K linear students with weight decays Ak and learning noises Tk (k = 1 . . . K). Each ,student has an ensemble weight Wk and is trained on N ak training examples, with students k and I sharing N akl training examples (of course, akk = ak). As above, we consider noisy training data generated by a linear target function. The resulting ensemble generalization error can be calculated by diagrammatic [10] or response function [11] methods. We refer the reader to a forthcoming publication for details and only state the result: (9) where (10) Here Pk is defined as Pk = AkG(ak, Ak). The Kronecker delta in the last term of (10) arises because the training noises of different students are uncorrelated. The generalization errors and ambiguities of the individual students are ak = ckk - 2 LWlckl + LWIWmclm; I 1m the result for the Ck can be shown to agree with the single student result (7). In the following sections, we shall explore the consequences of the general result (9). We will concentrate on the case where the training set of each student is sampled randomly from the total available data set of size NO', For the overlap of the training sets of students k and I (k 'II) one then has akl/a = (ak/a)(al/a) and hence ak/ = akal/a (11) up to fluctuations which vanish in the thermodynamic limit. For finite ensembles one can construct training sets for which akl < akal/a. This is an advantage, because it results in a smaller generalization error, but for simplicity we use (11). 4 LARGE ENSEMBLE LIMIT We now use our main result (9) to analyse the generalization performance of an ensemble with a large number K of students, in particular when the size of the training sets for the individual students are chosen optimally. If the ensemble weights Wk are approximately uniform (Wk ~ 1/ K) the off-diagonal elements of the matrix (ckl) dominate the generalization error for large K, and the contributions from the training noises n are suppressed. For the special case where all students are identical and are trained on training sets of identical size, ak = (1 - c)a, the ensemble generalization error is shown in Figure 1(left). The minimum at a nonzero value of c, which is the fraction of the total data set held out for testing each student, can clearly be seen. This confirms our intuition: when the students are trained on smaller, less overlapping training sets, the increase in error of the individual students can be more than offset by the corresponding increase in ambiguity. The optimal training set sizes ak can be calculated analytically: _ 1 - Ak/(T2 Ck = 1 - ak/a = 1 + G(a, (T2) ' (12) 194 w 1.0 r---,-----,r---.,----,.----:. 0.8 0.6 0.4 0.2 ,...------/ , / , / , 0.0 / , 0.0 0.2 0.4 0.6 0.8 1.0 C w P. SOLLICH, A. KROGH 1.0 r---,-----,---.----r----" 0.8 .' 0.6 0.2 ------0.0 ..... 0.0 0.2 0.4 0.6 0.8 1.0 C Figure 1: Generalization error and ambiguity for an infinite ensemble of identical students. Solid line: ensemble generalization error, fj dotted line: average generalization error of the individual students, l; dashed line: ensemble ambiguity, a. For both plots a = 1 and (72 = 0.2. The left plot corresponds to under-regularized students with A = 0.05 < (72. Here the generalization error of the ensemble has a minimum at a nonzero value of c. This minimum exists whenever>' < (72. The right plot shows the case of over-regularized students (A = 0.3 > (72), where the generalization error is minimal at c = O. The resulting generalization error is f = (72G(a, (72) + 0(1/ K), which is the globally minimal generalization error that can be obtained using all available training data, as explained in Section 3. Thus, a large ensemble with optimally chosen training set sizes can achieve globally optimal generalization performance. However, we see from (12) that a valid solution Ck > 0 exists only for Ak < (72, i.e., if the ensemble is under-regularized. This is exemplified, again for an ensemble of identical students, in Figure 1 (right) , which shows that for an over-regularized ensemble the generalization error is a: monotonic function of c and thus minimal at c = o. We conclude this section by discussing how the adaptation of the training set sizes could be performed in practice, for simplicity confining ourselves to an ensemble of identical students, where only one parameter c = Ck = 1- ak/a has to be adapted. If the ensemble is under-regularized one expects a minimum of the generalization error for some nonzero c as in Figure 1. One could, therefore, start by training all students on a large fraction of the total data set (corresponding to c ~ 0), and then gradually and randomly remove training examples from the students' training sets. Using (6), the generalization error of each student could be estimated by their performance on the examples on which they were not trained, and one would stop removing training examples when the estimate stops decreasing. The resulting estimate of the generalization error will be slightly biased; however, for a large enough ensemble the risk of a strongly biased estimate from systematically testing all students on too 'easy' training examples seems small, due to the random selection of examples. 5 REALISTIC ENSEMBLE SIZES We now discuss some effects that occur in learning with ensembles of 'realistic' sizes. In an over-regularized ensemble nothing can be gained by making the students more diverse by training them on smaller, less overlapping training sets. One would also Learning with Ensembles: How Overfitting Can Be Useful 195 Figure 2: The generalization error of an ensemble with 10 identical students as a function of the test set fraction c. From bottom to top the curves correspond to training noise T = 0,0.1,0.2, ... ,1.0. The star on each curve shows the error of the optimal single perceptron (i. e., with optimal weight decay for the given T) trained on all examples, which is independent of c. The parameters for this example are: a = 1, A = 0.05, 0'2 = 0.2. 0.2 0.0 L-_--'-_---' __ -'--_--'-_~ 0.0 0.2 0.4 0.6 0.8 1.0 C expect this kind of 'diversification' to be unnecessary or even counterproductive when the training noise is high enough to provide sufficient 'inherent' diversity of students. In the large ensemble limit, we saw that this effect is suppressed, but it does indeed occur in finite ensembles. Figure 2 shows the dependence of the generalization error on c for an ensemble of 10 identical, under-regularized students with identical training noises Tk = T. For small T, the minimum of f. at nonzero c persists. For larger T, f. is monotonically increasing with c, implying that further diversification of students beyond that caused by the learning noise is wasteful. The plot also shows the performance of the optimal single student (with A chosen to minimize the generalization error at the given T), demonstrating that the ensemble can perform significantly better by effectively averaging out learning noise. For realistic ensemble sizes the presence of learning noise generally reduces the potential for performance improvement by choosing optimal training set sizes. In such cases one can still adapt the ensemble weights to optimize performance, again on the basis of the estimate of the ensemble generalization error (6). An example is 1.0 1.0 I I 0.8 , 0.8 / , ,,I 0.6 I 0.6 tV tV 0.4 0.4 ..... -_ ................. 0.2 ---0.2 0.0 ....... 0.0 0.001 0.010 0 2 0.100 1.000 0.001 0.010 0 2 0.100 1.000 Figure 3: The generalization error of an ensemble of 10 students with different weight decays (marked by stars on the 0'2-axis) as a function of the noise level 0'2. Left: training noise T = 0; right: T = 0.1. The dashed lines are for the ensemble with uniform weights, and the solid line is for optimized ensemble weights. The dotted lines are for the optimal single perceptron trained on all data. The parameters for this example are: a = 1, c = 0.2. 196 P. SOu...ICH, A. KROGH shown in Figure 3 for an ensemble of size 1< = 10 with the weight decays >'k equally spaced on a logarithmic axis between 10-3 and 1. For both of the temperatures T shown, the ensemble with uniform weights performs worse than the optimal single student. With weight optimization, the generalization performance approaches that of the optimal single student for T = 0, and is actually better at T = 0.1 over the whole range of noise levels rr2 shown. Even the best single student from the ensemble can never perform better than the optimal single student, so combining the student outputs in a weighted ensemble average is superior to simply choosing the best member of the ensemble by cross-validation, i.e., on the basis of its estimated generalization error. The reason is that the ensemble average suppresses the learning noise on the individual students. 6 CONCLUSIONS We have studied ensemble learning in the simple, analytically solvable scenario of an ensemble of linear students. Our main findings are: In large ensembles, one should use under-regularized students in order to maximize the benefits of the variance-reducing effects of ensemble learning. In this way, the globally optimal generalization error on the basis of all the available data can be reached by optimizing the training set sizes of the individual students. At the same time an estimate of the generalization error can be obtained. For ensembles of more realistic size, we found that for students subjected to a large amount of noise in the training process it is unnecessary to increase the diversity of students by training them on smaller, less overlapping training sets. In this case, optimizing the ensemble weights can still yield substantially better generalization performance than an optimally chosen single student trained on all data with the same amount of training noise. This improvement is most insensitive to changes in the unknown noise levels rr2 if the weight decays of the individual students cover a wide range. We expect most of these conclusions to carryover, at least qualitatively, to ensemble learning with nonlinear models, and this correlates well with experimental results presented in [7]. References [1] C. Granger, Journal of Forecasting 8, 231 (1989). [2] D. Wolpert, Neural Networks 5, 241 (1992). [3] L. Breimann, Tutorial at NIPS 7 and personal communication. [4] L. Hansen and P. Salamon, IEEE Trans. Pattern Anal. and Mach. Intell. 12, 993 (1990). [5] M. P. Perrone and L. N. Cooper, in Neural Networks for Speech and Image processing, ed. R. J. Mammone (Chapman-Hall, 1993). [6] S. Hashem: Optimal Linear Combinations of Neural Networks. Tech. Rep. PNL-SA-25166, submitted to Neural Networks (1995). [7] A. Krogh and J. Vedelsby, in NIPS 7, ed. G. Tesauro et al., p. 231 (MIT Press, 1995). [8] R. Meir, in NIPS 7, ed. G. Tesauro et al., p. 295 (MIT Press, 1995). [9] A. Krogh and J. A. Hertz, J. Phys. A 25,1135 (1992). [10] J. A. Hertz, A. Krogh, and G. I. Thorbergsson, J. Phys. A 22, 2133 (1989). [11] P. Sollich, J. Phys. A 27, 7771 (1994).
|
1995
|
11
|
1,010
|
Softassign versus Softmax: Benchmarks in Combinatorial Optimization Steven Gold Department of Computer Science Yale University New Haven, CT 06520-8285 Anand Rangarajan Dept. of Diagnostic Radiology Yale University New Haven, CT 06520-8042 Abstract A new technique, termed soft assign, is applied for the first time to two classic combinatorial optimization problems, the traveling salesman problem and graph partitioning. Soft assign , which has emerged from the recurrent neural network/statistical physics framework, enforces two-way (assignment) constraints without the use of penalty terms in the energy functions. The soft assign can also be generalized from two-way winner-take-all constraints to multiple membership constraints which are required for graph partitioning. The soft assign technique is compared to the softmax (Potts glass). Within the statistical physics framework, softmax and a penalty term has been a widely used method for enforcing the two-way constraints common within many combinatorial optimization problems. The benchmarks present evidence that soft assign has clear advantages in accuracy, speed, parallelizabilityand algorithmic simplicity over softmax and a penalty term in optimization problems with two-way constraints. 1 Introduction In a series of papers in the early to mid 1980's, Hopfield and Tank introduced techniques which allowed one to solve combinatorial optimization problems with recurrent neural networks [Hopfield and Tank, 1985]. As researchers attempted to reproduce the original traveling salesman problem results of Hopfield and Tank, problems emerged, especially in terms of the quality of the solutions obtained. More recently however, a number of techniques from statistical physics have been adopted to mitigate these problems. These include deterministic annealing which convexifies the energy function in order help avoid some local minima and the Potts glass approximation which results in a hard enforcement of a one-way (one set of) winner-take-all (WTA) constraint via the softmax. In Softassign versus Softmax: Benchmarks in Combinatorial Optimization 627 the late 80's, armed with these techniques optimization problems like the traveling salesman problem (TSP) [Peterson and Soderberg, 1989] and graph partitioning [Peterson and Soderberg, 1989, Van den Bout and Miller III, 1990] were reexamined and much better results compared to the original Hopfield-Tank dynamics were obtained. However, when the problem calls for two-way interlocking WTA constraints, as do TSP and graph partitioning, the resulting energy function must still include a penalty term when the softmax is employed in order to enforce the second set of WTA constraints. Such penalty terms may introduce spurious local minima in the energy function and involve free parameters which are hard to set. A new technique, termed soft assign, eliminates the need for all such penalty terms. The first use of the soft assign was in an algorithm for the assignment problem [Kosowsky and Yuille, 1994]. It has since been applied to much more difficult optimization problems, including parametric assignment problems-point matching [Gold et aI., 1994, Gold et aI., 1995, Gold et aI., 1996] and quadratic assignment problems-graph matching [Gold et aI., 1996, Gold and Rangarajan, 1996, Gold, 1995]. Here, we for the first time apply the soft assign to two classic combinatorial optimization problems, TSP and graph partitioning. Moreover, we show that the soft assign can be generalized from two-way winner-take-all constraints to multiple membership constraints, which are required for graph partitioning (as described below). We then run benchmarks against the older softmax (Potts glass) methods and demonstrate advantages in terms of accuracy, speed, parallelizability, and simplicity of implementation. It must be emphasized there are other conventional techniques, for solving some combinatorial optimization problems such as TSP, which remain superior to this method in certain ways [Lawler et aI., 1985]. (We think for some problems-specifically the type of pattern matching problems essential for cognition [Gold, 1995]-this technique is superior to conventional methods.) Even within neural networks, elastic net methods may still be better in certain cases. However, the elastic net uses only a one-way constraint in TSP. The main goal of this paper is to provide evidence, that when minimizing energy functions within the neural network framework, which have two-way constraints, the soft assign should be the technique of choice. We therefore compare it to the current dominant technique, softmax with a penalty term. 2 Optimizing With Softassign 2.1 The Traveling Salesman Problem The traveling salesman problem may be defined in the following way. Given a set of intercity distances {hab} which may take values in R+ , find the permutation matrix M such that the following objective function is minimized. 1 N N N E 1(M) = 2 LLL habMai Mb(i6H) a==lb==li=l (1) subject to Va L~l Mai = 1 , Vi L~=l Mai = 1 , Vai Mai E {O, 1}. In the above objective hab represents the distance between cities a and b. M is a permutation matrix whose rows represent cities, and whose columns represent the day (or order) the city was visited and N is the number of cities. (The notation i EEl 1 628 S.GOLD,A.RANGARAJAN is used to indicate that subscripts are defined modulo N, i.e. Ma(N+I) = Mal.) So if Mai = 1 it indicates that city a was visited on day i. Then, following [Peterson and Soderberg, 1989, Yuille and Kosowsky, 1994] we employ Lagrange multipliers and an x log x barrier function to enforce the constraints, as well as a 'Y term for stability, resulting in the following objective: 1 N N N N N E2(M, 1',11) = 2 L L L babMaiMb(ieJ I ) ~ L L M;i a=l b=l i=l a=l i=l INN N N N N +p I: I: Mai(10g M ai - 1) + I: J.la(I: Mai - 1) + I: lIi(I: Mai - 1) (2) a=l i=l a=l i=l i=l a=l In the above we are looking for a saddle point by minimizing with respect to M and maximizing with respect to I' and 11, the Lagrange multipliers. 2.2 The Soft assign In the above formulation of TSP we have two-way interlocking WTA constraints. {Mai} must be a permutation matrix to ensure that a valid tour-one in which each city is visited once and only once-is described. A permutation matrix means all the rows and columns must add to one (and the elements must be zero or one) and therefore requires two-way WTA constraints-a set of WTA constraints on the rows and a set of WTA constraints on the columns. This set of two-way constraints may also be considered assignment constraints, since each city must be assigned to one and only one day (the row constraint) and each day must be assigned to one and only one city (the column constraint). These assignment constraints can be satisfied using a result from [Sinkhorn, 1964]. In [Sinkhorn, 1964] it is proven that any square matrix whose elements are all positive will converge to a doubly stochastic matrix just by the iterative process of alternatively normalizing the rows and columns. (A doubly stochastic matrix is a matrix whose elements are all positive and whose rows and columns all add up to one-it may roughly be thought of as the continuous analog of a permutation matrix). The soft assign simply employs Sinkhorn's technique within a deterministic annealing context. Figure 1 depicts the contrast between the soft assign and the softmax. In the softmax, a one-way WTA constraint is strictly enforced by normalizing over a vector. [Kosowsky and Yuille, 1994] used the soft assign to solve the assignment problem, i.e. minimize: 2:~=1 2:{=1 MaiQai. For the special case of the quadratic assignment problem, being solved here, by setting Q ai = -:J:i' and using the values of M from the previous iteration, we can at each iteration produce a new assignment problem for which the soft assign then returns a doubly stochastic matrix. As the temperature is lowered a series of assignment problems are generated, along with the corresponding doubly stochastic matrices returned by each soft assign , until a permutation matrix is reached. The update with the partial derivative in the preceding may be derived using a Taylor series expansion. See [Gold and Rangarajan, 1996, Gold, 1995] for details. The algorithm dynamics then become: Softassign versus Softmax: Benchmarks in Combinatorial Optimization 629 Softassign Softmax Positivity M.i = exP(I3Q.) 1 Positivity Mi = exP(I3Qi) Two-way constraints Row Normalization 1 ( ~) Mai--_ 1 l:M.i 0<;. """"'~_ M.i- l:~. a 1 One-way constraint M· M· ___ 1_ 1 l:M. i ) Figure 1: Softassign and softmax. This paper compares these two techniques. (3) Mai = Softassignai (Q) (4) E2 is E2 without the {3, J.l or II terms of (2), therefore no penalty terms are now included. The above dynamics are iterated as (3, the inverse temperature, is gradually increased. These dynamics may be obtained by evaluating the saddle points of the objective in (2). Sinkhorn's method finds the saddle points for the Lagrange parameters. 2.3 Graph Partitioning The graph partitioning problem maybe defined in the following way. Given an unweighted graph G, find the membership matrix M such that the following objective function is minimized. A I I E3(M) = - I:L:L:GijMaiMaj (5) a=1 i=1 j=1 subject to Va E;=1 Mai = IIA, Vi E:=1 Mai = 1, Vai Mai E to, I} where graph G has I nodes which should be equally partitioned into A bins. {Gij} is the adjacency matrix of the graph, whose elements must be 0 or 1. M is a membership matrix such that Mai = 1 indicates that node i is in bin a. The permutation matrix constraint present in TSP is modified to the membership constraint. Node i is a member of only bin a and the number of members in each bin is fixed at IIA. When the above objective is at a minimum, then graph G will be partitioned into A equal sized bins, such that the cutsize is minimum for all possible partitionings of G into A equal sized bins. We assume IIA is an integer. Then following the treatment for TSP, we derive the following objective: 630 S.GOLD,A. RANGARAJAN A I I A I E4(M,p,v) = - I: I:L: CijMaiMaj ~ L:L:M;i a=l i=l j=l a=l i=l 1AI A I I A +:8 I: I: Mai(lOgMai - 1) + I:Pa(2: Mai - [fA) + 2: Vi (2: Mai -1) (6) a=li=l a=l i=l i=l a=l which is minimized with a similar algorithm employing the softassign. Note however now in the soft assign the columns are normalized to [j A instead of 1. 8 Experimental Results Experiments on Euclidean TSP and graph partitioning were conducted. For each problem three different algorithms were run. One used the soft assign described above. The second used the Potts glass dynamics employing synchronous update as described in [Peterson and Soderberg, 1989]. The third used the Potts glass dynamics employing serial update as described in [Peterson and Soderberg, 1989]. Originally the intention was to employ just the synchronous updating version of the Potts glass dynamics, since that is the dynamics used in the algorithms employing soft assign and is the method that is massively parallelizable. We believe massive parallelism to be such a critical feature of the neural network architecture [Rumelhart and McClelland, 1986] that any algorithm that does not have this feature loses much of the power of the neural network paradigm. Unfortunately the synchronous updating algorithms just worked so poorly that we also ran the serial versions in order to get a more extensive comparison. Note that the results reported in [Peterson and Soderberg, 1989] were all with the serial versions. 3.1 Euclidean TSP Experiments Figure 2 shows the results of the Euclidean TSP experiments. 500 different 100city tours from points uniformly generated in the 2D unit square were used as input. The asymptotic expected length of an optimal tour for cities distributed in the unit square is given by L( n) = J( Vn where n is the number of cities and 0.765 ~ J( ~ 0.765 +.1 [Lawler et al., 1985]. This gives the interval [7.65,8.05] for the 100 city TSP. 95<70 of the tour lengths fall in the interval [8,11] when using the soft assign approach. Note the large difference in performance between the soft assign and the Potts glass algorithms. The serial Potts glass algorithm ran about 5 times slower than the soft assign version. Also as noted previously the serial version is not massively parallelizable. The synchronous Potts glass ran about 2 times slower. Also note the softassign algorithm is much simpler to implement-fewer parameters to tune. 3.2 Graph Partitioning Experiments Figure 3 shows the results of the graph partitioning experiments. 2000 different randomly generated 100 node graphs with 10% connectivity were used as input. These graphs were partitioned into four bins. The soft assign performs better than the Potts glass algorithms, however here the difference is more modest than in the TSP experiments. However the serial Potts glass algorithm again ran about 5 times slower then the soft assign version and as noted previously the serial version is not massively parallelizable. The synchronous Potts glass ran about 2 times slower. Softassign versus Softmax: Benchmarks in Combinatorial Optimization 631 r-"' • r• "' r-• ,. r,. • ,. •• • r• • • 11 It ,. " 11 • n Inn • ..--:I .r-I!'--,,:,:...u:~~-""'=---!;-~,........_---!,.. I II 11 "' n "I 11 ,.1 ,. '.1 1.~' " It II • . .. .. --"' ,. -......... ... ..... Figure 2: 100 City Euclidean TSP. 500 experiments. Left: Softassign .. Middle: Softmax (serial update). Right: Softmax (synchronous update). Also again note the softassign algorithm was much simpler to implement-fewer parameters to tune. .. ',.. e, r0.. til e.. "' rr,. , •• rInn. .n "' ,. .. til ,. • .n n_ ,. nn. • ~n ~ .. ... "' "' til _ ~ .. _ M .. "' _ .. Figure 3: 100 node Graph Partitioning, 4 bins. 2000 experiments. Left: Softassign •. Middle: Softmax (serial update). Right: Softmax (synchronous update). A relatively simple version of graph partitioning was run. It is likely that as the number of bins are increased the results on graph partitioning will come to resemble more closely the TSP results, since when the number of bins equal the number of nodes, the TSP can be considered a special case of graph partitioning (there are some additional restrictions). However even in this simple case the softassign has clear advantages over the softmax and penalty term. 4 Conclusion For the first time, two classic combinatorial optimization problems, TSP and graph partitioning, are solved using a new technique for constraint satisfaction, the soft assign. The softassign, which has recently emerged from the statistical physics/neural networks framework, enforces a two-way (assignment) constraint, without penalty terms in the energy function. We also show that the softassign can be generalized from two-way winner-take-all constraints to multiple membership constraints, which are required for graph partitioning. Benchmarks against the Potts glass methods, using softmax and a penalty term, clearly demonstrate its advantages in terms of accuracy, speed, parallelizability and simplicity of implementation. Within the neural network/statistical physics framework, soft assign should be considered the technique of choice for enforcing two-way constraints in energy functions. 632 S. GOLD,A. RANGARAJAN References [Gold, 1995] Gold, S ~ (1995). Matching and Learning Structural and Spatial Representations with Neural Networks. PhD thesis, Yale University. [Gold et al., 1995] Gold, S., Lu, C. P., Rangarajan, A., Pappu, S., and Mjolsness, E. (1995). New algorithms for 2-D and 3-D point matching: pose estimation and correspondence. In Tesauro, G., Touretzky, D. S., and Leen, T. K., editors, Advances in Neural Information Processing Systems 7, pages 957-964. MIT Press, Cambridge, MA. [Gold et al. , 1994] Gold, S., Mjolsness, E., and Rangarajan, A. (1994). Clustering with a domain specific distance measure. In Cowan, J., Tesauro, G., and AIspector, J., editors, Advances in Neural Information Processing Systems 6, pages 96-103. Morgan Kaufmann, San Francisco, CA. [Gold and Rangarajan, 1996] Gold, S. and Rangarajan, A. (1996). A graduated assignment algorithm for graph matching. IEEE Transactions on Pattern Analysis and Machine Intelligence, (in press). [Gold et al., 1996] Gold, S., Rangarajan, A., and Mjolsness, E. (1996). Learning with preknowledge: clustering with point and graph matching distance measures. Neural Computation, (in press). [Hopfield and Tank, 1985] Hopfield, J. J. and Tank, D. (1985). 'Neural' computation of decisions in optimization problems. Biological Cybernetics, 52:141-152. [Kosowsky and Yuille, 1994] Kosowsky, J . J . and Yuille, A. L. (1994). The invisible hand algorithm: Solving the assignment problem with statistical physics. Neural Networks, 7(3):477-490. [Lawler et al., 1985] Lawler, E. L., Lenstra, J. K., Kan, A. H. G. R., and Shmoys, D. B., editors (1985). The Traveling Salesman Problem. John Wiley and Sons, Chichester. [Peterson and Soderberg, 1989] Peterson, C. and Soderberg, B. (1989). A new method for mapping optimization problems onto neural networks. Inti. Journal of Neural Systems, 1(1):3-22. [Rumelhart and McClelland, 1986] Rumelhart, D. and McClelland, J. L. (1986). Parallel Distributed Processing, volume 1. MIT Press, Cambridge, MA. [Sinkhorn, 1964] Sinkhorn, R. (1964). A relationship between arbitrary positive matrices and doubly stochastic matrices. Ann. Math. Statist., 35:876-879. [Van den Bout and Miller III, 1990] Van den Bout, D. E. and Miller III, T . K. (1990). Graph partitioning using annealed networks. IEEE Trans. Neural Networks, 1(2):192-203. [Yuille and Kosowsky, 1994] Yuille, A. L. and Kosowsky, J. J. (1994). Statistical physics algorithms that converge. Neural Computation, 6(3):341-356.
|
1995
|
110
|
1,011
|
From Isolation to Cooperation: An Alternative View of a System of Experts Stefan Schaal:!:* Christopher C. Atkeson:!: sschaal@cc.gatech.edu cga@cc.gatech.edu http://www.cc.gatech.eduifac/Stefan.Schaal http://www.cc.gatech.eduifac/Chris.Atkeson +College of Computing, Georgia Tech, 801 Atlantic Drive, Atlanta, GA 30332-0280 * A TR Human Infonnation Processing, 2-2 Hikaridai, Seiko-cho, Soraku-gun, 619-02 Kyoto Abstract We introduce a constructive, incremental learning system for regression problems that models data by means of locally linear experts. In contrast to other approaches, the experts are trained independently and do not compete for data during learning. Only when a prediction for a query is required do the experts cooperate by blending their individual predictions. Each expert is trained by minimizing a penalized local cross validation error using second order methods. In this way, an expert is able to find a local distance metric by adjusting the size and shape of the receptive field in which its predictions are valid, and also to detect relevant input features by adjusting its bias on the importance of individual input dimensions. We derive asymptotic results for our method. In a variety of simulations the properties of the algorithm are demonstrated with respect to interference, learning speed, prediction accuracy, feature detection, and task oriented incremental learning. 1. INTRODUCTION Distributing a learning task among a set of experts has become a popular method in computationallearning. One approach is to employ several experts, each with a global domain of expertise (e.g., Wolpert, 1990). When an output for a given input is to be predicted, every expert gives a prediction together with a confidence measure. The individual predictions are combined into a single result, for instance, based on a confidence weighted average. Another approach-the approach pursued in this paper-of employing experts is to create experts with local domains of expertise. In contrast to the global experts, the local experts have little overlap or no overlap at all. To assign a local domain of expertise to each expert, it is necessary to learn an expert selection system in addition to the experts themselves. This classifier determines which expert models are used in which part of the input space. For incremental learning, competitive learning methods are usually applied. Here the experts compete for data such that they change their domains of expertise until a stable configuration is achieved (e.g., Jacobs, Jordan, Nowlan, & Hinton, 1991). The advantage of local experts is that they can have simple parameterizations, such as locally constant or locally linear models. This offers benefits in terms of analyzability, learning speed, and robustness (e.g., Jordan & Jacobs, 1994). For simple experts, however, a large number of experts is necessary to model a function. As a result, the expert selection system has to be more complicated and, thus, has a higher risk of getting stuck in local minima and/or of learning rather slowly. In incremental learning, another potential danger arises when the input distribution of the data changes. The expert selection system usually makes either implicit or explicit prior assumptions about the input data distribution. For example, in the classical mixture model (McLachlan & Basford, 1988) which was employed in several local expert approaches, the prior probabilities of each mixture model can be interpreted as 606 S. SCHAAL. C. C. ATKESON the fraction of data points each expert expects to experience. Therefore, a change in input distribution will cause all experts to change their domains of expertise in order to fulfill these prior assumptions. This can lead to catastrophic interference. In order to avoid these problems and to cope with the interference problems during incremental learning due to changes in input distribution, we suggest eliminating the competition among experts and instead isolating them during learning. Whenever some new data is experienced which is not accounted for by one of the current experts, a new expert is created. Since the experts do not compete for data with their peers, there is no reason for them to change the location of their domains of expertise. However, when it comes to making a prediction at a query point, all the experts cooperate by giving a prediction of the output together with a confidence measure. A blending of all the predictions of all experts results in the final prediction. It should be noted that these local experts combine properties of both the global and local experts mentioned previously. They act like global experts by learning independently of each other and by blending their predictions, but they act like local experts by confining themselves to a local domain of expertise, i.e., their confidence measures are large only in a local region. The topic of data fitting with structurally simple local models (or experts) has received a great deal of attention in nonparametric statistics (e.g., Nadaraya, 1964; Cleveland, 1979; Scott, 1992, Hastie & Tibshirani, 1990). In this paper, we will demonstrate how a nonparametric approach can be applied to obtain the isolated expert network (Section 2.1), how its asymptotic properties can be analyzed (Section 2.2), and what characteristics such a learning system possesses in terms of the avoidance of interference, feature detection, dimensionality reduction, and incremental learning of motor control tasks (Section 3). 2. RECEPTIVE FIELD WEIGHTED REGRESSION This paper focuses on regression problems, i.e., the learning of a map from 9tn ~ 9tm • Each expert in our learning method, Receptive Field Weighted Regression (RFWR), consists of two elements, a locally linear model to represent the local functional relationship, and a receptive field which determines the region in input space in which the expert's knowledge is valid. As a result, a given data set will be modeled by piecewise linear elements, blended together. For 1000 noisy data points drawn from the unit interval of the function z == max[exp(-10x 2),exp(-50l),1.25exp(-5(x2 + l)], Figure 1 illustrates an example of function fitting with RFWR. This function consists of a narrow and a wide ridge which are perpendicular to each other, and a Gaussian bump at the origin. Figure 1 b shows the receptive fields which the system created during the learning process. Each experts' location is at the center of its receptive field, marked by a $ in Figure 1 b. The recep0 . 5 0 -0.5 -1 1.5 ,1 10. 5% 0 I 1- 0 .5 1 0 - 0 .5 -1 x (a) 1.5 0.5 ,., 0 -0.5 -1 -1.5 -1.5 (b) -1 -0.5 o x 0.5 1.5 Figure 1: (a) result of function approximation with RFWR. (b) contour lines of 0.1 iso-activation of each expert in input space (the experts' centers are marked by small circles). From Isolation to Cooperation: An Alternative View of a System of Experts 607 tive fields are modeled by Gaussian functions, and their 0.1 iso-activation lines are shown in Figure 1 b as well. As can be seen, each expert focuses on a certain region of the input space, and the shape and orientation of this region reflects the function's complexity, or more precisely, the function's curvature, in this region. It should be noticed that there is a certain amount of overlap among the experts, and that the placement of experts occurred on a greedy basis during learning and is not globally optimal. The approximation result (Figure 1 a) is a faithful reconstruction of the real function (MSE = 0.0025 on a test set, 30 epochs training, about 1 minute of computation on a SPARC1O). As a baseline comparison, a similar result with a sigmoidal 3-layer neural network required about 100 hidden units and 10000 epochs of annealed standard backpropagation (about 4 hours on a SPARC1O). 2.1 THE ALGORITHM li'Iear ..•... '. ~"" " Galng Unrt WeighBd' / ~:~~ ConnectIOn Average centered at e Output y, Figure 2: The RFWR network RFWR can be sketched in network form as shown in Figure 2. All inputs connect to all expert networks, and new experts can be added as needed. Each expert is an independent entity. It consists of a two layer linear subnet and a receptive field subnet. The receptive field subnet has a single unit with a bell-shaped activation profile, centered at the fixed location c in input space. The maximal output of this unit is "I" at the center, and it decays to zero as a function of the distance from the center. For analytical convenience, we choose this unit to be Gaussian: (1) x is the input vector, and D the distance metric, a positive definite matrix that is generated from the upper triangular matrix M. The output of the linear subnet is: A Tb b -Tf3 y=x + o=x (2) The connection strengths b of the linear subnet and its bias bO will be denoted by the d-dimensional vector f3 from now on, and the tilde sign will indicate that a vector has been augmented by a constant "I", e.g., i = (x T , Il. In generating the total output, the receptive field units act as a gating component on the output, such that the total prediction is: (3) The parameters f3 and M are the primary quantities which have to be adjusted in the learning process: f3 forms the locally linear model, while M determines the shape and orientation of the receptive fields. Learning is achieved by incrementally minimizing the cost function: (4) The first term of this function is the weighted mean squared cross validation error over all experienced data points, a local cross validation measure (Schaal & Atkeson, 1994). The second term is a regularization or penalty term. Local cross validation by itself is consistent, i.e., with an increasing amount of data, the size of the receptive field of an expert would shrink to zero. This would require the creation of an ever increasing number of experts during the course of learning. The penalty term introduces some non-vanishing bias in each expert such that its receptive field size does not shrink to zero. By penalizing the squared coefficients of D, we are essentially penalizing the second derivatives of the function at the site of the expert. This is similar to the approaches taken in spline fitting 608 S. SCHAAL, C. C. A TI(ESON (deBoor, 1978) and acts as a low-pass filter: the higher the second derivatives, the more smoothing (and thus bias) will be introduced. This will be analyzed further in Section 2.2. The update equations for the linear subnet are the standard weighted recursive least squares equation with forgetting factor A (Ljung & SOderstrom, 1986): 1 ( pn- -Tpn ) f3n+1 =f3n+wpn+lxe wherepn+1 =_ pn_ xx ande =(y-xT f3n) cv' A Ajw + xTpnx cv (5) This is a Newton method, and it requires maintaining the matrix P, which is size 0.5d x (d + 1) . The update of the receptive field subnet is a gradient descent in J: Mn+l=Mn- a dJ!aM (6) Due to space limitations, the derivation of the derivative in (6) will not be explained here. The major ingredient is to take this derivative as in a batch update, and then to reformulate the result as an iterative scheme. The derivatives in batch mode can be calculated exactly due to the Sherman-Morrison-Woodbury theorem (Belsley, Kuh, & Welsch, 1980; Atkeson, 1992). The derivative for the incremental update is a very good approximation to the batch update and realizes incremental local cross validation. A new expert is initialized with a default M de! and all other variables set to zero, except the matrix P. P is initialized as a diagonal matrix with elements 11 r/, where the ri are usually small quantities, e.g., 0.01. The ri are ridge regression parameters. From a probabilistic view, they are Bayesian priors that the f3 vector is the zero vector. From an algorithmic view, they are fake data points of the form [x = (0, ... , '12 ,o, ... l,y = 0] (Atkeson, Moore, & Schaal, submitted). Using the update rule (5), the influence of the ridge regression parameters would fade away due to the forgetting factor A. However, it is useful to make the ridge regression parameters adjustable. As in (6), rj can be updated by gradient descent: 1'n+1 = 1'n - a aJ/ar I I I (7) There are d ridge regression parameters, one for each diagonal element of the P matrix. In order to add in the update of the ridge parameters as well as to compensate for the forgetting factor, an iterative procedure based on (5) can be devised which we omit here. The computational complexity of this update is much reduced in comparison to (5) since many computations involve multiplications by zero. Initialize the RFWR network. with no expert; For every new training sample (x,y): a) For k= I to #experts: b) c) d) e) end; - calculate the activation from (I) - update the expert's parameters according to (5), (6), and (7) end; Ir no expert was activated by more than W gen: - create a new expert with c=x end; Ir two experts are acti vated more than W pn .. ~ - erase the expert with the smaller receptive field end; calculate the mean, err ""an' and standard de viation errslIl of the incrementally accumulated error er,! of all experts; For k.= I to #experts: Ir (Itrr! - err_I> 9 er'Sld) reinitialize expert k with M = 2 • Mdef end; In sum, a RFWR expert consists of three sets of parameters, one for the locally linear model, one for the size and shape of the receptive fields, and one for the bias. The linear model parameters are updated by a Newton method, while the other parameters are updated by gradient descent. In our implementations, we actually use second order gradient descent based on Sutton (1992), since, with minor extra effort, we can obtain estimates of the second derivatives of the cost function with respect to all parameters. Finally, the logic of RFWR becomes as shown in the pseudo-code above. Point c) and e) of the algorithm introduce a pruning facility. Pruning takes place either when two experts overlap too much, or when an expert has an exceptionally large mean squared error. The latter method corresponds to a simple form of outlier detection. Local optimization of a distance metric always has a minimum for a very large receptive field size. In our case, this would mean that an expert favors global instead of locally linear regression. Such an expert will accumulate a very large error which can easily be detected From Isolation to Cooperation: An Alternative View of a System of Experts 609 in the given way. The mean squared error term, err, on which this outlier detection is based, is a bias-corrected mean squared error, as will be explained below. 2.2 ASYMPTOTIC BIAS AND PENALTY SELECTION The penalty term in the cost function (4) introduces bias. In order to assess the asymptotic value of this bias, the real function f(x) , which is to be learned, is assumed to be represented as a Taylor series expansion at the center of an expert's receptive field. Without loss of generality, the center is assumed to be at the origin in input space. We furthermore assume that the size and shape of the receptive field are such that terms higher than 0(2) are negligible. Thus, the cost (4) can be written as: J ~ (1w(f. +fTX+~XTFX-bo -bTx Y dx )/(1 wdx )+r~Dnm (8) where fo' f, and F denote the constant, linear, and quadratic terms of the Taylor series expansion, respectively. Inserting Equation (1), the integrals can be solved analytically after the input space is rotated by an orthonormal matrix transforming F to the diagonal matrix F'. Subsequently, bo' b, and D can be determined such that J is minimized: 0.25 ( ) ~ b~ = fa + bias = fa + ~075 ~ sgn(F:')~IF;,:I, b' = f, D:: = (2r)2 (9) This states that the linear model will asymptotically acquire the correct locally linear model, while the constant term will have bias proportional to the square root of the sum of the eigenvalues of F, i.e., the F:n • The distance metric D, whose diagonalized counterpart is D', will be a scaled image of the Hessian F with an additional square root distortion. Thus, the penalty term accomplishes the intended task: it introduces more smoothing the higher the curvature at an expert's location is, and it prevents the receptive field of an expert shrinking to zero size (which would obviously happen for r ~ 0). Additionally, Equation (9) shows how to determine rfor a given learning problem from an estimate of the eigenvalues and a permissible bias. Finally, it is possible to derive estimates of the bias and the mean squared error of each expert from the current distance metric D: biasesl = ~0 .5r IJeigenvalues(D)l.; en,,~, = r L D;m (10) n.m The latter term was incorporated in the mean squared error, err, in Section 2.1. Empirical evaluations (not shown here) verified the validity of these asymptotic results. 3. SIMULA TION RESULTS This section will demonstrate some of the properties of RFWR. In all simulations, the threshold parameters of the algorithm were set to e = 3.5, w prune = 0.9, and w min = 0.1. These quantities determine the overlap of the experts as well as the outlier removal threshold; the results below are not affected by moderate changes in these parameters. 3.1 AVOIDING INTERFERENCE In order to test RFWR's sensitivity with respect to changes in input data distribution, the data of the example of Figure 1 was partitioned into three separate training sets 1; = {(x, y, z) 1-1.0 < x < -O.2} , 1; = {(x, y, z) 1-0.4 < x < OA}, 1; = {(x, y, z) I 0.2 < x < 1.0}. These data sets correspond to three overlapping stripes of data, each having about 400 uniformly distributed samples. From scratch, a RFWR network was trained first on I; for 20 epochs, then on T2 for 20 epochs, and finally on 1; for 20 epochs. The penalty was chosen as in the example of Figure 1 to be r = I.e - 7 , which corresponds to an asymptotic bias of 610 S. SCHAAL, C. C. ATKESON 0.1 at the sharp ridge of the function. The default distance metric D was 50*1, where I is the identity matrix. Figure 3 shows the results of this experiment. Very little interference can be found. The MSE on the test set increased from 0.0025 (of the original experiment of Figure 1) to 0.003, which is still an excellent reconstruction of the real function. y 0 .5 -0 . 5 - 0 . 5 (a) (b) (c) -1 Figure 3: Reconstructed function after training on (a) 7;, (b) then ~,(c) and finally 1;. 3.2 LOCAL FEATURE DETECTION The examples of RFWR given so far did not require ridge regression parameters. Their importance, however, becomes obvious when dealing with locally rank deficient data or with irrelevant input dimensions. A learning system should be able to recognize irrelevant input dimensions. It is important to note that this cannot be accomplished by a distance metric. The distance metric is only able to decide to what spatial extent averaging over data in a certain dimension should be performed. However, the distance metric has no means to exclude an input dimension. In contrast, bias learning with ridge regression parameters is able to exclude input dimensions. To demonstrate this, we added 8 purely noisy inputs (N(0,0.3)) to the data drawn from the function of Figure 1. After 30 epochs of training on a 10000 data point training set, we analyzed histograms of the order of magnitude of the ridge regression parameters in all 100bias input dimensions over all the 79 experts that had been generated by the learning algorithm. All experts recognized that the input dimensions 3 to 8 did not contain relevant information, and correctly increased the corresponding ridge parameters to large values. The effect of a large ridge regression parameter is that the associated regression coefficient becomes zero. In contrast, the ridge parameters of the inputs 1, 2, and the bias input remained very small. The MSE on the test set was 0.0026, basically identical to the experiment with the original training set. 3.3 LEARNING AN INVERSE DYNAMICS MODEL OF A ROBOT ARM Robot learning is one of the domains where incremental learning plays an important role. A real movement system experiences data at a high rate, and it should incorporate this data immediately to improve its performance. As learning is task oriented, input distributions will also be task oriented and interference problems can easily arise. Additionally, a real movement system does not sample data from a training set but rather has to move in order to receive new data. Thus, training data is always temporally correlated, and learning must be able to cope with this. An example of such a learning task is given in Figure 4 where a simulated 2 DOF robot arm has to learn to draw the figure "8" in two different regions of the work space at a moderate speed (1.5 sec duration). In this example, we assume that the correct movement plan exists, but that the inverse dynamics model which is to be used to control this movement has not been acquired. The robot is first trained for 10 minutes (real movement time) in the region of the lower target trajectory where it performs a variety of rhythmic movements under simple PID control. The initial performance of this controller is shown in the bottom part of Figure 4a. This training enables the robot to learn the locally appropriate inverse dynamics model, a ~6 ~ ~2 continuous mapping. Subsequent perFrom Isolation to Cooperation: An Alternative View of a System of Experts 611 0.5 0.' t GralMy 0.' 0.2 ~ 8 0.1 ~t Z 8 8 ..,. ~. ·0.4 ~.5 (a) (b) (0) 0 0.1 0.2 0.3 0.4 0.!5 Figure 4: Learning to draw the figure "8" with a 2-joint arm: (a) Performance of a PID controller before learning (the dimmed lines denote the desired trajectories, the solid lines the actual performance); (b) Performance after learning using a PD controller with feedforward commands from the learned inverse model; (c) Performance of the learned controller after training on the upper "8" of (b) (see text for more explanations). formance using this inverse model for control is depicted in the bottom part of Figure 4b. Afterwards, the same training takes place in the region of the upper target trajectory in order to acquire the inverse model in this part of the world. The figure "8" can then equally well be drawn there (upper part of Figure 4a,b). Switching back to the bottom part of the work space (Figure 4c), the first task can still be performed as before. No interference is recognizable. Thus, the robot could learn fast and reliably to fulfill the two tasks. It is important to note that the data generated by the training movements did not always have locally full rank. All the parameters of RFWR were necessary to acquire the local inverse model appropriately. A total of 39 locally linear experts were generated. 4. DISCUSSION We have introduced an incremental learning algorithm, RFWR, which constructs a network of isolated experts for supervised learning of regression tasks. Each expert determines a locally linear model, a local distance metric, and local bias parameters by incrementally minimizing a penalized local cross validation error. Our algorithm differs from other local learning techniques by entirely avoiding competition among the experts, and by being based on nonparametric instead of parametric statistics. The resulting properties of RFWR are a) avoidance of interference in the case of changing input distributions, b) fast incremental learning by means of Newton and second order gradient descent methods, c) analyzable asymptotic properties which facilitate the selection of the fit parameters, and d) local feature detection and dimensionality reduction. The isolated experts are also ideally suited for parallel implementations. Future work will investigate computationally less costly delta-rule implementations of RFWR, and how well RFWR scales in higher dimensions. 5. REFERENCES Atkeson, C. G., Moore, A. W., & Schaal, S. (submitted). "Locally weighted learning." Artificial Intelligence Review. Atkeson, C. G. (1992). "Memory-based approaches to approximating continuous functions." In: Casdagli, M., & Eubank, S. (Eds.), Nonlinear Modeling and Forecasting, pp.503-521. Addison Wesley. Belsley, D. A., Kuh, E., & Welsch, R. E. (1980). Regression diagnostics: Identifying influential data and sources ofcollinearity. New York: Wiley. Cleveland, W. S. (1979). "Robust locally weighted regression and smoothing scatterplots." J. American Stat. Association, 74, pp.829-836. de Boor, C. (1978). A practical guide to splines. New York: Springer. Hastie, T. J., & Tibshirani, R. J. (1990). Generalized additive models. London: Chapman and Hall. Jacobs, R. A., Jordan, M. I., Nowlan, S. J., & Hinton, G. E. (1991). "Adaptive mixtures of local experts." Neural Computation, 3, pp.79-87. Jordan, M. I., & Jacobs, R. (1994). "Hierarchical mixtures of experts and the EM algorithm." Neural Computation, 6, pp.79-87. Ljung, L., & S_derstr_m, T. (1986). Theory and practice of recursive identification. Cambridge, MIT Press. McLachlan, G. J., & Basford, K. E. (1988). Mixture models. New York: Marcel Dekker. Nadaraya, E. A. (1964). "On estimating regression." Theor. Prob. Appl., 9, pp.141-142. Schaal, S., & Atkeson, C. G. (l994b). "Assessing the quality of learned local models." In: Cowan, J. ,Tesauro, G., & Alspector, J. (Eds.), Advances in Neural Information Processing Systems 6. Morgan Kaufmann. Scott, D. W. (1992). Multivariate Density Estimation. New York: Wiley. Sutton, R. S. (1992). "Gain adaptation beats least squares." In: Proc. of 7th Yale Workshop on Adaptive and Learning Systems, New Haven, CT. Wolpert, D. H. (1990). "Stacked genealization." Los Alamos Technical Report LA-UR-90-3460.
|
1995
|
111
|
1,012
|
Fast Learning by Bounding Likelihoods in Sigmoid Type Belief Networks Tommi Jaakkola tommi@psyche.mit.edu Lawrence K. Saul lksaul@psyche.mit.edu Michael I. Jordan jordan@psyche.mit.edu Department of Brain and Cognitive Sciences Massachusetts Institute of Technology Cambridge, MA 02139 Abstract Sigmoid type belief networks, a class of probabilistic neural networks, provide a natural framework for compactly representing probabilistic information in a variety of unsupervised and supervised learning problems. Often the parameters used in these networks need to be learned from examples. Unfortunately, estimating the parameters via exact probabilistic calculations (i.e, the EM-algorithm) is intractable even for networks with fairly small numbers of hidden units. We propose to avoid the infeasibility of the E step by bounding likelihoods instead of computing them exactly. We introduce extended and complementary representations for these networks and show that the estimation of the network parameters can be made fast (reduced to quadratic optimization) by performing the estimation in either of the alternative domains. The complementary networks can be used for continuous density estimation as well. 1 Introduction The appeal of probabilistic networks for knowledge representation, inference, and learning (Pearl, 1988) derives both from the sound Bayesian framework and from the explicit representation of dependencies among the network variables which allows ready incorporation of prior information into the design of the network. The Bayesian formalism permits full propagation of probabilistic information across the network regardless of which variables in the network are instantiated. In this sense these networks can be "inverted" probabilistically. This inversion, however, relies heavily on the use of look-up table representations Fast Learning by Bounding Likelihoods in Sigmoid Type Belief Networks 529 of conditional probabilities or representations equivalent to them for modeling dependencies between the variables. For sparse dependency structures such as trees or chains this poses no difficulty. In more realistic cases of reasonably interdependent variables the exact algorithms developed for these belief networks (Lauritzen & Spiegelhalter, 1988) become infeasible due to the exponential growth in the size of the conditional probability tables needed to store the exact dependencies. Therefore the use of compact representations to model probabilistic interactions is unavoidable in large problems. As belief network models move away from tables, however, the representations can be harder to assess from expert knowledge and the important role of learning is further emphasized. Compact representations of interactions between simple units have long been emphasized in neural networks. Lacking a thorough probabilistic interpretation, however, classical feed-forward neural networks cannot be inverted in the above sense; e.g. given the output pattern of a feed-forward neural network it is not feasible to compute a probability distribution over the possible input patterns that would have resulted in the observed output. On the other hand, stochastic neural networks such as Boltzman machines admit probabilistic interpretations and therefore, at least in principle, can be inverted and used as a basis for inference and learning in the presence of uncertainty. Sigmoid belief networks (Neal, 1992) form a subclass of probabilistic neural networks where the activation function has a sigmoidal form - usually the logistic function. Neal (1992) proposed a learning algorithm for these networks which can be viewed as an improvement ofthe algorithm for Boltzmann machines. Recently Hinton et al. (1995) introduced the wake-sleep algorithm for layered bi-directional probabilistic networks. This algorithm relies on forward sampling and has an appealing coding theoretic motivation. The Helmholtz machine (Dayan et al., 1995), on the other hand, can be seen as an alternative technique for these architectures that avoids Gibbs sampling altogether. Dayan et al. also introduced the important idea of bounding likelihoods instead of computing them exactly. Saul et al. (1995) subsequently derived rigorous mean field bounds for the likelihoods. In this paper we introduce the idea of alternative - extended and complementary - representations of these networks by reinterpreting the nonlinearities in the activation function. We show that deriving likelihood bounds in the new representational domains leads to efficient (quadratic) estimation procedures for the network parameters. 2 The probability representations Belief networks represent the joint probability of a set of variables {S} as a product of conditional probabilities given by n P(St, ... , Sn) = IT P(Sk Ipa[k]), (1) k=l where the notation pa[k], "parents of Sk", refers to all the variables that directly influence the probability of Sk taking on a particular value (for equivalent representations, see Lauritzen et al. 1988). The fact that the joint probability can be written in the above form implies that there are no "cycles" in the network; i.e. there exists an ordering of the variables in the network such that no variable directly influences any preceding variables. In this paper we consider sigmoid belief networks where the variables S are binary 530 T. JAAKKOLA, L. K. SAUL, M. I. JORDAN (0/1), the conditional probabilities have the form P(Ss:lpa[i]) = g( (2Ss: - 1) L WS:jSj) j (2) and the weights Wij are zero unless Sj is a parent of Si, thus preserving the feedforward directionality of the network. For notational convenience we have assumed the existence of a bias variable whose value is clamped to one. The activation function g(.) is chosen to be the cumulative Gaussian distribution function given by 1 jX .l ~ 1 1 00 .l( )~ g(x) = -e- 2 Z dz = -e- 2 z-x dz (3) ..j2; - 00 ..j2; 0 Although very similar to the standard logistic function, this activation function derives a number of advantages from its integral representation. In particular, we may reinterpret the integration as a marginalization and thereby obtain alternative representations for the network. We consider two such representations. We derive an extended representation by making explicit the nonlinearities in the activation function. More precisely, P(Silpa[i]) g( (2Si - 1) L WijSj) j (4) This suggests defining the extended network in terms of the new conditional probabilities P(Si, Zs:lpa[i]). By construction then the original binary network is obtained by marginalizing over the extra variables Z. In this sense the extended network is (marginally) equivalent to the binary network. We distinguish a complementary representation from the extended one by writing the probabilities entirely in terms of continuous variables!. Such a representation can be obtained from the extended network by a simple transformation of variables. The new continuous variables are defined by Zs: = (2Si - l)Zi, or, equivalently, by Zi = IZs: I and Si = O( Zs:) where 0(·) is the step function. Performing this transformation yields P(Z-'I [.]) - _1_ -MZi-L: . Wij9(Zj)1~ I pa z rn=e J V 211" (5) which defines a network of conditionally Gaussian variables. The original network in this case can be recovered by conditional marginalization over Z where the conditioning variables are O(Z). Figure 1 below summarizes the relationships between the different representations. As will become clear later, working with the alternative representations instead of the original binary representation can lead to more flexible and efficient (leastsquares) parameter estimation. 3 The learning problem We consider the problem of learning the parameters of the network from instantiations of variables contained in a training set. Such instantiations, however, need not 1 While the binary variables are the outputs of each unit the continuous variables pertain to the inputs - hence the name complementary. Fast Learning by Bounding Likelihoods in Sigmoid Type Belief Networks Extended network ___ --~-::a-z_.::~ ~ _::s. Z) Original network over {S} "tr:sfonnation of ~ariables Complementary network over {Z} Figure 1: The relationship between the alternative representations. 531 be complete; there may be variables that have no value assignments in the training set as well as variables that are always instantiated. The tacit division between hidden (H) and visible (V) variables therefore depends on the particular training example considered and is not an intrinsic property of the network. To learn from these instantiations we adopt the principle of maximum likelihood to estimate the weights in the network. In essence, this is a density estimation problem where the weights are chosen so as to match the probabilistic behavior of the network with the observed activities in the training set. Central to this estimation is the ability to compute likelihoods (or log-likelihoods) for any (partial) configuration of variables appearing in the training set. In other words, if we let XV be the configuration of visible or instantiated variables2 and XH denote the hidden or uninstantiated variables, we need to compute marginal probabilities of the form (6) XH If the training samples are independent, then these log marginals can be added to give the overall log-likelihood of the training set 10gP(training set) = L:logP(XVt) (7) Unfortunately, computing each of these marginal probabilities involves summing (integrating) over an exponential number of different configurations assumed by the hidden variables in the network. This renders the sum (integration) intractable in all but few special cases (e.g. trees and chains). It is possible, however, to instead find a manageable lower bound on the log-likelihood and optimize the weights in the network so as to maximize this bound. To obtain such a lower bound we resort to Jensen's inequality: 10gP(Xv) 10gL p(XH,XV) = 10gLQ(XH)P(XH,;V) XH XH Q(X ) > ~Q(XH)1 p(XH,XV) (8) f; og Q(XH) Although this bound holds for all distributions Q(X) over the hidden variables, the accuracy of the bound is determined by how closely Q approximates the posterior distribution p(XH IXv) in terms of the Kullback-Leibler divergence; if the approximation is perfect the divergence is zero and the inequality is satisfied with equality. Suitable choices for Q can make the bound both accurate and easy to compute. The feasibility of finding such Q, however, is highly dependent on the choice of the representation for the network. 2To postpone the issue of representation we use X to denote 5, {5, Z}, or Z depending on the particular representation chosen. 532 T. JAAKKOLA, L. K. SAUL, M. I. JORDAN 4 Likelihood bounds in different representations To complete the derivation of the likelihood bound (equation 8) we need to fix the representation for the network. Which representation to select, however, affects the quality and accuracy of the bound. In addition, the accompanying bound of the chosen representation implies bounds in the other two representational domains as they all code the same distributions over the observables. In this section we illustrate these points by deriving bounds in the complementary and extended representations and discuss the corresponding bounds in the original binary domain. Now, to obtain a lower bound we need to specify the approximate posterior Q. In the complementary representation the conditional probabilities are Gaussians and therefore a reasonable approximation (mean field) is found by choosing the posterior approximation from the family of factorized Gaussians: Q(Z) = IT _1_e-(Zi-hi)~/2 (9) i..;?:; Substituting this into equation 8 we obtain the bound log P(S*) ~ -~ L (hi - Ej Jij g(hj»2 - ~ L Ji~g(hj )g(-hj ) (10) i ij The means hi for the hidden variables are adjustable parameters that can be tuned to make the bound as tight as possible. For the instantiated variables we need to enforce the constraints g( hi) = S: to respect the instantiation. These can be satisfied very accurately by setting hi = 4(2S: - 1). A very convenient property of this bound and the complementary representation in general is the quadratic weight dependence - a property very conducive to fast learning. Finally, we note that the complementary representation transforms the binary estimation problem into a continuous density estimation problem. We now turn to the interpretation of the above bound in the binary domain. The same bound can be obtained by first fixing the inputs to all the units to be the means hi and then computing the negative total mean squared error between the fixed inputs and the corresponding probabilistic inputs propagated from the parents. The fact that this procedure in fact gives a lower bound on the log-likelihood would be more difficult to justify by working with the binary representation alone. In the extended representation the probability distribution for Zi is a truncated Gaussian given Si and its parents. We therefore propose the partially factorized posterior approximation: (11) where Q(ZiISi) is a truncated Gaussian: Q(Zi lSi) = 1 _1_e- t(Zi-(2S,-1)hi)~ g« 2Si- 1)hi ) ..;?:; (12) As in the complementary domain the resulting bound depends quadratically on the weights. Instead of writing out the bound here, however, it is more informative to see its derivation in the binary domain. A factorized posterior approximation (mean field) Q(S) = n. q~i(1 - qi)l-S, for the binary network yields a bound I I 10gP(S*) > L {(Si 10gg(Lj J,jSj») + (1- Si) 10g(l- 9(L; Ji;S;»)} i Fast Learning by Bounding Likelihoods in Sigmoid Type Belief Networks 533 (13) where the averages (.) are with respect to the Q distribution. These averages, however, do not conform to analytical expressions. The tractable posterior approximation in the extended domain avoids the problem by implicitly making the following Legendre transformation: 1 2 1 2 1 2 logg(x) = ["2x + logg(x)] -"2x ~ AX - G(A) - "2x (14) which holds since x 2/2 + logg(x) is a convex function. Inserting this back into the relevant parts of equation 13 and performing the averages gives 10gP(S*) > L {[qjAj - (1- qj),Xd Lhjqj - qjG(Ai) - (1- qj)G('xi)} j I", 2 1",2 ( -"2(L.,.. Jijqj) -"2 L.,.. Jjjqj 1- gj) j ij (15) which is quadratic in the weights as expected. The mean activities q for the hidden variables and the parameters A can be optimized to make the bound tight. For the instantiated variables we set qi = S; . 5 Numerical experiments To test these techniques in practice we applied the complementary network to the problem of detecting motor failures from spectra obtained during motor operation (see Petsche et al. 1995). We cast the problem as a continuous density estimation problem. The training set consisted of 800 out of 1283 FFT spectra each with 319 components measured from an electric motor in a good operating condition but under varying loads. The test set included the remaining 483 FFTs from the same motor in a good condition in addition to three sets of 1340 FFTs each measured when a particular fault was present. The goal was to use the likelihood of a test FFT with respect to the estimated density to determine whether there was a fault present in the motor. We used a layered 6 -+ 20 -+ 319 generative model to estimate the training set density. The resulting classification error rates on the test set are shown in figure 2 as a function of the threshold likelihood. The achieved error rates are comparable to those of Petsche et al. (1995). 6 Conclusions Network models that admit probabilistic formulations derive a number of advantages from probability theory. Moving away from explicit representations of dependencies, however, can make these properties harder to exploit in practice. We showed that an efficient estimation procedure can be derived for sigmoid belief networks, where standard methods are intractable in all but a few special cases (e.g. trees and chains). The efficiency of our approach derived from the combination of two ideas. First, we avoided the intractability of computing likelihoods in these networks by computing lower bounds instead. Second, we introduced new representations for these networks and showed how the lower bounds in the new representational domains transform the parameter estimation problem into 534 0.0 ..... 0.8 0.7 0.8 ',_ fo.s "\ D:' ' , ... .. d.. ''', 0.3 " --0.2 . .. '0.1 , . , , , , , , , '. " , , . T. JAAKKOLA, L. K. SAUL, M. 1. JORDAN Figure 2: The probability of error curves for missing a fault (dashed lines) and misclassifying a good motor (solid line) as a function of the likelihood threshold. quadratic optimization. Acknowledgments The authors wish to thank Peter Dayan for helpful comments. This project was supported in part by NSF grant CDA-9404932, by a grant from the McDonnellPew Foundation, by a grant from ATR Human Information Processing Research Laboratories, by a grant from Siemens Corporation, and by grant N00014-94-10777 from the Office of Naval Research. Michael I. Jordan is a NSF Presidential Young Investigator. References P. Dayan, G. Hinton, R. Neal, and R. Zemel (1995). The helmholtz machine. Neural Computation 7: 889-904. A. Dempster, N. Laird, and D. Rubin. Maximum likelihood from incomplete data via the EM algorithm (1977). J. Roy. Statist. Soc. B 39:1-38. G. Hinton, P. Dayan, B. Frey, and R. Neal (1995). The wake-sleep algorithm for unsupervised neural networks. Science 268: 1158-1161. S. L. Lauritzen and D. J. Spiegelhalter (1988). Local computations with probabilities on graphical structures and their application to expert systems. J. Roy. Statist. Soc. B 50:154-227. R. Neal. Connectionist learning of belief networks (1992). Artificial Intelligence 56: 71-113. J. Pearl (1988). Probabilistic Reasoning in Intelligent Systems. Morgan Kaufmann: San Mateo. T. Petsche, A. Marcantonio, C. Darken, S. J. Hanson, G. M. Kuhn, I. Santoso (1995). A neural network autoassociator for induction motor failure prediction. In Advances in Neural Information Processing Systems 8. MIT Press. 1. K. Saul, T. Jaakkola, and M. I. Jordan (1995). Mean field theory for sigmoid belief networks. M.l. T. Computational Cognitive Science Technical Report 9501.
|
1995
|
112
|
1,013
|
Neural Control for Nonlinear Dynamic Systems Ssu-Hsin Yu Department of Mechanical Engineering Massachusetts Institute of Technology Cambridge, MA 02139 Email: hsin@mit.edu Anuradha M. Annaswamy Department of Mechanical Engineering Massachusetts Institute of Technology Cambridge, MA 02139 Email: aanna@mit.edu Abstract A neural network based approach is presented for controlling two distinct types of nonlinear systems. The first corresponds to nonlinear systems with parametric uncertainties where the parameters occur nonlinearly. The second corresponds to systems for which stabilizing control structures cannot be determined. The proposed neural controllers are shown to result in closed-loop system stability under certain conditions. 1 INTRODUCTION The problem that we address here is the control of general nonlinear dynamic systems in the presence of uncertainties. Suppose the nonlinear dynamic system is described as x= f(x , u , 0) , y = h(x, u, 0) where u denotes an external input, y is the output, x is the state, and 0 is the parameter which represents constant quantities in the system. The control objectives are to stabilize the system in the presence of disturbances and to ensure that reference trajectories can be tracked accurately, with minimum delay. While uncertainties can be classified in many different ways, we focus here on two scenarios. One occurs because the changes in the environment and operating conditions introduce uncertainties in the system parameter O. As a result, control objectives such as regulation and tracking, which may be realizable using a continuous function u = J'(x, 0) cannot be achieved since o is unknown. Another class of problems arises due to the complexity of the nonlinear function f. Even if 0, f and h can be precisely determined, the selection of an appropriate J' that leads to stabilization and tracking cannot be made in general. In this paper, we present two methods based on neural networks which are shown to be applicable to both the above classes of problems. In both cases, we clearly outline the assumptions made, the requirements for adequate training of the neural network, and the class of engineering problems where the proposed methods are applicable. The proposed approach significantly expands the scope of neural controllers in relation to those proposed in (Narendra and Parthasarathy, 1990; Levin and Narendra, 1993; Sanner and Slotine, 1992; Jordan and Rumelhart, 1992). Neural Control for Nonlinear Dynamic Systems 1011 The first class of problems we shall consider includes nonlinear systems with parametric uncertainties. The field of adaptive control has addressed such a problem, and over the past thirty years, many results have been derived pertaining to the control of both linear and nonlinear dynamic systems (Narendra and Annaswamy, 1989). A common assumption in almost all of the published work in this field is that the uncertain parameters occur linearly. In this paper, we consider the control of nonlinear dynamic systems with nonlinear parametrizations. We design a neural network based controller that adapts to the parameter o and show that closed-loop system stability can be achieved under certain conditions. Such a controller will be referred to as a O-adaptive neural controller. Pertinent results to this class are discussed in section 2. The second class of problems includes nonlinear systems, which despite being completely known, cannot be stabilized by conventional analytical techniques. The obvious method for stabilizing nonlinear systems is to resort to linearization and use linear control design methods. This limits the scope of operation of the stabilizing controller. Feedback linearization is another method by which nonlinear systems can be stably controlled (lsidori, 1989). This however requires fairly stringent set of conditions to be satisfied by the functions! and h. Even after these conditions are satisfied, one cannot always find a closed-form solution to stabilize the system since it is equivalent to solving a set of partial differential equations. We consider in this paper, nonlinear systems, where system models as well as parameters are known, but controlIer structures are unknown. A neural network based controller is shown to exist and trained so that a stable closed-loop system is achieved. We denote this class of controllers as a stable neural controller. Pertinent results to this class are discussed in section 3. 2 O-ADAPTIVE NEURAL CONTROLLER The focus of the nonlinear adaptive controller to be developed in this paper is on dynamic systems that can be written in the d-step ahead predictor form as follows: Yt+d = !r(Wt,Ut,O) (I) where wi = [Yt," . ,Yt-n+l, Ut-I, ' ", Ut-m-d+l], n ~ I, m ~ 0, d ~ I, m + d = n, YI, UI C ~ containing the origin and 8 1 C ~k are open, ir : Y1 x U;n+d x 8 1 -t ~, Yt and Ut are the output and the input of the system at time t respectively, and 0 is an unknown parameter and occurs nonlinearly in ir.1 The goal is to choose a control input 'It such that the system in (1) is stabilized and the plant output is regulated around zero. Letxi ~ [Yt+d-I , '" ,Yt+l ,wil T , Am = [e2,"', en+d-I, 0, en+d+I,"', en+m+2d-2, 0], Bm = [el' en+d], where e, is an unit vector with the i-th term equal to I. The following assumptions are made regarding the system in Eq. (I ). (AI) For every 0 E 8 1, ir(O,O, O) = 0. CA2) There exist open and convex neighborhoods of the origin Y2 C YI and U2 C UI, an open and convex set 82 C 8 1, and a function K : 0.2 x Y2 x 8 2 ---> U I such that for every Wt E 0.2, Yt+d E Y2 and 0 E 8 2, Eq. (1) can be written as Ut = K(wt, Yt+d, 0), where 0.2 ~ Y2 X u;,,+d-I. (A3) K is twice differentiable and has bounded first and second derivatives on EI ~ 0.2 X Y2 X 8 2, while ir is differentiable and has a bounded derivative on 0.2 x I{ (E I ) x 8 2. (A4) There exists bg > ° such that for every YI E ir(o.2, K(o.2' 0, 8 2), 8 2), W E 0.2 and o BE 8 11 - (8K(w,y ,O) _ 8K(w,y,9))1 _ . 8f,(w ,u ,O) I - I > b ,2, ay ay Y YI au U-UI g' 1 Here, as well as in the following sections, An denotes the n-th product space of the set A. 1012 S. YU, A. M. ANNASW AMY (A5) There exist positive definite matrices P and Q of dimensions (n + m + 2d - 2) such T T' -T T T T T that xt (AmPAm - P)Xt+ J( BrnPBmK + 2xt ArnPBmK ~ -Xt QXt, where [( = [0, K(wt, 0, O)]T. Since the objective is to control the system ~n (1) where 0 is unknown, in order to stabilize the output Y at the origin with an estimate Of, we choose the control input as (2) 2.1 PARAMETER ESTIMATION SCHEME Suppose the estimation algorithm for updating Ot is defined recursively as /10t ~ OtOt-I = R(Yt,Wt-d,Ut-d,Ot-l) the problem is to determine the function R such that Ot converges to 0 asymptotical1y. In general, R is chosen to depend on Yt, Wt-d, 1£t-d and Ot-l since they are measurable and contain information regarding O. For example, in the case of linear systems which can be cast in the input predictor form, 1£t = <b[ 0, a wel\known linear parameter estimation method is to adjust /10 as (Goodwin and Sin, 1984) /10t = 1+4>'t~~t-'/ [1£t-d - ¢LdOt-d· In other words, the mechanism for carrying out parameter estimation is realized by R. In the case of general nonlinear systems, the task of determining such a function R is quite difficult, especial\y when the parameters occur nonlinearly. Hence, we propose the use of a neural network parameter estimation algorithm denoted O-adaptive neural network (TANN) (Annaswamy and Yu, 1996). That is, we adjust Ot as if /1Vd, < -f otherwise (3) where the inputs of the neural network are Yt, Wt-d, 1£t-d and Ot-I, the output is /10t, and f defines a dead-zone where parameter adaptation stops. The neural network is to be trained so that the resulting network can improve the parameter estimation over time for any possible 0 in a compact set. In addition, the trained network must ensure that the overal1 system in Eqs. (1), (2) and (3) is stable. Toward this end, N in TANN algorithm is required to satisfy the fol1owing two properties: (PI) IN(Yt,wt- d,1£t- d,Ot-l)12 ~ a(llfJt;~~,~1~2)2uLd' and (P2) /1Vt -/1Vd, < fl, > 0 where A Tf Iii 12 _ Iii 12 ii II _ II AV, -a 2+IC( <f;,_,,)1 1 1£-2 fl , L.l.Vt Ut Ut- I, Ut Ut u, L.l. d, ( IC( )12)2 t- d' 1+ , 4>,-./ C(¢t) = (~~ (Wt,Yt+d,O)lo=oo)T, Ut = Ut - K(Wt,Yt+d,Ot+d- I), (fit = [WT,Yt+djT, a E (0, I) and 00 is the point where K is linearized and often chosen to be the mean value of parameter variation. 2.2 TRAINING OF TANN FOR CONTROL In the previous section, we proposed an algorithm using a neural network for adjusting the control parameters. We introduced two properties (PI) and (P2) of the identification algorithm that the neural network needs to possess in order to maintain stability of the closed-loop system. In this section, we discuss the training procedure by which the weights of the neural network are to be adjusted so that the network retains these properties, The training set is constructed off-line and should compose of data needed in the training phase. If we wan..!. the algorithm in Eq. (3) to be valid on the specified sets Y3 and U3 for various 0 and 0 in 83, the training set should cover those variables appearing in Eq. (3) in their respective ranges. Hence, we first sample W in the set Y;- x U;:+d-I, Neural Control for Nonlinear Dynamic Systems 1013 and B, 8 in the set 83. Their values are, say, WI, BI and 81 respectively. For the particular fh and BI we sample 8 again in the set {B E 8 31 IB BII :s: 181 BI I}, and its value is, say, 8t. Once WI, BI , 81 and 8t are sampled, other data can then be calculated, such as UI = K(WI' 0, 8d and YI = fr(WI, UI, Bd. We can also ob. th d· C(A-) BK ( B) All: 2+IC(¢dI2 ( ~)2 d tam ecorrespon mg '1'1 - ao WI , YI, 0, il d l -a(I+IC(¢I)i2)2 UI -'ttl an _ IC(¢IW ~ 2 _ T T ~ _ ~d LI a (1+IC(<I>I)I2)2 (UI UI) ,where ¢I [WI ,yJ} and UI K(WI' YI,( 1 )· A data element can then be formed as (Yl ,WI ,UI, 8t, BI , ~ Vd l , Ld. Proceeding in the same manner, by choosing various ws , Bs , 1f. and 8~ in their respective ranges, we form a typical training set Ttram = { (Ys , W s, us,1f~ , Bs, ~ Vd d Ls) 11 :s: s :s: M}, where M denotes the total number of patterns in the training set. If the quadratic penalty function method (Bertsekas, 1995) is used, properties (PI) and (P2) can be satisfied by training the network on the training set to minimize the following cost function: M mJpl ~ mJ,n~~{(max{0 , ~VeJ)2+ ;2 (max{0, INi(W)12 - L t})2} (4) To find a W which minimizes the above unconstrained cost function 1, we can apply algorithms such as the gradient method and the Gauss-Newton method. 2.3 STABILITY RESULT With the plant given by Eq. (1), the controller by Eq. (2), and the TANN parameter estimation algorithm by Eq. (3), it can be shown that the stability of the closed-loop system is guaranteed. Based on the assumptions of the system in (1) and properties (PI) and (P2) that TANN satisfies, the stability result of the closed-loop system can be concluded in the following theorem. We refer the reader to (Yu and Annaswamy, 1996) for further detail. Theorem 1 Given the compact sets Y;+ I X U:;+d x 8 3 where the neural network in Eq. (3) is trained. There exist EI, E > 0 such that for any interior point B of 8 3, there exist open sets Y4 C Y3, U4 C U3 and a neighborhood 8 4 of B such that if Yo , ... , Yn+d-2 E Y4, Uo, .. . , U n -2 E U4, and 8n - l , ... ,8n+d - 2 E 8 4, then all the signals in the closed-loop system remain bounded and Yt converges to a neighborhood of the origin. 2.4 SIMULATION RESULTS In this section, we present a simulation example of the TANN controller proposed in this . Th . f h f lIy, ( I-y,) h B· h b sectton. e system IS 0 t e orm Yt+1 = I+e U.USH", + Ut, were IS t e parameter to e determined on-line. Prior information regarding the system is that () E [4, 10]. Based on 8 (I) ~ Eq. (2), the controller was chosen to be Ut = ,y, 0 ,-;'Y' ' where Bt denotes the parameter I+e- · "" estimate at time t. According to Eq. (3), B was estimated using the TANN algorithm with inputs YHI, Yt. Ut and~, and E = 0.01. N is a Gaussian network with 700 centers. The training set and the testing set were composed of 6,040 and 720 data elements respectively. After the training was completed, we tested the TANN controller on the system with six different values of B, 4.5, 5.5, 6.5, 7.~, 8.5 and 9.5, while the initial parameter estimate and the initial output were chosen as BI = 7 and Yo = -0.9 respectively. The results are plotted in Figure 1. It can be seen that Yt can be stabilized at the origin for all these values of B. For comparison, we also simulated the system under the same conditions but with 8 1014 -1 -2 o 50 10 100 4 ~ 0 -1 -2 o 50 S. YU, A. M. ANNASWAMY 10 100 4 Figure 1: Yt (TANN Controller) Figure 2: Yt (Extended Kalman Filter) estimated using the extended Kalman filter (Goodwin and Sin, 1984). Figure 2 shows the output responses. It is not surprising that for some values of fJ, especially when the initial estimation error is large, the responses either diverge or exhibit steady state error. 3 STABLE NEURAL CONTROLLER 3.1 STATEMENT OF THE PROBLEM Consider the following nonlinear dynamical system X= j(x,u), Y = h(x) (5) where x E Rn and u E RTn. Our goal is to construct a stabilizing neural controller as u = N(y; W) where N is a neural network with weights W, and establish the conditions under which the closed-loop system is stable. The nonlinear system in (5) is expressed as a combination of a higher-order linear part and a nonlinear part as x= Ax + Bu + RI (x, u) and y = Cx + R 2(x), where j(O,O) = 0 and h(O) = O. We make the following assumptions: (AI) j, h are twice continuously differentiable and are completely known. (A2) There exists a K such that (A - BKC) is asymptotically stable. 3.2 TRAINING OF THE STABLE NEURAL CONTROLLER In order for the neural controller in Section 3.1 to result in an asymptotically stable c1osedloop system, it is sufficient to establish that a continuous positive definite function of the state variables decreases monotonically through output feedback. In other words, if we can find a scalar definite function with a negative definite derivative of all points in the state space, we can guarantee stability of the overall system. Here, we limit our choices of the Lyapunov function candidates to the quadratic form, i.e. V = xT Px, where P is positive definite, and the goal is to choose the controller so that V < 0 where V = 2xT P j(x, N(h(x), W)). Based on the above idea, we define a "desired" time-derivative V d as V d= -xTQx where Q = QT > O. We choose P and Q matrices as follows. First, according to (AI), we can find a matrix K to make (A - BKC) asymptotically stable. We can then find a (P, Q) pair by choosing an arbitrary positive definite matrix Q and solving the Lyapunov equation, (A - BKC)T P + P(A - BKC) = -Q to obtain a positive definite P. Neural Control for Nonlinear Dynamic Systems 1015 With the contro\1er of the form in Section 3.1, the goal is to find W in the neural network which yields V:::; V d along the trajectories in a neighborhood X C ~n of the origin in the state space. Let Xi denote the value of a sample point where i is an index to the sample variable X E X in the state space. To establish V:::; V d, it is necessary that for every Xl in a neighborhood X C ~n of the origin, Vi:::;Vd" where Vi= 2x;Pf(x l ,N(h(:rl ) , W)) and V d, = -x; QXi . That is, the goal is to find a W such that the inequality constraints tlVe , :::; 0, where i = 1,··· , M, is satisfied, where tlVe , =V l V d, and M denotes the total number of sample points in X. As in the training of TANN controller, this can also be posed as an optimization problem. If the same quadratic penalty function method is used, the problem is to find W to minimize the fo\1owing cost function over the training set, which is described as Ttrain = {(Xl' Yi, V d.)\l :::; i :::; M}: 1M rwn J 6. mJp 2 I: (max{O, tlVe,})2 (6) i= 1 3.3 STABILITY OF THE CLOSED-LOOP SYSTEM Assum~tions (A 1) and (A2) imply that a stabilizing controller u = - J( y exists so that V = X Px is a candidate Lyapunov function. More genera\1y, suppose a continuous but unknown function ,,((y) exists such that for V = xT Px, a control input 1t = "((y) leads to V:::; -xT Qx, then we can find a neural network N (y) which approximates "((y) arbitrarily closely in a compact set leading to closed-loop stability. This is summarized in Theorem 2 (Yu and Annaswamy, 1995). Theorem 2 Let there be a continuous function "((h(x)) such that 2xT P f(x , "((h(x))) + xT Qx :::; 0 for every X E X where X is a compact set containing the origin as an interior point. Then, given a neighborhood 0 C X of the origin, there exists a neural controlierH = N(h(x); W) and a compact set Y E X such that the solutions of x= f(x , N(h(x); W)) converge to 0, for every initial condition x(to) E y. 3.4 SIMULATION RESULTS In this section, we show simulation results for a discrete-time nonlinear systems using the proposed neural network contro\1er in Section 3, and compare it with a linear contro\1er to illustrate the difference. The system we considered is a second-order nonlinear system Xt = f(xt-I , Ut-I), where f = [II, 12]T, h = Xl t _ 1 X (1 +X2'_1 )+X2t-1 x (l-ut- I +uLI) and 12 = XT'_I + 2X2'_1 +Ut-I (1 + X2'_I)· It was assumed that X is measurable, and we wished to stabilize the system around the origin. The controller is of the form Ht = N (x It, X2 t ). The neural network N used is a Gaussian network with 120 centers. The training set and the testing set were composed of 441 and 121 data elements respectively. After the training was done, we plotted the actual change of the Lyapunov function, tl V, using the linear controller U = - K x and the neural network controller in Figures 3 and 4 respectively. It can be observed from the two figures that if the neural network contro\1er is used, tl V is negative definite except in a small neighborhood of the origin, which assures that the closed-loop system would converge to vicinity of the origin; whereas, if the linear controller is used, tl V becomes positive in some region away from the origin, which implies that the system can be unstable for some initial conditions. Simulation results confirmed our observation. 1016 S. YU, A. M. ANNASW AMY -0 01 - 0 J - 0 I -0 J -()2 -O J Figure 3: ~V(u = -Kx ) Figure 4: ~V(u = N(x)) Acknowledgments This work is supported in part by Electrical Power Research Institute under contract No. 8060-13 and in part by National Science Foundation under grant No. ECS-9296070. References [1] A. M. Annaswamy and S. Yu. O-adaptive neural networks: A new approach to parameter estimation. IEEE Transactions on Neural Networks, (to appear) 1996. [2] D. P. Bertsekas. Nonlinear Programming. Athena Scientific, Belmont, MA, 1995. [3] G. C. Goodwin and K. S. Sin. Adaptive Filtering Prediction and Control. PrenticeHall, Inc., 1984. [4] A. Isidori. Nonlinear Control Systems. Springer-Verlag, New York, NY, 1989. [5] M. L Jordan and D. E. Rumelhart. Forward models: Supervised learning with a distal teacher. Cognitive Science, 16:307-354, 1992. [6] A. U. Levin and K. S. Narendra. Control of nonlinear dynamical systems using neural networks: Controllability and stabilization. IEEE Transactions on Neural Networks, 4(2): 192-206, March 1993. [7] K. S. Narendra and A. M. Annaswamy. Stable Adaptive Systems. Prentice-Hall, Inc., 1989. [8] K. S. Narendra and K. Parthasarathy. Identification and control of dynamical systems using neural networks. IEEE Transactions on Neural Networks, 1 (I ):4-26, March 1990. [9] R. M. Sanner and J.-J. E. Slotine. Gaussian networks for direct adaptive control. IEEE Transactions on Neural Networks, 3(6):837-863, November 1992. [10] S. Yu and A. M. Annaswamy. Adaptive control of nonlinear dynamic systems using O-adaptive neural networks. Technical Report 9601 , Adaptive Control Laboratory, Department of Mechanical Engineering, M.LT., 1996. [11] S.-H. Yu and A. M. Annaswamy. Control of nonlinear dynamic systems using a stability based neural network approach. In Technical report 9501, Adaptive Control Laboratory, MIT, Submitted to Proceedings of the 34th IEEE Conference on Decision and Control, New Orleans, LA, 1995.
|
1995
|
113
|
1,014
|
Sample Complexity for Learning Recurrent Percept ron Mappings Bhaskar Dasgupta Department of Computer Science University of Waterloo Waterloo, Ontario N2L 3G 1 CANADA bdasgupt~daisy.uwaterloo.ca Eduardo D. Sontag Department of Mathematics Rutgers University New Brunswick, NJ 08903 USA sontag~control.rutgers.edu Abstract Recurrent perceptron classifiers generalize the classical perceptron model. They take into account those correlations and dependences among input coordinates which arise from linear digital filtering. This paper provides tight bounds on sample complexity associated to the fitting of such models to experimental data. 1 Introduction One of the most popular approaches to binary pattern classification, underlying many statistical techniques, is based on perceptrons or linear discriminants; see for instance the classical reference (Duda and Hart, 1973). In this context, one is interested in classifying k-dimensional input patterns V=(Vl, . . . ,Vk) into two disjoint classes A + and A -. A perceptron P which classifies vectors into A + and A - is characterized by a vector (of "weights") C E lR k, and operates as follows. One forms the inner product C.V = CIVI + ... CkVk . If this inner product is positive, v is classified into A +, otherwise into A - . In signal processing and control applications, the size k of the input vectors v is typically very large, so the number of samples needed in order to accurately "learn" an appropriate classifying perceptron is in principle very large. On the other hand, in such applications the classes A + and A-often can be separated by means of a dynamical system of fairly small dimensionality. The existence of such a dynamical system reflects the fact that the signals of interest exhibit context dependence and Sample Complexity for Learning Recurrent Perceptron Mappings 205 correlations, and this prior information can help in narrowing down the search for a classifier. Various dynamical system models for classification appear from instance when learning finite automata and languages (Giles et. al., 1990) and in signal processing as a channel equalization problem (at least in the simplest 2-level case) when modeling linear channels transmitting digital data from a quantized source, e.g. (Baksho et. al., 1991) and (Pulford et. al., 1991). When dealing with linear dynamical classifiers, the inner product c. v represents a convolution by a separating vector c that is the impulse-response of a recursive digital filter of some order n ~ k. Equivalently, one assumes that the data can be classified using a c that is n-rec'Ursive, meaning that there exist real numbers TI, ... , Tn SO that n Cj = 2: Cj-iTi, j = n + 1, ... , k . i=1 Seen in this context, the usual perceptrons are nothing more than the very special subclass of "finite impulse response" systems (all poles at zero); thus it is appropriate to call the more general class "recurrent" or "IIR (infinite impulse response)" perceptrons. Some authors, particularly Back and Tsoi (Back and Tsoi, 1991; Back and Tsoi, 1995) have introduced these ideas in the neural network literature. There is also related work in control theory dealing with such classifying, or more generally quantized-output, linear systems; see (Delchamps, 1989; Koplon and Sontag, 1993). The problem that we consider in this paper is: if one assumes that there is an n-recursive vector c that serves to classify the data, and one knows n but not the particular vector, how many labeled samples v(i) are needed so as to be able to reliably estimate C? More specifically, we want to be able to guarantee that any classifying vector consistent with the seen data will classify "correctly with high probability" the unseen data as well. This is done by computing the VC dimension of the related concept class and then applying well-known results from computational learning theory. Very roughly speaking, the main result is that the number of samples needed is proportional to the logarithm of the length k (as opposed to k itself, as would be the case if one did not take advantage of the recurrent structure). Another application of our results, again by appealing to the literature from computational learning theory, is to the case of "noisy" measurements or more generally data not exactly classifiable in this way; for example, our estimates show roughly that if one succeeds in classifying 95% of a data set of size logq, then with confidence ~ lone is assured that the prediction error rate will be < 90% on future (unlabeled) samples. Section 5 contains a result on polynomial-time learnability: for n constant, the class of concepts introduced here is PAC learnable. Generalizations to the learning of real-valued (as opposed to Boolean) functions are discussed in Section 6. For reasons of space we omit many proofs; the complete paper is available by electronic mail from the authors. 2 Definitions and Statements of Main Results Given a set X, and a subset X of X, a dichotomy on X is a function fJ: X - {-I, I}. Assume given a class F of functions X {-I, I}, to be called the class of classifier functions. The subset X ~ X is shattered by F if each dichotomy on X is the restriction to X of some <P E F. The Vapnik-Chervonenkis dimension vc (F) is the supremum (possibly infinite) of the set of integers K for which there is some subset 206 B. DASGUPTA,E.D.SONTAG x ~ X of cardinality K, which can be shattered by:F. Due to space limitations, we omit any discussion regarding the relevance of the VC dimension to learning problems; the reader is referred to the excellent surveys in (Maass, 1994; Thran, 1994) regarding this issue. Pick any two integers n>O and q~O. A sequence C= (Cl, ... , cn+q) E lR.n+q is said to be n-recursive if there exist real numbers r1, .. . , rn so that n cn+j = 2: cn+j-iri, j = 1, . .. , q. i=l (In particular, every sequence of length n is n-recursive, but the interesting cases are those in which q i= 0, and in fact q ~ n.) Given such an n-recursive sequence C, we may consider its associated perceptron classifier. This is the map ¢c: lR.n+q --+{-1,1}: (X1, ... ,Xn+q) H sign (I:CiXi) .=1 where the sign function is understood to be defined by sign (z) = -1 if z ~ 0 and sign (z) = 1 otherwise. (Changing the definition at zero to be + 1 would not change the results to be presented in any way.) We now introduce, for each two fixed n, q as above, a class of functions: :Fn,q := {¢cl cE lR.n+q is n-recursive}. This is understood as a function class with respect to the input space X = lR. n+q, and we are interested in estimating vc (:Fn,q). Our main result will be as follows (all logs in base 2): Theorem 1 Imax {n, nLlog(L1 + ~ J)J} ~ vc (:Fn ,q) ~ min {n + q, 18n + 4n log(q + 1)} I Note that, in particular, when q> max{2 + n2 , 32}, one has the tight estimates n "2 logq ~ vc (:Fn ,q) ~ 8n logq . The organization of the rest of the paper is as follows. In Section 3 we state an abstract result on VC-dimension, which is then used in Section 4 to prove Theorem 1. Finally, Section 6 deals with bounds on the sample complexity needed for identification of linear dynamical systems, that is to say, the real-valued functions obtained when not taking "signs" when defining the maps ¢c. 3 An Abstract Result on VC Dimension Assume that we are given two sets X and A, to be called in this context the set of inputs and the set of parameter values respectively. Suppose that we are also given a function F: AxX--+{-1,1}. Associated to this data is the class of functions :F := {F(A,·): X --+ {-1, 1} I A E A} Sample Complexity for Learning Recurrent Perceptron Mappings 207 obtained by considering F as a function of the inputs alone, one such function for each possible parameter value A. Note that, given the same data one could, dually, study the class F*: {F(-,~) : A-{-I,I}I~EX} which obtains by fixing the elements of X and thinking of the parameters as inputs. It is well-known (and in any case, a consequence of the more general result to be presented below) that vc (F) ~ Llog(vc (F*»J, which provides a lower bound on vc (F) in terms of the "dual VC dimension." A sharper estimate is possible when A can be written as a product of n sets A = Al X A2 X • • . x An and that is the topic which we develop next. (1) We assume from now on that a decomposition of the form in Equation (1) is given, and will define a variation of the dual VC dimension by asking that only certain dichotomies on A be obtained from F*. We define these dichotomies only on "rectangular" subsets of A, that is, sets of the form L = Ll X .•. x Ln ~ A with each Li ~ Ai a nonempty subset. Given any index 1 ::; K ::; n, by a K-axis dichotomy on such a subset L we mean any function 6 : L {-I, I} which depends only on the Kth coordinate, that is, there is some function ¢ : Lit - {-I, I} so that 6(Al, . . . ,An ) = ¢(AIt) for all (Al, . . . ,An ) E L; an axis dichotomy is a map that is a K-axis dichotomy for some K. A rectangular set L will be said to be axisshattered if every axis dichotomy is the restriction to L of some function of the form F(·,~): A - {-I, I}, for some ~ EX. Theorem 2 If L = Ll X ... x Ln ~ A can be axis-shattered and each set Li has cardinality ri, then vc (F) ~ Llog(rt)J + ... + Llog(rn)J . (In the special case n=1 one recovers the classical result vc (F) ~ Llog(vc (F*)J.) The proof of Theorem 2 is omitted due to space limitations. 4 Proof of Main Result We recall the following result; it was proved, using Milnor-Warren bounds on the number of connected components of semi-algebraic sets, by Goldberg and Jerrum: Fact 4.1 (Goldberg and Jerrum, 1995) Assume given a function F : A x X {-I, I} and the associated class of functions F:= {F(A,·): X - {-I, I} I A E A} . Suppose that A = ~ k and X = ~ n, and that the function F can be defined in terms of a Boolean formula involving at most s polynomial inequalities in k + n variables, each polynomial being of degree at most d. Then, vc (F) ::; 2k log(8eds). 0 Using the above Fact and bounds for the standard "perceptron" model, it is not difficult to prove the following Lemma. Lemma 4.2 vc (Fn,q) ::; min{n + q, 18n + 4nlog(q + I)} Next, we consider the lower bound of Theorem 1. Lemma 4.3 vc (Fn,q) ~ maxin, nLlog(Ll + q~1 J)J} 208 B. DASGUPTA, E. D. SONTAG Proof As Fn,q contains the class offunctions <Pc with c= (C1, ... , cn, 0, ... ,0), which in turn being the set of signs of an n-dimensional linear space of functions, has VC dimension n, we know that vc (Fn,q) ~ n. Thus we are left to prove that if q > n then vc(Fn,q) ~ nLlog(l1 + ~J)J. The set of n-recursive sequences of length n + q includes the set of sequences of the following special form: n ~. 1 Cj = L.Jlf, i=l j=I, ... ,n+q (2) where ai, h E lR for each i = 1, ... , n. Hence, to prove the lower bound, it is sufficient to study the class of functions induced by F : lll.n x lll.n+. ~ {-I, I}, (~I"'" ~n, XI,···, x n+.) >-> sign (t, ~ ~i-I Xj) . Let r = L q+~-l J and let L1, ... ,Ln be n disjoint sets of real numbers (if desired, integers), each of cardinality r. Let L = U:'::l Lj . In addition, if rn < q+n-1, then select an additional set B of (q+n-rn-1) real numbers disjoint from L. We will apply Theorem 2, showing that the rectangular subset L1 x ... x Ln can be axis-shattered. Pick any,.. E {1, ... , n} and any <P : L,. ~ {-1, 1}. Consider the ( unique) interpolating polynomial n+q peA) = L XjAj- 1 j=l in A of degree q + n - 1 such that peA) = { ~(A) if A E L,. if A E (L U B) - L,.. Now pick e = (Xl, ... , Xn +q-1). Observe that F(lt, I" ... , In, Xl, .. . , xn+.) = sign (t, P(I'») = ¢(I.) for all (11, ... , In) E L1 X .•• X Ln, since p(l) = 0 fori ¢ L,. and p(l) = <P(I) otherwise. It follows from Theorem 2 that vc (Fn,q) ~ nLlog(r)J, as desired. • [) The Consistency Problem We next briefly discuss polynomial time learnability of recurrent perceptron mappings. As discussed in e.g. (Turan, 1994), in order to formalize this problem we need to first choose a data structure to represent the hypotheses in Fn,q. In addition, since we are dealing with complexity of computation involving real numbers, we must also clarify the meaning of "finding" a hypothesis, in terms of a suitable notion of polynomial-time computation. Once this is done, the problem becomes that of solving the consistency problem: Given a set ofs ~ s(c,8) inputs6,6, ... ,e& E lR n +q, and an arbitrary dichotomy ~ : {e1, 6, ... , e&} ~ {-I, I} find a representation of a hypothesis <Pc E Fn,q such that the restriction of <Pc to the set {e1,6, ... ,e&} is identical to the dichotomy ~ (or report that no such hypothesis exists). Sample Complexity for Learning Recurrent Perceptron Mappings 209 The representation to be used should provide an "efficient encoding" of the values of the parameters rl, • .. , rn , Cl , . . . , cn: given a set of inputs (Xl" ' " Xn+q) E jRn+q, one should be able to efficiently check concept membership (that is, compute sign (L:7~l CjXj)). Regarding the precise meaning of polynomial-time computation, there are at least two models of complexity possible: the unit cost model which deals with algebraic complexity (arithmetic and comparison operations take unit time) and the logarithmic cost model (computation in the Turing machine sense; inputs (Xl , . . . , X n+q ) are rationals, and the time involved in finding a representation of rl , . .. , r n , Cl, . .. , Cn is required to be polynomial on the number of bits L. Theorem 3 For each fixed n > 0, the consistency problem for :Fn,q can be solved in time polynomial in q and s in the unit cost model, and time polynomial in q, s, and L in the logarithmic cost model. Since vc (:Fn ,q) = O(n + nlog(q + 1)), it follows from here that the class :Fn,q is learnable in time polynomial in q (and L in the log model). Due to space limitations, we must omit the proof; it is based on the application of recent results regarding computational complexity aspects of the first-order theory of real-closed fields. 6 Pseudo-Dimension Bounds In this section, we obtain results on the learnability of linear systems dynamics, that is, the class of functions obtained if one does not take the sign when defining recurrent perceptrons. The connection between VC dimension and sample complexity is only meaningful for classes of Boolean functions; in order to obtain learnability results applicable to real-valued functions one needs metric entropy estimates for certain spaces of functions. These can be in turn bounded through the estimation of Pollard's pseudo-dimension. We next briefly sketch the general framework for learning due to Haussler (based on previous work by Vapnik, Chervonenkis, and Pollard) and then compute a pseudo-dimension estimate for the class of interest. The basic ingredients are two complete separable metric spaces X and If (called respectively the sets of inputs and outputs), a class :F of functions f : X -+ If (called the decision rule or hypothesis space), and a function f : If x If -+ [0, r] C jR (called the loss or cost function). The function f is so that the class of functions (x, y) ~ f(f(x), y) is "permissible" in the sense of Haussler and Pollard. Now, one may introduce, for each f E :F, the function AJ,l : X x If x jR -+ {-I, I} : (x, y, t) ~ sign (f(f(x) , y) - t) as well as the class A.1",i consisting of all such A/,i ' The pseudo-dimension of :F with respect to the loss function f, denoted by PO [:F, f], is defined as: PO [:F,R] := vc (A.1",i). Due to space limitations, the relationship between the pseudo-dimension and the sample complexity of the class :F will not be discussed here; the reader is referred to the references (Haussler, 1992; Maass, 1994) for details. For our application we define, for any two nonnegative integers n, q, the class :F~ , q := {¢<! ICE jRn+q is n-recursive} where ¢c jRn+q -+ jR: (Xl , .. . , Xn+q) ~ L:7~l CjXj . The following Theorem can be proved using Fact 4.1. Theorem 4 Let p be a positive integer and assume that the loss function f is given byf(Yl,Y2) = IYl- Y2IP • Then, PO [:F~ , q,f] ~ 18n+4nlog(p(q+ 1)) . 210 B. DASGUPTA, E. D. SONTAG Acknowledgements This research was supported in part by US Air Force Grant AFOSR-94-0293. References A.D. BACK AND A.C. TSOI, FIR and IIR synapses, a new neural network architecture for time-series modeling, Neural Computation, 3 (1991), pp. 375-385. A .D. BACK AND A .C. TSOI, A comparison of discrete-time operator models for nonlinear system identification, Advances in Neural Information Processing Systems (NIPS'94), Morgan Kaufmann Publishers, 1995, to appear. A.M . BAKSHO, S. DASGUPTA, J .S. GARNETT, AND C.R. JOHNSON, On the similarity of conditions for an open-eye channel and for signed filtered error adaptive filter stability, Proc. IEEE Conf. Decision and Control, Brighton, UK, Dec. 1991, IEEE Publications, 1991, pp. 1786-1787. A. BLUMER, A. EHRENFEUCHT, D. HAUSSLER, AND M . WARMUTH, Learnability and the Vapnik-Chervonenkis dimension, J. of the ACM, 36 (1989), pp. 929-965. D.F. DELCHAMPS, Extracting State Information from a Quantized Output Record, Systems and Control Letters, 13 (1989), pp. 365-372. R .O. DUDA AND P.E. HART, Pattern Classification and Scene Analysis, Wiley, New York, 1973. C.E. GILES, G.Z. SUN, H.H. CHEN, Y.C. LEE, AND D. CHEN, Higher order recurrent networks and grammatical inference, Advances in Neural Information Processing Systems 2, D.S. Touretzky, ed., Morgan Kaufmann, San Mateo, CA, 1990. P . GOLDBERG AND M. JERRUM, Bounding the Vapnik-Chervonenkis dimension of concept classes parameterized by real numbers, Mach Learning, 18, (1995): 131-148. D. HAUSSLER, Decision theoretic generalizations of the PAC model for neural nets and other learning applications, Information and Computation, 100, (1992): 78-150. R. KOPLON AND E.D. SONTAG, Linear systems with sign-observations, SIAM J. Control and Optimization, 31(1993): 1245 - 1266. W. MAASS, Perspectives of current research about the complexity of learning in neural nets, in Theoretical Advances in Neural Computation and Learning, V.P. Roychowdhury, K.Y. Siu, and A. Orlitsky, eds., Kluwer, Boston, 1994, pp. 295-336. G.W. PULFORD, R.A. KENNEDY, AND B.D.O. ANDERSON, Neural network structure for emulating decision feedback equalizers, Proc. Int. Conf. Acoustics, Speech, and Signal Processing, Toronto, Canada, May 1991, pp. 1517-1520. E.D. SONTAG, Neural networks for control, in Essays on Control: Perspectives in the Theory and its Applications (H.L. Trentelman and J .C. Willems, eds.), Birkhauser, Boston, 1993, pp. 339-380. GYORGY TURAN, Computational Learning Theory and Neural Networks:A Survey of Selected Topics, in Theoretical Advances in Neural Computation and Learning, V.P. Roychowdhury, K.Y. Siu,and A. Orlitsky, eds., Kluwer, Boston, 1994, pp. 243-293. L.G. VALIANT A theory of the learnable, Comm. ACM, 27, 1984, pp. 1134-1142. V.N .VAPNIK, Estimation of Dependencies Based on Empirical Data, Springer, Berlin, 1982.
|
1995
|
114
|
1,015
|
Is Learning The n-th Thing Any Easier Than Learning The First? Sebastian Thrun I Computer Science Department Carnegie Mellon University Pittsburgh, PA 15213-3891 World Wide Web: http://www.cs.cmu.edul'''thrun Abstract This paper investigates learning in a lifelong context. Lifelong learning addresses situations in which a learner faces a whole stream of learning tasks. Such scenarios provide the opportunity to transfer knowledge across multiple learning tasks, in order to generalize more accurately from less training data. In this paper, several different approaches to lifelong learning are described, and applied in an object recognition domain. It is shown that across the board, lifelong learning approaches generalize consistently more accurately from less training data, by their ability to transfer knowledge across learning tasks. 1 Introduction Supervised learning is concerned with approximating an unknown function based on examples. Virtually all current approaches to supervised learning assume that one is given a set of input-output examples, denoted by X, which characterize an unknown function, denoted by f. The target function f is drawn from a class of functions, F, and the learner is given a space of hypotheses, denoted by H, and an order (preference/prior) with which it considers them during learning. For example, H might be the space of functions represented by an artificial neural network with different weight vectors. While this formulation establishes a rigid framework for research in machine learning, it dismisses important aspects that are essential for human learning. Psychological studies have shown that humans often employ more than just the training data for generalization. They are often able to generalize correctly even from a single training example [2, 10]. One of the key aspects of the learning problem faced by humans, which differs from the vast majority of problems studied in the field of neural network learning, is the fact that humans encounter a whole stream of learning problems over their entire lifetime. When faced with a new thing to learn, humans can usually exploit an enormous amount of training data and I also affiliated with: Institut fur Informatik III, Universitat Bonn, Romerstr. 164, Germany Is Learning the n-th Thing Any Easier Than Learning the First? 641 experiences that stem from other, related learning tasks. For example, when learning to drive a car, years of learning experience with basic motor skills, typical traffic patterns, logical reasoning, language and much more precede and influence this learning task. The transfer of knowledge across learning tasks seems to play an essential role for generalizing accurately, particularly when training data is scarce. A framework for the study of the transfer of knowledge is the lifelong learning framework. In this framework, it is assumed that a learner faces a whole collection of learning problems over its entire lifetime. Such a scenario opens the opportunity for synergy. When facing its n-th learning task, a learner can re-use knowledge gathered in its previous n - 1 learning tasks to boost the generalization accuracy. In this paper we will be interested in the most simple version of the lifelong learning problem, in which the learner faces a family of concept learning tasks. More specifically, the functions to be learned over the lifetime of the learner, denoted by 11 , 12, 13, .. . E F , are all of the type I : I --+ {O, I} and sampled from F. Each function I E {II , h ,13, ... } is an indicator function that defines a particular concept: a pattern x E I is member of this concept if and only if I(x) = 1. When learning the n-th indicator function, In , the training set X contains examples of the type (x , In(x)) (which may be distorted by noise). In addition to the training set, the learner is also given n - 1 sets of examples of other concept functions, denoted by Xk (k = 1, .. . , n - I). Each Xk contains training examples that characterize Ik. Since this additional data is desired to support learning In, Xk is called a support set for the training set X . An example of the above is the recognition of faces [5, 7]. When learning to recognize the n-th person, say IBob, the learner is given a set of positive and negative example of face images of this person. In lifelong learning, it may also exploit training information stemming from other persons, such as I E {/Rieh, IMike , IDave , ... }. The support sets usually cannot be used directly as training patterns when learning a new concept, since they describe different concepts (hence have different class labels). However, certain features (like the shape of the eyes) are more important than others (like the facial expression, or the location of the face within the image). Once the invariances of the domain are learned, they can be transferred to new learning tasks (new people) and hence improve generalization. To illustrate the potential importance of related learning tasks in lifelong learning, this paper does not present just one particular approach to the transfer of knowledge. Instead, it describes several, all of which extend conventional memory-based or neural network algorithms. These approaches are compared with more traditional learning algorithms, i.e., those that do not transfer knowledge. The goal of this research is to demonstrate that, independent of a particular learning approach, more complex functions can be learned from less training data iflearning is embedded into a lifelong context. 2 Memory-Based Learning Approaches Memory-based algorithms memorize all training examples explicitly and interpolate them at query-time. We will first sketch two simple, well-known approaches to memory-based learning, then propose extensions that take the support sets into account. 2.1 Nearest Neighbor and Shepard's Method Probably the most widely used memory-based learning algorithm is J{ -nearest neighbor (KNN) [15]. Suppose x is a query pattern, for which we would like to know the output y. KNN searches the set of training examples X for those J{ examples (Xi, Yi) E X whose input patterns Xi are nearest to X (according to some distance metric, e.g., the Euclidian distance). It then returns the mean output value k 2:= Yi of these nearest neighbors. Another commonly used method, which is due to Shepard [13], averages the output values 642 s. THRUN of all training examples but weights each example according to the inverse distance to the query :~~~t x. ( ) ( I) -I L Ilx - ~: II + E· L Ilx - Xi II + E (x"y.)EX (x. ,y.)EX (1) Here E > 0 is a small constant that prevents division by zero. Plain memory-based learning uses exclusively the training set X for learning. There is no obvious way to incorporate the support sets, since they carry the wrong class labels. 2.2 Learning A New Representation The first modification of memory-based learning proposed in this paper employs the support sets to learn a new representation of the data. More specifically, the support sets are employed to learn a function, denoted by 9 : I --+ I', which maps input patterns in I to a new space, I' . This new space I' forms the input space for a memory-based algorithm. Obviously, the key property of a good data representations is that multiple examples of a single concept should have a similar representation, whereas the representation of an example and a counterexample of a concept should be more different. This property can directly be transformed into an energy function for g: n-I ( ) E:= ~ (X ,y~EXk (X"y~EXk Ilg(x)-g(x')11 (X"y~EXk Ilg(x)-g(x')11 (2) Adjusting 9 to minimize E forces the distance between pairs of examples of the same concept to be small, and the distance between an example and a counterexample of a concept to be large. In our implementation, 9 is realized by a neural network and trained using the Back-Propagation algorithm [12]. Notice that the new representation, g, is obtained through the support sets. Assuming that the learned representation is appropriate for new learning tasks, standard memory-based learning can be applied using this new representation when learning the n-th concept. 2.3 Learning A Distance Function An alternative way for exploiting support sets to improve memory-based learning is to learn a distance function [3, 9]. This approach learns a function d : I x I --+ [0, I] which accepts two input patterns, say x and x', and outputs whether x and x' are members of the same concept, regardless what the concept is. Training examples for d are ((x, x'),I) ify=y'=l ((x, x'), 0) if(y=IAy'=O)or(y=OAy'=I). They are derived from pairs of examples (x , y) , (x', y') E Xk taken from a single support set X k (k = 1, . .. , n I). In our implementation, d is an artificial neural network trained with Back-Propagation. Notice that the training examples for d lack information concerning the concept for which they were originally derived. Hence, all support sets can be used to train d. After training, d can be interpreted as the probability that two patterns x, x' E I are examples of the same concept. Once trained, d can be used as a generalized distance function for a memory-based approach. Suppose one is given a training set X and a query point x E I. Then, for each positive example (x' , y' = I) EX, d( x, x') can be interpreted as the probability that x is a member of the target concept. Votes from multiple positive examples (XI, I) , (X2' I), ... E X are combined using Bayes' rule, yielding Prob(fn(x)=I) .1- (I + II I:(~(::~,))-I (3) (x' ,y'=I)EXk Is Learning the n-th Thing Any Easier Than Learning the First? 643 Notice that d is not a distance metric. It generalizes the notion of a distance metric, because the triangle inequality needs not hold, and because an example of the target concept x' can provide evidence that x is not a member of that concept (if d(x, x') < 0.5). 3 Neural Network Approaches To make our comparison more complete, we will now briefly describe approaches that rely exclusively on artificial neural networks for learning In. 3.1 Back-Propagation Standard Back-Propagation can be used to learn the indicator function In, using X as training set. This approach does not employ the support sets, hence is unable to transfer knowledge across learning tasks. 3.2 Learning With Hints Learning with hints [1, 4, 6, 16] constructs a neural network with n output units, one for each function Ik (k = 1,2, .. . , n). This network is then trained to simultaneously minimize the error on both the support sets {Xk} and the training set X. By doing so, the internal representation of this network is not only determined by X but also shaped through the support sets {X k }. If similar internal representations are required for al1 functions Ik (k = 1,2, .. . , n), the support sets provide additional training examples for the internal representation. 3.3 Explanation-Based Neural Network Learning The last method described here uses the explanation-based neural network learning algorithm (EBNN), which was original1y proposed in the context of reinforcement learning [8, 17]. EBNN trains an artificial neural network, denoted by h : I ----+ [0, 1], just like Back-Propagation. However, in addition to the target values given by the training set X, EBNN estimates the slopes (tangents) of the target function In for each example in X. More specifically, training examples in EBNN are of the sort (x, In (x), \7 xl n (x)), which are fit using the Tangent-Prop algorithm [14]. The input x and target value In(x) are taken from the trai ning set X. The third term, the slope \7 xl n ( X ), is estimated using the learned distance function d described above. Suppose (x', y' = 1) E X is a (positive) training example. Then, the function dx ' : I ----+ [0, 1] with dx ' (z) := d(z , x') maps a single input pattern to [0, 1], and is an approximation to In. Since d( z, x') is represented by a neural network and neural networks are differentiable, the gradient 8dx ' (z) /8z is an estimate of the slope of In at z. Setting z := x yields the desired estimate of \7 xln (x) . As stated above, both the target value In (x) and the slope vector \7 x In (x) are fit using the Tangent-Prop algorithm for each training example x EX. The slope \7 xln provides additional information about the target function In. Since d is learned using the support sets, EBNN approach transfers knowledge from the support sets to the new learning task. EBNN relies on the assumption that d is accurate enough to yield helpful sensitivity information. However, since EBNN fits both training patterns (values) and slopes, misleading slopes can be overridden by training examples. See [17] for a more detailed description of EBNN and further references. 4 Experimental Results All approaches were tested using a database of color camera images of different objects (see Fig. 3.3). Each of the object in the database has a distinct color or size. The n-th 644 l1 .... 'I I't' 'I • < , > '" -.... . ~ 1:1 ,I , , c_. ML~._ ... , I '''!!!i!~, =' ;~~~ , ...... ~.. ' . ~ ~ <.:t " ~~- -_,1~ ~_ ~,-l/> ;' ;'j III ... '1 ~' ''',ll t! ~[~ .d!t~)ltI!{iH-"" :. ~~~ -~""":::.~ ~ -,:~~,} I , £ ».~ <.~ ,,, ~ -... ~ ~_l_~ __ E~ '~~ II _e·m;, ;1 ~ t ~,~,AA( , ~ " :R;1-; , ""111':'i, It f4~ r S. THRUN Figure 1: The support sets were compiled out of a hundred images of a bottle, a hat, a hammer, a coke can, and a book. The n-th learning tasks involves distinguishing the shoe from the sunglasses. Images were subsampled to a 100x 100 pixel matrix (each pixel has a color, saturation, and a brightness value), shown on the right side. learning task was the recognition of one of these objects, namely the shoe. The previous n 1 learning tasks correspond to the recognition of five other objects, namely the bottle, the hat, the hammer, the coke can, and the book. To ensure that the latter images could not be used simply as additional training data for In, the only counterexamples of the shoe was the seventh object, the sunglasses. Hence, the training set for In contained images of the shoe and the sunglasses, and the support sets contained images of the other five objects. The object recognition domain is a good testbed for the transfer of knowledge in lifelong learning. This is because finding a good approximation to In involves recognizing the target object invariant of rotation, translation, scaling in size, change of lighting and so on. Since these invariances are common to all object recognition tasks, images showing other objects can provide additional information and boost the generalization accuracy. Transfer of knowledge is most important when training data is scarce. Hence, in an initial experiment we tested all methods using a single image of the shoe and the sunglasses only. Those methods that are able to transfer knowledge were also provided 100 images of each of the other five objects. The results are intriguing. The generalization accuracies KNN Shepard repro g+Shep. distanced Back-Prop hints EBNN 60.4% 60.4% 74.4% 75.2% 59.7% 62.1% 74.8% ±8.3% ±8.3% ±18.5% ±18.9% ±9.0% ±10.2% ±11.1% illustrate that all approaches that transfer knowledge (printed in bold font) generalize significantly better than those that do not. With the exception of the hint learning technique, the approaches can be grouped into two categories: Those which generalize approximately 60% of the testing set correctly, and those which achieve approximately 75% generalization accuracy. The former group contains the standard supervised learning algorithms, and the latter contains the "new" algorithms proposed here, which are capable of transferring knOWledge. The differences within each group are statistically not significant, while the differences between them are (at the 95% level). Notice that random guessing classifies 50% of the testing examples correctly. These results suggest that the generalization accuracy merely depends on the particular choice of the learning algorithm (memory-based vs. neural networks). Instead, the main factor determining the generalization accuracy is the fact whether or not knowledge is transferred from past learning tasks. Is Learning the n-th Thing Any Easier Than Learning the First? 645 95% 85% ~ 80% ~ 15% 70% 65% 60% distance function d hepard 's method with representation g Shepard's method 55% 50%~2--~~----~10~~1~2~1~4--1~6--1~.--~20 training example. 95% , ,,'~ . . </'~ 70% if . 65% /f Back-Propagauon 60% ;;./ 55% ~%~2--~~----~1~O~1~2--1~4--1~6--~1B--~20 training exampletl Figure 2: Generalization accuracy as a function of training examples, measured on an independent test set and averaged over 100 experiments. 95%-confidence bars are also displayed. What happens as more training data arrives? Fig. 2 shows generalization curves with increasing numbers of training examples for some of these methods. As the number of training examples increases, prior knowledge becomes less important. After presenting 20 training examples, the results KNN Shepard repro g+Shep. distance d Back-Prop hints EBNN 81.0% 70.5% 81.7% 87.3% 88.4% n_avail. 90.8% ±3.4% ±4.9% ±2.7% ±O_9% ±2.5% ±2.7% illustrate that some of the standard methods (especially Back-Propagation) generalize about as accurately as those methods that exploit support sets. Here the differences in the underlying learning mechanisms becomes more dominant. However, when comparing lifelong learning methods with their corresponding standard approaches, the latter ones are stiIl inferior: BackPropagation (88.4%) is outperformed by EBNN (90.8%), and Shepard's method (70.5%) generalizes less accurately when the representation is learned (81.7%) or when the distance function is learned (87.3%). All these differences are significant at the 95% confidence level. 5 Discussion The experimental results reported in this paper provide evidence that learning becomes easier when embedded in a lifelong learning context. By transferring knowledge across related learning tasks, a learner can become "more experienced" and generalize better. To test this conjecture in a more systematic way, a variety of learning approaches were evaluated and compared with methods that are unable to transfer knowledge. It is consistently found that lifelong learning algorithms generalize significantly more accurately, particularly when training data is scarce. Notice that these results are well in tune with other results obtained by the author. One of the approaches here, EBNN, has extensively been studied in the context of robot perception [11], reinforcement learning for robot control, and chess [17]. In all these domains, it has consistently been found to generalize better from less training data by transferring knowledge from previous learning tasks. The results are also consistent with observations made about human learning [2, 10], namely that previously learned knowledge plays an important role in generalization, particularly when training data is scarce. [18] extends these techniques to situations where most support sets are not related.w However, lifelong learning rests on the assumption that more than a single task is to be learned, and that learning tasks are appropriately related. Lifelong learning algorithms are particularly well-suited in domains where the costs of collecting training data is the dominating factor in learning, since these costs can be amortized over several learning tasks. Such domains include, for example, autonomous service robots which are to learn and improve over their entire lifetime. They include personal software assistants which have 646 S. THRUN to perform various tasks for various users. Pattern recognition, speech recognition, time series prediction, and database mining might be other, potential application domains for the techniques presented here. References [1] Y. S. Abu-Mostafa. Learning from hints in neural networks. Journal of Complexity, 6: 192-198, 1990. [2] W-K. Ahn and W F. Brewer. Psychological studies of explanation-based learning. In G. Dejong, editor, Investigating Explanation-Based Learning. Kluwer Academic Publishers, BostonlDordrechtILondon, 1993. [3] c. A. Atkeson. Using locally weighted regression for robot learning. In Proceedings of the 1991 1EEE International Conference on Robotics and Automation, pages 958-962, Sacramento, CA, April 1991. [4] J. Baxter. Learning internal representations. In Proceedings of the Conference on Computation Learning Theory, 1995. [5] D. Beymer and T. Poggio. Face recognition from one model view. In Proceedings of the International Conference on Computer Vision, 1995. [6] R. Caruana. MuItitask learning: A knowledge-based of source of inductive bias. In P. E. Utgoff, editor, Proceedings of the Tenth International Conference on Machine Learning, pages 41-48, San Mateo, CA, 1993. Morgan Kaufmann. [7] M. Lando and S. Edelman. Generalizing from a single view in face recognition. Technical Report CS-TR 95-02, Department of Applied Mathematics and Computer Science, The Weizmann Institute of Science, Rehovot 76100, Israel, January 1995. [8] T. M. Mitchell and S. Thrun. Explanation-based neural network learning for robot control. In S. J. Hanson, J. Cowan, and C. L. Giles, editors, Advances in Neural Information Processing Systems 5, pages 287-294, San Mateo, CA, 1993. Morgan Kaufmann. [9] A. W Moore, D. 1. Hill, and M. P. Johnson. An Empirical Investigation of Brute Force to choose Features, Smoothers and Function Approximators. In S. Hanson, S. Judd, and T. Petsche, editors, Computational Learning Theory and Natural Learning Systems, Volume 3. MIT Press, 1992. [10] Y. Moses, S. Ullman, and S. Edelman. Generalization across changes in illumination and viewing position in upright and inverted faces. Technical Report CS-TR 93-14, Department of Applied Mathematics and Computer Science, The Weizmann Institute of Science, Rehovot 76100, Israel, 1993. [11] J. O'Sullivan, T. M. Mitchell, and S. Thrun. Explanation-based neural network learning from mobile robot perception. In K. Ikeuchi and M. Veloso, editors, Symbolic Visual Learning. Oxford University Press, 1995. [12] D. E. Rumelhart, G. E. Hinton, and R. J. Williams. Learning internal representations by error propagation. In D. E. Rumelhart and 1. L. McClelland, editors, Parallel Distributed Processing. Vol. I + II. MIT Press, 1986. [13] D. Shepard. A two-dimensional interpolation function for irregularly spaced data. In 23rd National Conference ACM, pages 517-523, 1968. [14] P. Simard, B. Victorri, Y. LeCun, and J. Denker. Tangent prop - a formalism for specifying selected invariances in an adaptive network. In 1. E. Moody, S. J. Hanson, and R. P. Lippmann, editors, Advances in Neural Information Processing Systems 4, pages 895-903, San Mateo, CA, 1992. Morgan Kaufmann. [15] c. Stanfill and D. Waltz. Towards memory-based reasoning. Communications of the ACM, 29(12): 1213-1228, December 1986. [16] S. C. Suddarth and A. Holden. Symbolic neural systems and the use of hints for developing complex systems. International Journal of Machine Studies, 35, 1991. [17] S. Thrun. Explanation-Based Neural Network Learning: A Lifelong Learning Approach. Kluwer Academic Publishers, Boston, MA, 1996. to appear. [18] S. Thrun and J. O'Sullivan. Clustering learning tasks and the selective cross-task transfer of knowledge. Technical Report CMU-CS-95-209, Carnegie Mellon University, School of Computer Science, Pittsburgh, PA 15213, November 1995.
|
1995
|
115
|
1,016
|
Parallel analog VLSI architectures for computation of heading direction and time-to-contact Giacomo Indiveri giacomo@klab.caltech.edu Jorg Kramer kramer@klab.caltech.edu Division of Biology California Institute of Technology Pasadena, CA 91125 Abstract Christof Koch koch@klab.caltech.edu We describe two parallel analog VLSI architectures that integrate optical flow data obtained from arrays of elementary velocity sensors to estimate heading direction and time-to-contact. For heading direction computation, we performed simulations to evaluate the most important qualitative properties of the optical flow field and determine the best functional operators for the implementation of the architecture. For time-to-contact we exploited the divergence theorem to integrate data from all velocity sensors present in the architecture and average out possible errors. 1 Introduction We have designed analog VLSI velocity sensors invariant to absolute illuminance and stimulus contrast over large ranges that are able to achieve satisfactory performance in a wide variety of cases; yet such sensors, due to the intrinsic nature of analog processing, lack a high degree of precision in their output values. To exploit their properties at a system level, we developed parallel image processing architectures for applications that rely mostly on the qualitative properties of the optical flow, rather than on the precise values of the velocity vectors. Specifically, we designed two parallel architectures that employ arrays of elementary motion sensors for the computation of heading direction and time-to-contact. The application domain that we took into consideration for the implementation of such architectures, is the promising one of vehicle navigation. Having defined the types of images to be analyzed and the types of processing to perform, we were able to use a priori inforVLSI Architectures for Computation of Heading Direction and Time-to-contact 721 mation to integrate selectively the sparse data obtained from the velocity sensors and determine the qualitative properties of the optical flow field of interest. 2 The elementary velocity sensors A velocity sensing element, that can be integrated into relatively dense arrays to estimate in parallel optical flow fields, has been succesfully built [Kramer et al., 1995]. Unlike most previous implementations of analog VLSI motion sensors, it unambiguously encodes 1-D velocity over considerable velocity, contrast, and illuminance ranges, while being reasonably compact. It implements an algorithm that measure'3 the time of travel of features (here a rapid temporal change in intensity) stimulus between two fixed locations on the chip. In a first stage, rapid dark-to-bright irradiance changes or temporal ON edges are converted into short current pulses. Each current pulse then gives rise to a sharp voltage spike and a logarithmically-decaying voltage signal at each edge detector location. The sharp spike from one location is used to sample the analog voltage of the slowly-decaying signal from an adjacent location. The sampled output voltage encodes the relative time delay of the two signals, and therefore velocity, for the direction of motion where the onset of the slowly-decaying pulse precedes the sampling spike. In the other direction, a lower voltage is sampled. Each direction thus requires a separate output stage. '05 ~ OIlS j " > o. Oil!> O' 40 .. VIIIoaIy (mmsec) Figure 1: Output voltage of a motion sensing element for the preferred direction of motion of a sharp high-contrast ON edge versus image velocity under incandescent room illumination. Each data point represents the average of 5 successive measurements. As implemented with a 2 J.tm CMOS process, the size of an elementary bi-directional motion element (including 30 transistors and 8 capacitances) is 0.045 mm2. Fig. 1 shows that the experimental data confirms the predicted logarithmic encoding of velocity by the analog output voltage. The data was taken by imaging a moving high-contrast ON edge onto the chip under incandescent room illumination. The calibration of the image velocity in the focal plane is set by the 300 J.tm spacing of adjacent photoreceptors on the chip. 3 Heading direction computation To simplify the computational complexity of the problem of heading direction detection we restricted our analysis to pure translational motion, taking advantage of the 722 G. INDIVERI, 1. KRAMER, C. KOCH fact that for vehicle navigation it is possible to eliminate the rotational component of motion using lateral accelerometer measurements from the vehicle. Furthermore, to analyze the computational properties of the optical flow for typical vehicle navigation scenes, we performed software simulations on sequences of images obtained from a camera with a 64 x 64 pixel silicon retina placed on a moving truck (courtesy of B. Mathur at Rockwell Corporation). The optical flow fields have been computed Figure 2: The sum of the horizontal components of the optical flow field is plotted on the bottom of the figure. The presence of more than one zero-crossing is due to different types of noise in the optical flow computation (e.g. quantization errors in software simulations or device mismatch in analog VLSI circuits). The coordinate of the heading direction is computed as the abscissa of the zero-crossing with maximum steepness and closest to the abscissa of the previously selected unit. by implementing an algorithm based on the image brightness consia'f)cy equation [Verri et al., 1992] [Barron et al., 1994]. For the application domain considered and the types of optical flow fields obtained from the simulations, it is reasonable to assume that the direction of heading changes smoothly in time. Furthermore, being interested in determining, and possibly controlling, the heading direction mainly along the horizontal axis, we can greatly reduce the complexity of the problem by considering one-dimensional arrays of velocity sensors. In such a case, if we assign positive values to vectors pointing in one direction and negative values to vectors pointing in the opposite direction, the heading direction location will correspond to the point closest to the zero-crossing. Under these assumptions, the computation of the horizontal coordInate of the heading direction has been carried out using the following functional operators: thresholding on the horizontal components of the optical flow vectors; spatial smoothing on the resulting values; detection and evaluation of the steepness of the zero-crossings present in the array and finally selection of the zero-crossing with maximum steepness. The zero-crossing with maximum steepness is selected only if its position is in the neighborhood of the previously selected zero-crossing. This helps to eliminate errors due to noise and device mismatch and assures that the computed heading direction location will shift smoothly in time. Fig. 2 shows a result of the software simulations, on an image of a road VLSI Architectures for Computation of Heading Direction and Time-to-contact 723 with a shadow on the left side. All of the operators used in the algorithm have been implemented with analog circuits (see Fig. 3 for a block diagram of the architecture). Specifically, we have WIIb .. Figure 3: Block diagram of the architecture for detecting heading direction: the first layer of the architecture computes the velocity of the stimulus; the second layer converts the voltage output of the velocity sensors into a positive/negative current performing a threshold operation; the third layer performs a linear smoothing operation on the positive and negative halfs of the input current; the fourth layer detects zero-crossings by comparing the intensity of positive currents from one pixel with negative currents from the neighboring pixel; the top layer implements a winner-take-all network with distributed excitation, which selects the zero-crossing with maximum steepness. designed test chips in which the thresholding function has been implemented using a transconductance amplifier whose current represents the output signal [Mead, 1989], spatial smoothing has been obtained using a circuit that separates positive currents and negative currents into two distinct paths and feeds them into two layers of current-mode diffuser networks [Boahen and Andreou, 1992], the zerocrossing detection and evaluation of its steepness has been implemented using a newly designed circuit block based on a modification of the simple current-correlator [Delbriick, 1991], and the selection of the zero-crossing with maximum steepness closest to the previously selected unit has been implemented using a winner-takeall circuit with distributed excitation [Morris et al., 1995]. The schematics of the former three circuits, which implement the top three layers of the diagram of Fig. 3, are shown in Fig. 4. Fig. 5 shows the output of a test chip in which all blocks up to the diffuser network (without the zero-crossing detection stages) were implemented. The velocity sensor layout was modified to maximize the number of units in the 1-D array. Each velocity sensor measures 60pm x 802pm. On a (2.2mm)2 size chip we were able to fit 23 units. The shown results have been obtained by imaging on the chip expanding or contracting stimuli using black and white edges wrapped around a rotating drum and reflected by an adjacent tilted mirror. The point of contact between drum and mirror corresponding to the simulated heading direction has been imaged approximately onto the 15th unit of the array. As shown, the test chip considered does not achieve 100% correct performance due to errors that arise mainly from the presence of parasitic capacitors in the modified part of the velocity sensor circuits; nonetheless, at least from a qualitative point of view, the data confirms the results obtained from software simulations and demonstrates the validity of the approach considered. 724 O. INDNERI. J. KRAMER. C. KOCH - - - - - ~- - - - - -Ir - - - - 1 ~ II rln II ~ I ~ '" ,.. II ~ Cd •• II cp I II II I II II ""'>---t~;-------'~:-----<-' II I.. II *"J>--........,'t--t--t-<.,fJ "'.>----t ....... -----i'-----<,..t II ill !P II II II II II II II .. _"II II I , .. , .. II I I ------------------- ______ 1 ______ ---Figure 4: Circuit schematics of the smoothing, zero-detection and winner-take-all blocks respectively. " .................. -r-~...,..... ........ ......-........ -.-................. -r-.,.......,.--.-, " .B .0, .0. 'O·.~~'~~~~7~.~.~'.~'~,~I2~"~'~.~'.~"~'~"~.~,,~~~'~ ' 22 lk1I P9ton (a) .0' "!-, ~ , ~,~,~. ~'~'''''7-' ~,~,."""~' ~,,~,,~,,~,"". ~,.""",,:-',~, .,.." ~'O~"~22' """""'''''' (b) Figure 5: Zero crossings computed as difference between smoothed positive currents and smoothed negative currents: (a) for expanding stimuli; (b) for contracting stimuli. The "zero" axis is shifted due to is a systematic offset of 80 nA. 4 Time-to-contact The time-to-contact can be computed by exploiting qualitative properties of the optical flow field such as expansion or contraction [Poggio et al., 1991]. The divergence theorem, or Gauss theorem, as applied to a plane, shows that the integral over a surface patch of the divergence of a vector field is equal to the line integral along the patch boundary of the component of the field normal to the boundary. Since a camera approaching a rigid object sees a linear velocity field, where the velocity vectors are proportional to their distance from the focus-of-expansion, the divergence is constant over the image plane. By integrating the radial component of the optical flow field along the circumference of a circle, the time-to-contact can thus be estimated, independently of the position of the focus-of-expansion. We implemented this algorithm with an analog integrated circuit, where an array of twelve motion sensing elements is arranged on a circle, such that each element measures velocity radially. According to the Gauss theorem, the time-to-contact is VLSI Architectures for Computation of Heading Direction and Time-to-contact 725 then approximated by N·R T= N ' 2:k=l Vk (1) where N denotes the number of elements, R the radius of the circle, and Vk the radial velocity components at the locations of the elements. For each clement, temporal aliasing is prevented by comparing the output voltages of the two directions of motion and setting the lower one, corresponding to the null direction, to zero. The output voltages are then used to control subthreshold transistor currents. Since these voltages are logarithmically dependent on velocity, the transistor currents are proportional to the measured velocities. The sum of the velocity components is thus calculated by aggregating the currents from all elements on two lines, one for outward motion and one for inward motion, and taking the difference of the total currents. The resulting bi-directional output current is an inverse function of the signed time-to-contact. ;: 1 o~--------~----~~ I -, -0 25 OOS 01 015 02 025 Time-1c>ContKt (sec) Figure 6: Output current of the time-to-contact sensor as a function of simulated time-to-contact under incandescent room illumination. The theoretical fit predicts an inverse relationship. The circuit has been implemented on a chip with a size of (2.2mm)2 using 2 pm technology. The photo diodes of the motion sensing elements are arranged on two concentric circles with radii of 400 pm and 600 pm respectively. In order to simulate an approaching or withdrawing object, a high-contrast spiral stimulus was printed onto a rotating disk. Its image was projected onto the chip with a microscope lens under incandescent room illumination. The focus-of-expansion was approximately centered with respect to the photo diode circles. The averaged output current is shown as a function of simulated time-to-contact with a theoretical fit in Fig. 6. The expected inverse relationship is qualitatively observed and the sign (expansion or contraction) is robustly encoded. However, the deviation of the output current from its average can be substantial: Since the output voltage of each motion sensing element decays slowly due to leak currents and since the spiral stimulus causes a serial update of the velocity values along the array, a step change in the output current is observed upon each update, followed by a slow decay. The effect is aggravated, if the individual motion sensing elements measure significantly differing velocities. This is generally the case, because the focus-of-expansion is usually not centered with respect to the sensor and because of inaccuracies in the velocity measurements due to circuit offsets, noise, and the aperture problem [Verri et al., 1992]. The integrative property of the algorithm is thus highly desirable, and more robust data would be obtained from an array with more elements and stimuli with higher edge densities. 726 G. INDIVERI. J. KRAMER. C. KOCH 5 Conclusions We have developed parallel architectures for motion analysis that bypass the problem of low precision in analog VLSI technology by exploiting qualitative properties of the optical flow. The correct functionality of the devices built, at least from a qualitative point of view, have confirmed the validity of the approach followed and induced us to continue this line of research. We are now in the process of designing more accurate circuits that implement the operators used in the architectures proposed. Acknowledgments This work was supported by grants from ONR, ERe and Daimler-Benz AG. The velocity sensor was developed in collaboration with R. Sarpeshkar. The chips were fabricated through the MOSIS VLSI Fabrication Service. References [Barron et al., 1994] J.1. Barron, D.J. Fleet, and S.S. Beauchemin. Performance of optical flow techniques. International Journal on Computer Vision, 12(1):43-77, 1994. [Boahen and Andreou, 1992] K.A. Boahen and A.G. Andreou. A contrast sensitive silicon retina with reciprocal synapses. In NIPS91 Proceedings. IEEE, 1992. [Delbriick, 1991] T. Delbriick. "Bump" circuits for computing similarity and dissimilarity of analog voltages. In Proc. IJCNN, pages 1-475-479, June 1991. [Kramer et al., 1995] J. Kramer, R. Sarpeshkar, and C. Koch. An analog VLSI velocity sensor. In Proc. Int. Symp. Circuit and Systems ISCAS '95, pages 413-416, Seattle, WA, May 1995. [Mead, 1989] C.A. Mead. Analog VLSI and Neural Systems. Addison-Wesley, Reading, 1989. [Morris et al., 1995] T .G. Morris, D.M. Wilson, and S.P. DeWeerth. Analog VLSI circuits for manufacturing inspection. In Conference for Advanced Research in VLSI-Chapel Hill, North Carolina, March 1995. [Poggio et al., 1991] T. Poggio, A. Verri, and V. Torre. Green theorems and qualitative properties of the optical flow. Technical report, MIT, 1991. Internal Lab. Memo 1289. [Verri et al., 1992] A. Verri, M. Straforini, and V. Torre. Computational aspects of motion perception in natural and artificial vision systems. Phil. Trans. R. Soc. Lond. B, 337:429-443, 1992. PART VI SPEECH AND SIGNAL PROCESSING
|
1995
|
116
|
1,017
|
A Dynamical Systems Approach for a Learnable Autonomous Robot J un Tani and N aohiro Fukumura Sony Computer Science Laboratory Inc. Takanawa Muse Building, 3-14-13 Higashi-gotanda, Shinagawa-ku,Tokyo, 141 JAPAN Abstract This paper discusses how a robot can learn goal-directed navigation tasks using local sensory inputs. The emphasis is that such learning tasks could be formulated as an embedding problem of dynamical systems: desired trajectories in a task space should be embedded into an adequate sensory-based internal state space so that an unique mapping from the internal state space to the motor command could be established. The paper shows that a recurrent neural network suffices in self-organizing such an adequate internal state space from the temporal sensory input. In our experiments, using a real robot with a laser range sensor, the robot navigated robustly by achieving dynamical coherence with the environment. It was also shown that such coherence becomes structurally stable as the global attractor is self-organized in the coupling of the internal and the environmental dynamics. 1 Introd uction Conventionally, robot navigation problems have been formulated assuming a global view of the world. Given a detailed map of the workspace, described in a global coordinate system, the robot navigates to the specified goal by following this map. However, in situations where robots have to acquire navigational knowledge based on their own behaviors, it is important to describe the problems from the internal views of the robots. [Kuipers 87], [Mataric 92] and others have developed an approach based on landmark detection. The robot acquires a graph representation of landmark types as a topological modeling of the environment through its exploratory travels using the local sensory inputs. In navigation, the robot can identify its topological position by anticipating the landmark types in the graph representation obtained. It is, however, considered that this navigation strategy might be susceptible to erroneous landmark-matching. If the robot is once lost by such a catastrophe, its recoverance of the positioning might be difficult. We need certain mechanisms by which the 990 J. TANI, N. FUKUMURA robot can recover autonomously from such failures. We study the above problems by using the dynamical systems approach, expecting that this approach would provide an effective representational and computational framework. The approach focuses on the fundamental dynamical structure that arises from coupling the internal and the environmental dynamics [Beer 95]. Here, the objective of learning is to adapt the internal dynamical function such that the resultant dynamical structure might generate the desired system behavior. The system's performance becomes structurally stable if the dynamical structure maintains a sufficiently large basin of attraction against possible perturbations. We verify our claims through the implementation of our scheme on YAMABICO mobile robot equipped with a laser range sensor. The robot conducts navigational tasks under the following assumptions and conditions. (1) The robot cannot access its global position, but it navigates depending on its local sensory (range image) input. (2) There is no explicit landmarks accessible to the robot in the adopted workspace. (3) The robot learns tasks of cyclic routing by following guidance of a trainer. (4) The navigation should be robust enough against possible noise in the environment. 2 NAVIGATION ARCHITECTURE The YAMABICO mobile robot [Yuta and Iijima 90] was used as an experimental platform. The robot can obtain range images by a range finder consisting of laser projectors and three CCD cameras. The ranges for 24 directions, covering a 160 degree arc in front of the robot, are measured every 150 milliseconds. In our formulation, maneuvering commands are generated as the output of a composite system consisting of two levels [Tani and Fukumura 94]. The control level generates a collision-free, smooth trajectory using the range image, while the navigation level directs the control level in a macroscopic sense, responding to the sequential branching that appears in the sensory flows. The control level is fixed; the navigation level, on the other hand, can be adapted through learning. Firstly, let us describe the control level. The robot can sense the forward range readings of the surrounding environment, given in robot-centered polar coordinates by ri (1 ~ i ~ N). The angular range profile Ri is obtained by smoothing the original range readings through applying an appropriate Gaussian filter. The maneuvering focus of the robot is the maximum (the angular direction of the largest range) in this range profile. The robot proceeds towards the maximum of the profile (an open space in the environment). The navigation level focuses on the topological changes in the range profile as the robot moves. As the robot moves through a given workspace, the profile gradually changes until another local peak appears when the robot reaches a branching point. At this moment of branching the navigation level decides whether to transfer the focus to the new local peak or to remain with the current one. It is noted that this branching could be quite undeterministic one if applied to rugged obstacle environment. The robot is likely to fail to detect branching points frequently in such environment. The navigation level determines the branching by utilizing the range image obtained at branch points. Since the pertinent information in the range profile at a given moment is assumed to be only a small fraction of the total, we employ a vector quantization technique, known as the Kohonen network [Kohonen 82]' so that the information in the profile may be compressed into specific lower-dimensional data. The Kohonen network employed here consists of an I-dimensional lattice with m nodes along each dimension (l=3 and m=6 for the experiments with YAMABICO). The range image consisting of 24 values is input to the lattice, then the most A Dynamical Systems Approach for a Learnable Autonomous Robot Pn : sensory inputs TPM's output space • of (6.6.6) • , , , , , , , , F\,: range profile cn: context units Figure 1: Neural architecture for skill-based learning. 991 highly activated unit in the lattice, the "winner" unit, is found. The address of the winner unit in the lattice denotes the output vector of the network. Therefore, the navigation level receives the sensory input compressed into three dimensional data. The next section will describe how the robot can generate right branching sequences upon receiving the compressed range image. 3 Formulation 3.1 Learning state-action map The neural adaptation schemes are applied to the navigation level so that it can generate an adequate state-action map for a given task. Although some might consider that such map can be represented by using a layered feed-forward network with the inputs of the sensory image and the outputs of the motor command, this is not always true. The local sensory input does not always correspond uniquely to the true state of the robot (the sensory inputs could be the same for different robot positions). Therefore, there exists an ambiguity in determining the motor command solely from sensory inputs. This is a typical example of so-called non-Markovian problems which have been discussed by Lin and Mitchell [Lin and Mitchell 92]. In order to solve this ambiguity, a representation of contexts which are memories of past sensory sequences is required. For this purpose, a recurrent neural network (RNN) [Elman 90] was employed since its recurrent context states could represent the memory of past sequences. The employed neural architecture is shown in Figure. 1. The sensory input Pn and the context units en determine the appropriate motor command Xn+l' The motor command Xn takes a binary value of 0 (staying at the current branch) or 1 (a transit to a new branch). The RNN learning of sensorymotor (Pn,xn+d sequences, sampled through the supervised training, can build the desired state-action map by self-organizing adequate internal representation in time. 992 J. TANI, N. FUKUMURA (a) task space internal state space (b) task space internal state space Figure 2: The desired trajectories in the task space and its mapping to the internal state space. 3.2 Embedding problem The objective of the neural learning is to embed a task into certain global attractor dynamics which are generated from the coupling of the internal neural function and the environment. Figure 2 illustrates this idea. We define the internal state of the robot by the state of the RNN. The internal dynamics, which are coupled with the environmental dynamics through the sensory-motor loop, evolve as the robot travels in the task space. We assume that the desired vector field in the task space forms a global attractor, such as a fixed point for a homing task or limit cycling for a cyclic routing task. All that the robot has to do is to follow this vector flow by means of its internal state-action map. This requires a condition: the vector field in the internal state space should be self-organized as being topologically equivalent to that in the task space in order that the internal state determine the action (motor command) uniquely. This is the embedding problem from the task space to the internal state space, and RNN learning can attain this, using various training trajectories. This analysis conjectured that the trajectories in the task space can always converge into the desired one as long as the task is embedded into the global attractor in the internal state space. 4 Experiment 4.1 Task and training procedure Figure 3 shows an example of the navigation task, (which is adopted for the physical experiment in a later section). The task is for the robot to repeat looping of a figure of '8' and '0' in sequence. The task is not trivial because at the branching position A the robot has to decide whether to go '8' or '0' depending on its memory of the last sequence. The robot learns this navigation task through supervision by a trainer. The trainer repeatedly guides the robot to the desired loop from a set of arbitrarily selected A Dynamical Systems Approach for a Learnable Autonomous Robot 993 CJ Figure 3: Cyclic routing task, in which YAMABICO has to trace a figure of eight followed by a single loop. Figure 4: Trace of test travels for cyclic routing. initial locations. (The training was conducted with starting the robot from 10 arbitrarily selected initial locations in the workspace.) In actual training, the robot moves by the navigation of the control level and stops at each branching point, where the branching direction is taught by the trainer. The sequence of range images and teaching branching commands at those bifurcation points are fed into the neural architecture as training data. The objective of training RNN is to find the optimal weight matrix that minimizes the mean square error of the training output (branching decision) sequences associating with sensory inputs (outputs of Kohonen network). The weight matrix can be obtained through an iterative calculation of back-propagation through time (BPTT) [Rumelhart et al. 86]. 4.2 Results After the training, we examined how the robot achieves the trained task. The robot was started from arbitrary initial positions for this test. Fig. 4 shows example test travels. The result showed that the robot always converged to the desired loop regardless of its starting position. The time required to converge, however, took a 994 J. TANI. N. FUKUMURA AA Jt.JLGlLrr1L .dLA[L[L O:t....an..lIIlITIl. nlLllil..lliLrriL lliLalLlliLlliL lllllLllll11 Q. oIL .fiL .fiL .ilL .. (A') 'iii '" lLlLlLlL c :2 0 rn:L 0....... rn:L 0....... c ~ .Q dlLdILdlLdIL dLdLdLdL rrllrrllrrllrrll J1..J1..JlJ1.. .ill. .ill. JlI. JlI. (A) [hJ[hJ[hJ[hJ ~uJrr.J~ • cycle Figure 5: The sequence of activations in input and context units during the cycling travel. certain period that depended on the case. The RNN initially could not function correctly because of the arbitrary initial setting of the context units. However, while the robot wandered around the workspace, the RNN became situated (recovered the context) as it encountered pre-learned sensory sequences. Thereafter, its navigation converged to the cycling loop. Even after convergence, the robot could, by chance, leave the loop, under the influence of noise. However, the robot always came back to the loop after a while. These observations indicate that the robot learned the objective navigational task as embedded in a global attractor of limit cycling. It is interesting to examine how the task is encoded in the internal dynamics of the RNN. We investigated the activation patterns of RNN after its convergence into the loop. The results are shown in Fig. 5. The input and context units at each branching point are shown as three white and two black bars, respectively. One cycle (the completion of two routes of '0' and '8') are aligned vertically as one column. The figure shows those of four continuous cycles. It can be seen that robot navigation is exposed to much noise; the sensing input vector becomes unstable at particular locations, and the number of branchings in one cycle is not constant (i.e. some branching points are undeterministic). The rows labeled as (A) and (A') are branches to the routes of '0' and '8', respectively. In this point, the sensory input receives noisy chattering of different patterns independent of (A) or (A'). The context units, on the other hand, is completely identifiable between (A) and (A'), which shows that the task sequence between two routes (a single loop and an eight) is rigidly encoded internally, even in a noisy environment. In further experiments in more rugged obstacle environments, we found that this sort of structural stability could not be always assured. When the undeterministicity in the branching exceeds a certain limit, the desired dynamical structure cannot be preserved. A Dynamical Systems Approach for a Learnable Autonomous Robot 995 5 Summary and Discussion The navigation learning problem was formulated from the dynamical systems perspective. Our experimental results showed that the robot can learn the goal-directed navigation by embedding the desired task trajectories in the internal state space through the RNN training. It was also shown that the robot achieves the navigational tasks in terms of convergence of attract or dynamics which emerge in the coupling of the internal and the environmental dynamics. Since the dynamical coherence arisen in this coupling leads to the robust navigation of the robot, the intrinsic mechanism presented here is characterized by the term "autonomy". Finally, it is interesting to study how robots can obtain analogical models of the environment rather than state-action maps for adapting to flexibly changed goals. We discuss such formulation based on the dynamical systems approach elsewhere [Tani 96]. References [Beer 95] R.D. Beer. A dynamical systems perspective on agent-environment interaction. Artificial Intelligence, Vol. 72, No.1, pp.173- 215, 1995. [Elman 90] J.L. Elman. Finding structure in time. Cognitive Science, Vol. 14, pp.179- 211, 1990. [Kohonen 82] T. Kohonen. Self-Organized Formation of Topographically Correct Feature Maps. Biological Cybernetics, Vol. 43, pp.59- 69, 1982. [Kuipers 87] B. Kuipers. A Qualitative Approach to Robot Exploration and Map Learning. In AAAI Workshop Spatial Reasoning and Multi-Sensor Fusion {Chicago),1987. [Lin and Mitchell 92] L.-J. Lin and T.M. Mitchell. Reinforcement learning with hidden states. In Proc. of the Second Int. Conf. on Simulation of Adaptive Behavior, pp. 271- 280, 1992. [Mataric 92] M. Mataric. Integration of Representation into Goal-driven Behaviorbased Robot. IEEE Trans. Robotics and Automation, Vol. 8, pp.304- 312, 1992. [Rumelhart et al. 86] D.E. Rumelhart, G.E. Hinton, and R.J. Williams. Learning Internal Representations by Error Propagation. In Parallel Distributed Processing. MIT Press, 1986. [Tani 96] J. Tani. Model-Based Learning for Mobile Robot Navigation from the Dynamical Systems Perspective. IEEE Trans. System, Man and Cybernetics Part B, Special issue on robot learning, Vol. 26, No.3, 1996. [Tani and Fukumura 94] J. Tani and N. Fukumura. Learning goal-directed sensorybased navigation of a mobile robot. Neural Networks, Vol. 7, No.3, pp.553- 563, 1994. [Yuta and Iijima 90] S. Yuta and J. Iijima. State Information Panel for InterProcessor Communication in an Autonomous Mobile Robot Controller. In proc. of IROS'90, 1990.
|
1995
|
117
|
1,018
|
Optimization Principles for the Neural Code Michael DeWeese Sloan Center, Salk Institute La Jolla, CA 92037 deweese@salk.edu Abstract Recent experiments show that the neural codes at work in a wide range of creatures share some common features. At first sight, these observations seem unrelated. However, we show that these features arise naturally in a linear filtered threshold crossing (LFTC) model when we set the threshold to maximize the transmitted information. This maximization process requires neural adaptation to not only the DC signal level, as in conventional light and dark adaptation, but also to the statistical structure of the signal and noise distributions. We also present a new approach for calculating the mutual information between a neuron's output spike train and any aspect of its input signal which does not require reconstruction of the input signal. This formulation is valid provided the correlations in the spike train are small, and we provide a procedure for checking this assumption. This paper is based on joint work (DeWeese [1], 1995). Preliminary results from the LFTC model appeared in a previous proceedings (DeWeese [2], 1995), and the conclusions we reached at that time have been reaffirmed by further analysis of the model. 1 Introduction Most sensory receptor cells produce analog voltages and currents which are smoothly related to analog signals in the outside world. Before being transmitted to the brain, however, these signals are encoded in sequences of identical pulses called action potentials or spikes. We would like to know if there is a universal principle at work in the choice of these coding strategies. The existence of such a potentially powerful theoretical tool in biology is an appealing notion, but it may not turn out to be useful. Perhaps the function of biological systems is best seen as a complicated compromise among constraints imposed by the properties of biological materials, the need to build the system according to a simple set of development rules, and 282 M. DEWEESE the fact that current systems must arise from their ancestors by evolution through random change and selection. In this view, biology is history, and the search for principles (except for evolution itself) is likely to be futile. Obviously, we hope that this view is wrong, and that at least some of biology is understandable in terms of the same sort of universal principles that have emerged in the physics of the inanimate world. Adrian noticed in the 1920's that every peripheral neuron he checked produced discrete, identical pulses no matter what input he administered (Adrian, 1928). From the work of Hodgkin and Huxley we know that these pulses are stable non-linear waves which emerge from the non-linear dynamics describing the electrical properties of the nerve cell membrane These dynamics in turn derive from the molecular dynamics of specific ion channels in the cell membrane. By analogy with other nonlinear wave problems, we thus understand that these signals have propagated over a long distance e.g. ~ one meter from touch receptors in a finger to their targets in the spinal cord so that every spike has the same shape. This is an important observation since it implies that all information carried by a spike train is encoded in the arrival times of the spikes. Since a creature's brain is connected to all of its sensory systems by such axons, all the creature knows about the outside world must be encoded in spike arrival times. Until recently, neural codes have been studied primarily by measuring changes in the rate of spike production by different input signals. Recently it has become possible to characterize the codes in information-theoretic terms, and this has led to the discovery of some potentially universal features of the code (Bialek, 1996) (or see (Bialek, 1993) for a brief summary). They are: 1. Very high information rates. The record so far is 300 bits per second in a cricket mechanical sensor. 2. High coding efficiency. In cricket and frog vibration sensors, the information rate is within a factor of 2 of the entropy per unit time of the spike train. 3. Linear decoding. Despite evident non-linearities ofthe nervous system, spike trains can be decoded by simple linear filters. Thus we can write an estimate of the analog input signal s(t) as Sest (t) = Ei Kl (t - td, with Kl chosen to minimize the mean-squared errors (X2 ) in the estimate. Adding non-linear K2(t - ti, t - tj) terms does not significantly reduce X2 . 4. Moderate signal-to-noise ratios (SNR). The SNR in these experiments was defined as the ratio of power spectra of the input signal to the noise referred back to the input; the power spectrum of the noise was approximated by X2 defined above. All these examples of high information transmission rates have SNR of order unity over a broad bandwidth, rather than high SNR in a narrow band. We will try to tie all of these observations together by elevating the first to a principle: The neural code is chosen to maximize information transmission where information is quantified following Shannon. We apply this principle in the context of a simple model neuron which converts analog signals into spike trains. Before we consider a specific model, we will present a procedure for expanding the information rate of any point process encoding of an analog signal about the limit where the spikes are uncorrelated. We will briefly discuss how this can be used to measure information rates in real neurons. Optimization Principles for the Neural Code 283 This work will also appear in Network. 2 Information Theory In the 1940's, Shannon proposed a quantitative definition for "information" (Shannon, 1949). He argued first that the average amount of information gained by observing some event Z is the entropy of the distribution from which z is chosen, and then showed that this is the only definition consistent with several plausible requirements. This definition implies that the amount of information one signal can provide about some other signal is the difference between the entropy of the first signal's a priori distribution and the entropy of its conditional distribution. The average of this quantity is called the mutual (or transmitted) information. Thus, we can write the amount of information that the spike train, {td, tells us about the time dependent signal, s(t), as (1) where I1Jt; is shorthand for integration over all arrival times {til from 0 to T and summation over the total number of spikes, N (we have divided the integration measure by N! to prevent over counting due to equivalent permutations of the spikes, rather than absorb this factor into the probability distribution as we did in (DeWeese [1], 1995)). < ... >8= I1JsP[sO]'" denotes integration over the space offunctions s(t) weighted by the signal's a priori distribution, P[{t;}ls()] is the probability distribution for the spike train when the signal is fixed and P[{t;}] is the spike train's average distribution. 3 Arbitrary Point Process Encoding of an Analog Signal In order to derive a useful expression for the information given by Eq. (1), we need an explicit representation for the conditional distribution of the spike train. If we choose to represent each spike as a Dirac delta function, then the spike train can be defined as N p(t) = L c5(t - t;). (2) ;=1 This is the output spike train for our cell, so it must be a functional of both the input signal, s(t), and all the noise sources in the cell which we will lump together and call '7(t). Choosing to represent the spikes as delta functions allows us to think of p(t) as the probability of finding a spike at time t when both the signal and noise are specified. In other words, if the noise were not present, p would be the cell's firing rate, singular though it is. This implies that in the presence of noise the cell's observed firing rate, r(t), is the noise average of p(t): r(t) = J 1J'7P ['70Is0]p(t) = (p(t))'1' (3) Notice that by averaging over the conditional distribution for the noise rather than its a priori distribution as we did in (DeWeese [1], 1995), we ensure that this expression is still valid if the noise is signal dependent, as is the case in many real neurons. For any particular realization of the noise, the spike train is completely specified which means that the distribution for the spike train when both the signal and 284 M. DEWEESE noise are fixed is a modulated Poisson process with a singular firing rate, p(t). We emphasize that this is true even though we have assumed nothing about the encoding of the signal in the spike train when the noise is not fixed. One might then assume that the conditional distribution for the spike tra.in for fixed signal would be the noise average of the familiar formula for a modulated Poisson process: (4) However, this is only approximately true due to subtleties arising from the singular nature of p(t). One can derive the correct expression (DeWeese [1], 1995) by carefully taking the continuum limit of an approximation to this distribution defined for discrete time. The result is the same sum of noise averages over products of p's produced by expanding the exponential in Eq. (4) in powers of f dtp(t) except that all terms containing more than one factor of p(t) at equal times are not present. The exact answer is: (5) where the superscripted minus sign reminds us to remove all terms containing products of coincident p's after expanding everything in the noise average in powers of p. 4 Expanding About the Poisson Limit An exact solution for the mutual information between the input signal and spike train would be hopeless for all but a few coding schemes. However, the success of linear decoding coupled with the high information rates seen in the experiments suggests to us that the spikes might be transmitting roughly independent information (see (DeWeese [1], 1995) or (Bialek, 1993) for a more fleshed out argument on this point). If this is the case, then the spike train should approximate a Poisson process. We can explicitly show this relationship by performing a cluster expansion on the right hand side of Eq. (5): (6) where we have defined ~p(t) == p(t)- < p(t) >'1= p(t) - r(t) and introduced C'1(m) which collects all terms containing m factors of ~p. For example, C (2) == ~ ,,(~Pi~Pj}q - J dt' ~ (~p' ~Pi}q + ~ J dt'dt"(~ '~ ")-. '1 2 L..J r·r · L..J r · 2 p P '1 i¢j , J i=l' (7) Clearly, if the correlations between spikes are small in the noise distribution, then the C'1 's will be small, and the spike train will nearly approximate a modulated Poisson process when the signal is fixed. Optimization Principles for the Neural Code 285 Performing the cluster expansion on the signal average of Eq. (5) yields a similar expression for the average distribution for the spike train: (8) where T is the total duration of the spike train, r is the average firing rate, and C'1. 8 (m) is identical to C'1(m) with these substitutions: r(t) --+ r, ~p(t) --+ ap(t) == p(t) - f, and ( ... ); --+ {{ .. ·);)8. In this case, the distribution for a homogeneous Poisson process appears in front of the square brackets, and inside we have 1 + corrections due to correlations in the average spike train. 5 The Transmitted Information Inserting these expressions for P[ {til IsO] and P[ {til] (taken to all orders in ~p and ap, respectively) into Eq. (1), and expanding to second non-vanishing order in fTc results in a useful expression for the information (DeWeese [1], 1995): (9) where we have suppressed the explicit time notation in the correction term inside the double integral. If the signal and noise are stationary then we can replace the I; dt in front of each of these terms by T illustrating that the information does indeed grow linearly with the duration of the spike train. The leading term, which is exact if there are no correlations between the spikes, depends only on the firing rate, and is never negative. The first correction is positive when the correlations between pairs of spikes are being used to encode the signal, and negative when individual spikes carry redundant information. This correction term is cumbersome but we present it here because it is experimentally accessible, as we now describe. This formula can be used to measure information rates in real neurons without having to assume any method of reconstructing the signal from the spike train. In the experimental context, averages over the (conditional) noise distribution become repeated trials with the same input signal, and averages over the signal are accomplished by summing over all trials. r(t), for example, is the histogram of the spike trains resulting from the same input signal, while f(t) is the histogram of all spike trains resulting from all input signals. If the signal and noise are stationary, then f will not be time dependent. {p(t)p(t'))'1 is in general a 2-dimensional histogram which is signal dependent: It is equal to the number of spike trains resulting from some specific input signal which simultaneously contain a spike in the time bins containing t and t'. If the noise is stationary, then this is a function of only t - t', and it reduces to a 1-dimensional histogram. In order to measure the full amount of information contained in the spike train, it is crucial to bin the data in small enough time bins to resolve all of the structure in 286 M. DEWEESE r(t), (p(t)p(t'))'l' and so on. We have assumed nothing about the noise or signal; in fact, they can even be correlated so that the noise averages are signal dependent without changing the experimental procedure. The experimenter can also choose to fix only some aspects of the sensory data during the noise averaging step, thus measuring the mutual information between the spike train and only these aspects of the input. The only assumption we have made up to this point is that the spikes are roughly uncorrelated which can be checked by comparing the leading term to the first correction, just as we do for the model we discuss in the next section. 6 The Linear Filtered Threshold Crossing Model As we reported in a previous proceedings (DeWeese [2], 1995) (and see (DeWeese [1], 1995) for details), the leading term in Eq. (9) can be calculated exactly in the case of a linear filtered threshold crossing (LFTC) model when the signal and noise are drawn from independent Gaussian distributions. Unlike the Integrate and Fire (IF) model, the LFTC model does not have a "renewal process" which resets the value of the filtered signal to zero each time the threshold is reached. Stevens and Zador have developed an alternative formulation for the information transmission which is better suited for studying the IF model under some circumstances (Stevens, 1995), and they give a nice discussion on the way in which these two formulations compliment each other. For the LFTC model, the leading term is a function of only three variables: 1) The threshold height; 2) the ratio of the variances of the filtered signal and the filtered noise, (s2(t)),/(7J2(t))'l' which we refer to as the SNR; 3) and the ratio of correlation times ofthe filtered signal and the filtered noise, T,/T'l' where T; == (S2(t)),/(S2(t)), and similarly for the noise. In the equations in this last sentence, and in what follows, we absorb the linear filter into our definitions for the power spectra of the signal and noise. Near the Poisson limit, the linear filter can only affect the information rate through its generally weak influence on the ratios of variances and correlation times of the signal and noise, so we focus on the threshold to understand adaptation in our model cell. When the ratio of correlation times of the signal and noise is moderate, we find a maximum for the information rate near the Poisson limit the leading term ~ lOx the first correction. For the interesting and physically relevant case where the noise is slightly more broadband than the signal as seen through the cell's prefiltering, we find that the maximum information rate is achieved with a threshold setting which does not correspond to the maximum average firing rate illustrating that this optimum is non-trivial. Provided the SNR is about one or less, linear decoding does well a lower bound on the information rate based on optimal linear reconstruction of the signal is within a factor of two of the total available information in the spike train. As SNR grows unbounded, this lower bound asymptotes to a constant. In addition, the required timing resolution for extracting the information from the spike train is quite modest discretizing the spike train into bins which are half as wide as the correlation time of the signal degrades the information rate by less than 10%. However, at maximum information transmission, the information per spike is low Rmaz/r ~ .7 bits/spike, much lower than 3 bits/spike seen in the cricket. This low information rate drives the efficiency down to 1/3 of the experimental values despite the model's robustness to timing jitter. Aside from the low information rate, the optimized model captures all the experimental features we set out to explain. Optimization Principles for the Neural Code 287 7 Concluding Remarks We have derived a useful expression for the transmitted information which can be used to measure information rates in real neurons provided the correlations between spikes are shorter range than the average inter-spike interval. We have described a method for checking this hypothesis experimentally. The four seemingly unrelated features that were common to several experiments on a variety of neurons are actually the natural consequences of maximizing the transmitted information. Specifically, they are all due to the relation between if and Tc that is imposed by the optimization. We reiterate our previous prediction (DeWeese [2], 1995; Bialek, 1993): Optimizing the code requires that the threshold adapt not only to cancel DC offsets, but it must adapt to the statistical structure of the signal and noise. Experimental hints at adaptation to statistical structure have recently been seen in the fly visual system (de Ruyter van Steveninck, 1994) and in the salamander retina (Warland, 1995). 8 References M. DeWeese 1995 Optimization Principles for the Neural Code (Dissertation, Princeton University) M. DeWeese and W. Bialek 1995 Information flow in sensory neurons II Nuovo Cimento l7D 733-738 E. D. Adrian 1928 The Basis of Sensation (New York: W. W. Norton) F. Rieke, D. Warland, R. de Ruyter van Steveninck, and W. Bialek 1996 Neural Coding (Boston: MIT Press) W. Bialek, M. DeWeese, F. Rieke, and D. Warland 1993 Bits and Brains: Information Flow in the Nervous System Physica A 200 581-593 C. E. Shannon 1949 Communication in the presence of noise, Proc. I. R. E. 37 10-21 C. Stevens and A. Zador 1996 Information Flow Through a Spiking Neuron in M. Hasselmo ed Advances in Neural Information Processing Systems, Vol 8 (Boston: MIT Press) (this volume) R.R. de Ruyter van Steveninck, W. Bialek, M. Potters, R.H. Carlson 1994 Statistical adaptation and optimal estimation in movement computation by the blowfly visual system, in IEEE International Conference On Systems, Man, and Cybernetics pp 302-307 D. Warland, M. Berry, S. Smirnakis, and M. Meister 1995 personal communication
|
1995
|
118
|
1,019
|
A Framework for Non-rigid Matching and Correspondence Suguna Pappu, Steven Gold, and Anand Rangarajan1 Departments of Diagnostic Radiology and Computer Science and the Yale Neuroengineering and Neuroscience Center Yale University New Haven, CT 06520-8285 Abstract Matching feature point sets lies at the core of many approaches to object recognition. We present a framework for non-rigid matching that begins with a skeleton module, affine point matching, and then integrates multiple features to improve correspondence and develops an object representation based on spatial regions to model local transformations. The algorithm for feature matching iteratively updates the transformation parameters and the correspondence solution, each in turn. The affine mapping is solved in closed form, which permits its use for data of any dimension. The correspondence is set via a method for two-way constraint satisfaction, called softassign, which has recently emerged from the neural network/statistical physics realm. The complexity of the non-rigid matching algorithm with multiple features is the same as that of the affine point matching algorithm. Results for synthetic and real world data are provided for point sets in 2D and 3D, and for 2D data with multiple types of features and parts. 1 Introduction A basic problem of object recognition is that of matching- how to associate sensory data with the representation of a known object. This entails finding a transformation that maps the features of the object model onto the image, while establishing a correspondence between the spatial features. However, a tractable class of transformation, e.g., affine, may not be sufficient if the object is non-rigid or has relatively independent parts. If there is noise or occlusion, spatial information alone may not be adequate to determine the correct correspondence. In our previous work in spatial point matching [1], the 2D affine transformation was decomposed into its Ie-mail address of authors: lastname-firstname@cs.yale.edu 796 S. PAPPU, S. GOLD, A. RANGARAJAN physical component elements, which does not generalize easily to 3D, and so, only a rigid 3D transformation was considered. We present a framework for non-rigid matching that begins with solving the basic affine point matching problem. The algorithm iteratively updates the affine parameters and correspondence in turn, each as a function of the other. The affine transformation is solved in closed form, which lends tremendous flexibility- the formulation can be used in 2D or 3D. The correspondence is solved by using a softassign [1] procedure, in which the two-way assignment constraints are solved without penalty functions. The accuracy of the correspondence is improved by the integration of multiple features. A method for non-rigid parameter estimation is developed, based on the assumption of a well-articulated model with distinct regions, each of which may move in an affine fashion, or can be approximated as such. Umeyama [3] has done work on parameterized parts using an exponential time tree search technique, and Wakahara [4] on local affine transforms, but neither integrates multiple features nor explicitly considers the non-rigid matching case, while expressing a one-to-one correspondence between points. 2 Affine Point Matching The affine point matching problem is formulated as an optimization problem for determining the correspondence and affine transformation between feature points. Given two sets of data points Xj E Rn-l, n = 3,4 .. . , i = 1, ... , J, the image, and Yk E Rn-l, n = 3,4, ... , k = 1, ... , K, the model, find the correspondence and associated affine transformation that best maps a subset of the image points onto a subset of the model point set. These point sets are expressed in homogeneous coordinates, Xj = (l,Xj), Yk = (1, Yk). {aij} = A E Rnxn is the affine transformation matrix. Note that{alj = 0 Vi} because of the homogeneous coordinates. Define the match variable Mjk where Mjk E [0,1]. For a given match matrix {Mjd, transformation A and I, an identity matrix of dimension n, Lj,k MjkllXj - (A + I)Yk112 expresses the similarity between the point sets. The term -a Lj,k Mjk, with parameter a > 0 is appended to this to encourage matches (else Mjk = 0 V i, k minimizes the function). To limit the range of transformations, the terms of the affine matrix are regularized via a term Atr(AT A) in the objective function, with parameter A, where tr(.) denotes the trace of the matrix. Physically, Xj may fully match to one Yk, partially match to several, or may not match to any point. A similar constraint holds for Yk. These are expressed as the constraints in the following optimization problem: (1) s.t. LMjk::S 1, Vk, LMjk::S 1, Vi and Mjk ~ 0 j k To begin, slack variables Mj,K+l and MJ+l,k are introduced so that the inequality constraints can be transformed into equality constraints: Lf~t Mjk = 1, Vk and Lf:/ Mjk = 1, Vi. Mj,K+l = 1 indicates that Xj does not match to any point in Yk. An equivalent unconstrained optimization problem to (2) is derived by relaxing the constraints via Lagrange parameters Ilj, l/k, and introducing an x log x barrier function, indexed by a parameter {3. A similar technique was used A Framework for Nonrigid Matching and Correspondence 797 [2] to solve the assignment problem. The energy function used is: J K+1 min max LMjkllXj - (A+ J)Yk112 + Atr(AT A) - a LMjk + LJLj(L Mjk -1) A,M ~,v . . . ),k ),k ) k=l K J+1 1 J+1 K+1 + LlIk(LMjk -1) + (j L L Mjk(1ogMjk -1) k j=l j=l k=l This is to be minimized with respect to the match variables and affine parameters while satisfying the constraints via Lagrange parameters. Using the recently developed soft assign technique, we satisfy the constraints explicitly. When A is fixed, we have an assignment problem. Following the development in [1], the assignment constraints are satisfied using soft assign , a technique for satisfying two-way (assignment) constraints without a penalty term that is analogous to softmax which enforces a one-way constraint. First, the match variables are initialized: (2) This is followed by repeated row-column normalization of the match variables until a stopping criterion is reached: M)"k = Mjk Mjk '""' then M j k = '""' M L--j' Mj'k L--k' jk' (3) When the correspondence between the two point sets is fixed, A can be solved in closed form, by holding M fixed in the objective function, and differentiating and solving for A: A = A*(M) = (L Mjk(Xj Y[ - YkY{»(L MjkYkY[ + AI)-l (4) j,k j,k The algorithm is summarized as: 1. INITIALIZE: Variables: A = 0, M = 0 Parameters: .Binitial, .Bupdate, .Bfinal T = Inner loop iterations, A 2. ITERATE: Do T times for a fixed value of .B Softassign: Re-initialize M*(A) and then (Eq. 2) until ilM small A*(M) updated (Eq. 4) 3. UPDATE: While.B < .Bfinal, .B ~.B * .Bupdate, Return to 2. The complexity of the algorithm is O(J K). Starting with small .Binitial permits many partial correspondences in the initial solution for M. As.B increases the correspondence becomes more refined. For large .Bfinal, M approaches a permutation matrix (adjusting appropriately for the slack variables). 3 Nonrigid Feature Matching: Affine Quilts Recognition of an object requires many different types of information working in concert. Spatial information alone may not be sufficient for representation, especially in the presence of noise. Additionally the affine transformation is limited in its inability to handle local variation in an object, due to the object's non-rigidity or to the relatively independent movement of its parts, e.g., in human movement. The optimization problem (2) easily generalizes to integrate multiple invariant features. A representation with multiple features has a spatial component indicating 798 S. PAPPU, S. GOLD, A. RANGARAJAN the location of a feature element. At that location, there may be invariant geometric characteristics, e.g., this point belongs on a curve, or non-geometric invariant features such as color, and texture. Let Xjr be the value of feature r associated with point Xj. The location of point Xj is the null feature. There are R features associated with each point Xj and Yk. Note that the match variable remains the same. The new objective function is identical to the original objective function, (2), appended by the term "£j,k,r MjkWr(Xjr - Ykr)2. The (Xjr - Ykr)2 quantity captures the similarity between invariant types of features, with Wr a weighting factor for feature r. Non-invariant features are not considered. In this way, the point matching algorithm is modified only in the re-initialization of M(A): Mjk = exp(-,8(IIXj - (I + A)Yk112 + "£rWr(Xjr - ykr )2 - a)) The rest of the algorithm remains unchanged. Decomposition of spatial transformations motivates classification of the B individual regions of an object and use of a "quilt" of local affine transformations. In the multiple affine scenario, membership to a region is known on the well-articulated model, but not on the image set. It is assumed that all points that are members of one region undergo the same affine transformation. The model changes by the addition of one subscript to the affine matrix, Ab(k) where b(k) is an operator that indicates which transformation operates on point k. In the algorithm, during the A(M) update, instead of a single update, B updates are done. Denote K(b) = {klb(k) = b}, i.e., all the points that are within region b. Then in the affine update, Ab = Ab(M) = (L:j, kEK(b) Mjk(Xj Y{ - YkY{))("£j, kEK(b) MjkYkY{ + AbI)-l However, the theoretical complexity does not change, since the B updates still only require summing over the points. 4 Experimental Results: Hand Drawn and Synthetic The speed for matching point sets of 50 points each is around 20 seconds on an SGI workstation with a R4400 processor. This is true for points in 2D, 3D and with extra features . This can be improved with a tradeoff in accuracy by adopting a looser schedule for the parameter ,8 or by changing the stopping criterion. In the hand drawn examples, the contours of the images are drawn, discretized and then expressed as a set of points in the plane. In Figure (1), the contours of the boy's face were drawn in two different positions, and a subset of the points were extracted to make up the point sets. In each set this was approximately 250 points. Note that even with the change in mood in the two pictures, the corresponding parts of the face are found. However, in Figure (2) spatial information alone is Figure 1: Correspondence with simple point features insufficient. Although the rotation ofthe head is not a true affine transformation, it A Framework for Nonrigid Matching and Correspondence 799 is a weak perspective projection for which the approximation is valid. Each photo is outlined, generating approximately 225 points in each face. A point on a contour Figure 2: Correspondence with multiple features has associated with it a feature marker indicating the incident textures. For a human face, we use a binary 4-vector, with a 1 in position r if feature r is present. Specifically, we have used a vector with elements [skin, hair, lip, eye]. For example, a point on the line marking the mouth segment the lip from the skin has a feature vector [1,0,1,0]. Perceptual organization of the face motivates this type of feature marking scheme. The correspondence is depicted in Figure (2) for a small subset of matches. Next, we demonstrate how the multiple affine works in recovering the correct correspondence and transformation. The points associated with the standing figure have a marker indicating its part membership. There are six parts in this figure: head, torso, each arm and each leg. The correspondence is shown in Figure (3). For synthetic data, all 2D and 3D single part experiments used this protocol: The model set was generated uniformly on a unit square. A random affine matrix is generated, whose parameters, aij are chosen uniformly on a certain interval, which is used to generate the image set. Then, Pd image points are deleted, and Gaussian noise, N(O, u) is added. Finally, spurious points, Ps are added. For the multiple feature scenario, the elements of the feature vector are randomly mislabelled with probability, Pr , to represent distortion. For these experiments, 50 model points were generated, and aij are uniform on an interval of length 1.5. u E {0.01, 0.02, ... , 0.08}. Point deletions and spurious additions range from 0% to 50% of the image points. The random feature noise associated with nonspatial features has a probability of Pr = 0.05. The error measure we use is ea = C Li,j laij -a.ij I where c = # par:meters interva~ length· aij and a.ij are the correct parameter and the computed value, respectively. The constant term c normalizes the measure so that the error equals 1 in the case that the aij and aij are chosen at random on this interval. The factor 3 in the numerator of this formula follows since 800 S. PAPPU, S. GOLD, A. RANGARAJAN ~·~--··············~-··-····· ·········R Il · ·······························~ '-' . . _ ................... , ---................ ~ • . . .;.;;:...':":-:-:~~ ... -... ~~:~~.~.:~~g ~··~·~-·-··-···············:······<:"!»>,.~Il ;']~f Figure 3: Articulated Matching: Figure with six parts Elx - yl = ~, when x and yare chosen randomly on the unit interval, and we want to normalize the error. The parameters used in all experiments were: ,Binitial = .091, ,Bfmal = 100, ,Bupdate = 1.075, and T = 4. The model has four regions, 24 parameters. Points corresponding to part 1 were centered at (.5, .5), and generated randomly with a diameter of 1.0. For the image set, an affine transformation was applied with a translation diameter of .5, i.e., for a21, an, and the remaining four parameters have a diameter of 1. Points corresponding to regions 2, 3, and 4 were centered at (-.5, .5), (-.5, -.5), (.5, -.5) with model points and transformations generated in a similar fashion. 120 points were generated for the model point set, divided equally among the four parts. Image points were deleted with equal probability from each region. Spurious point were not explicitly added, since the overlapping of parts provides implicit spurious points. Results for the 2D and 3D (simple point) experiments are in Figure (4). Each data point represents 500 runs for a different randomly generated affine transformation. In all experiments, note that the error for small amounts of noise is approximately equal to that when there is no noise. We performed similar experiments for point sets that are 3-dimensional (12 parameters), but without any feature information. For the experiments with features, shown in Figure (5) we used R = 4 features, and Wr = 0.2, Vr. Each data point represents 500 runs.As expected, the inclusion of feature information reduces the error, especially for large u. Additionally, Figure (5) details synthetic results for experiments with multiple affines (2D). Each data point represents 70 runs. 5 Conclusion We have developed an affine point matching module, robust in the presence of noise and able to accommodate data of any dimension. The module forms the basis for a non-rigid feature matching scheme in which multiple types of features interact to establish correspondence. Modeling an object in terms of its spatial regions and then using multiple affines to capture local transformations results in a tractable method for non-rigid matching. This non-rigid matching framework arising out of A Framework for Nonrigid Matching and Correspondence 20 Results 0 .25r--~--~-~--..., ! 0.2 ~ 0.15 ..e-g 0.1 Q) ~0 .05 o o o o x o x o x O~----~-----~ 0.02 0.04 0.06 0.08 Standard deviation: Jitter -. : Pd = 0%,P8 = 0%, + : Pd = 10%,P8 = 10%, 3D Resutts 0.25~~--~-~-~-, ! 0.2 ~ 0.15 0 0 .e0 g 0.1 0 0 0 0 0 x x : x x x x x + + ~0 .05 + + _ .~ . J_!.-+- . - ' -' o~----~-----~ 0.02 0.04 0.06 0.08 Standard deviation: Jitter 0: Pd = 50%,P8 = 10% X: Pd = 30%,P8 = 10% Figure 4: Synthetic Experiments: 2D and 3D iii "ai ~ 0.1 ., Q. .." o 4 Features --ai 0.05 " • • '''' ... . -., ...... _.-' • • .-iii 0.25 "ai ~ 0.2 c;; %0.15 e iii 0.1 ~0 .05 4 Parts x x 0 x 0 ." . "" ~ .9·-" . " -.x x 0 0 ..0 ..o~----~-----~ 0.02 0.04 0.06 0.08 o~----~-----~ 0.02 0.04 0.06 0.08 Standard deviation: Jitter Standard deviation: Jitter .- :Pd = 0%,P8 = 0% * : Pd = 10%,P8 = 10% 0: Pd = 10%,P8 = 0% . : Pd = 30%,P8 = 10%, X: Pd = 25%,P8 = 0% -- : Pd = 50%,P8 = 10%, - : Pd = 40%,P8 = 0% Figure 5: Synthetic Experiments: Multiple features and parts 801 neural computation is widely applicable in object recognition. Acknowledgements: Our thanks to Eric Mjolsness for many interesting discussions related to the present work. References [1] S. Gold, C. P. Lu, A. Rangarajan, S. Pappu, and E. Mjolsness. New algorithms for 2D and 3D point matching: Pose estimation and correspondence. In G. Tesauro, D. Touretzky, and J. Alspector, editors, Advances in Neural Information Processing Systems, volume 7, San Francisco, CA, 1995. Morgan Kaufmann Publishers. [2] J. Kosowsky and A. Yuille. The invisible hand algorithm: Solving the assignment problem with statistical physics. Neural Networks, 7:477-490, 1994. [3] S. Umeyama. Parameterized point pattern matching and its application to recognition of object families. IEEE Trans. on Pattern Analysis and Machine Intelligence, 15:136-144,1993. [4] T . Wakahara. Shape matching using LAT and its application to handwritten numeral recognition. IEEE Trans. in Pattern Analysis and Machine Intelligence, 16:618- 629, 1994.
|
1995
|
119
|
1,020
|
Analog VLSI Processor Implementing the Continuous Wavelet Transform R. Timothy Edwards and Gert Cauwenberghs Department of Electrical and Computer Engineering Johns Hopkins University 3400 North Charles Street Baltimore, MD 21218-2686 {tim,gert}@bach.ece.jhu.edu Abstract We present an integrated analog processor for real-time wavelet decomposition and reconstruction of continuous temporal signals covering the audio frequency range. The processor performs complex harmonic modulation and Gaussian lowpass filtering in 16 parallel channels, each clocked at a different rate, producing a multiresolution mapping on a logarithmic frequency scale. Our implementation uses mixed-mode analog and digital circuits, oversampling techniques, and switched-capacitor filters to achieve a wide linear dynamic range while maintaining compact circuit size and low power consumption. We include experimental results on the processor and characterize its components separately from measurements on a single-channel test chip. 1 Introduction An effective mathematical tool for multiresolution analysis [Kais94], the wavelet transform has found widespread use in various signal processing applications involving characteristic patterns that cover multiple scales of resolution, such as representations of speech and vision. Wavelets offer suitable representations for temporal data that contain pertinent features both in the time and frequency domains; consequently, wavelet decompositions appear to be effective in representing wide-bandwidth signals interfacing with neural systems [Szu92]. The present system performs a continuous wavelet transform on temporal one-dimensional analog signals such as speech, and is in that regard somewhat related to silicon models of the cochlea implementing cochlear transforms [Lyon88], [Liu92] , [Watt92], [Lin94]. The multiresolution processor we implemented expands on the architecture developed in [Edwa93], which differs from the other analog auditory processors in the way signal components in each frequency band are encoded. The signal is modulated with the center Analog VLSI Processor Implementing the Continuous Wavelet Transform 693 _: '\/'V 1m - I s'(t) set) LPF ~ Lff ~ LPF x(t) ~ yet) x(t) x ~ ~ yet) get) h(t) Multiplier h(t) Prefilter Multiplexer (a) (b) Figure 1: Demodulation systems, (a) using multiplication, and (b) multiplexing. frequency of each channel and subsequently lowpass filtered, translating signal components taken around the center frequency towards zero frequency. In particular. we consider wavelet decomposition and reconstruction of analog continuous-time temporal data with a complex Gaussian kernel according to the following formulae: Yk(t) {teo x(e) exp (jWke- Q(Wk(t - e))2) de (decomposition) (1) x'(t) C 2::y\(t) exp(-jwkt) k (reconstruction) where the center frequencies Wk are spaced on a logarithmic scale. The constant Q sets the relative width of the frequency bins in the decomposition, and can be adjusted (together with C) alter the shape of the wavelet kernel. Successive decomposition and reconstruction transforms yield an approximate identity operation; it cannot be exact as no continuous orthonormal basis function exists for the CWT [Kais94]. 2 Architecture The above operations are implemented in [Edwa93] using two demodulator systems per channel, one for the real component of (1), and another for the imaginary component, 90° out of phase with the first. Each takes the form of a sinusoidal modulator oscillating at the channel center frequency, followed by a Gaussian-shaped lowpass filter, as shown in Figure 1 (a). This arrangement requires a precise analog sine wave generator and an accurate linear analog multiplier. In the present implementation, we circumvent both requirements by using an oversampled binary representation of the modulation reference signal. 2.1 Multiplexing vs. MUltiplying Multiplication of an analog signal x(t) with a binary (± 1) sequence is naturally implemented with high precision using a mUltiplexer, which alternates between presenting either the input or its inverse -x(t) to the output. This principle is applied to simplify harmonic modulation. and is illustrated in Figure 1 (b). The multiplier has been replaced by an analog inverter followed by a multiplexer, where the multiplexer is controlled by an oversampled binary periodic sequence representing the sine wave reference. The oversampled binary sequence is chosen to approximate the analog sine wave as closely as possible. disregarding components at high frequency which are removed by the subsequent lowpass filter. The assumption made is that no high frequency components are present in the input signal 694 SiglUll ,---- • ________ ----I In : eLK/: , , , , In Seltrt CLK2 , ' I. ________________ : Reconstruction Input Mult;pl;er R. T. EDWARDS, G. CAUWENBERGHS eLK] f ---<.-iK4 ----1:CLKS---: ;-----------------1 II • I II " E I ' 'I .---f-..., :: : : Ret'onJlructed : : cmLy, 0.,. i " , .--...L..--,:: : I I : ~~--+,~, , '-==:.J : ~ ____ __________ __ J Gaussian Filter Output Muxing Wavelet Reconstruction Figure 2: Block diagram of a single channel in the wavelet processor, showing test points A through E. under modulation, which otherwise would convolve with corresponding high frequency components in the binary sequence to produce low frequency distortion components at the output. To that purpose, an additionallowpass filter is added in front of the multiplexer. Residual low-frequency distortion at the output is minimized by maximizing roll-off of the filters, placing proper constraints on their cutofffrequencies, and optimally choosing the bit sequence in the oversampled reference [Edwa95]. Clearly, the signal accuracy that can be achieved improves as the length N of the sequence is extended. Constraints on the length N are given by the implied overhead in required signal bandwidth, power dissipation, and complexity of implementation. 2.2 Wavelet Gaussian Function The reason for choosing a Gaussian kernel in (l) is to ensure optimal support in both time and frequency [Gros89]. A key requirement in implementing the Gaussian filter is linear phase, to avoid spectral distortion due to non-uniform group delays. A worryfree architecture would be an analog FIR filter; however the number of taps required to accommodate the narrow bandwidth required would be prohibitively large for our purpose. Instead, we approximate a Gaussian filter by cascading several first-order lowpass filters. From probabilistic arguments, the obtained lowpass filter approximates a Gaussian filter increasingly well as the number of stages increases [Edwa93]. 3 Implementation Two sections of a wavelet processor, each containing 8 parallel channels, were integrated onto a single 4 mm x 6 mm die in 2 /lm CMOS technology. Both sections can be configured to perform wavelet decomposition as well as reconstruction. The block diagram for one of the channels is shown in Figure 2. In addition, a separate test chip was designed which performs one channel of the wavelet function. Test points were made available at various points for either input or output, as indicated in boldface capitals, A through E, in Figure 2. Each channel performs complex harmonic modulation and Gaussian lowpass filtering, as defined above. At the front end of the chip is a sample-and-hold section to sample timemultiplexed wavelet signals for reconstruction. In cases of both signal decomposition and reconstruction, each channel removes the input DC component removed, filters the result through the premultiplication lowpass (PML) filter, inverts the result, and passes both non-inverted and inverted signals onto the multiplexer. The multiplexer output is passed through a postmultiplication lowpass filter (PML, same architecture) to remove high frequency components of the oversampled sequence, and then passed through the Gaussianshaped lowpass filter. The cutoff frequencies of all filters are controlled by the clock rates Analog VLSI Processor Implementing the Continuous Wavelet Transform 695 (CLKI to CLK4 in Figure 2). The remainder of the system is for reconstruction and for time-multiplexing the output. 3.1 MUltiplier The multiplier is implemented by use of the above multiplexing scheme, driven by an oversampled binary sequence representing a sine wave. The sequence we used was 256 samples in length, created from a 64-sample base sequence by reversal and inversion. The sequence length of256 generates a modulator wave of 4 kHz (useful for speech applications) from a clock of about 1 MHz. We derived a sequence which, after postfiltering through a 3rd-order lowpass filter of the fonn of the PML prefilter (see below), produces a sine wave in which all hannonics are more than 60 dB down from the primary [Edwa95]. The optimized 64-bit base sequence consists of 11 zeros and 53 ones, allowing a very simple implementation in which an address decoder decodes the "zero" bits. The binary sequence is shown in Figure 4. The magnitude of the prime hannonic of the sequence is approximately 1.02, within 2% of unity. The process of reversing and inverting the sequence is simplified by using a gray code counter to produce the addresses for the sequence, with only a small amount of combinatorial logic needed to achieve the desired result [Edwa95]. It is also straightforward to generate the addresses for the cosine channel, which is 90° out of phase with the original. 3.2 Linear Filtering All filters used are implemented as linear cascades of first-order, single-pole filter sections. The number of first-order sections for the PML filters is 3. The number of sections for the "Gaussian" filter is 8, producing a suitable approximation to a Gaussian filter response for all frequencies of interest (Figure 5). Figure 3 shows one first-order lowpass section of the filters as implemented. This standard >-.+--o va"' v,,, + Figure 3: Single discrete-time lowpass filter section. switched-capacitor circuit implements a transfer function containing a single pole, approximately located in the Laplace domain at s = Is / a for large values of the parameter a, with Is being the sampling frequency. The value for this parameter a is fixed at the design stage as the ratio of two capacitors in Figure 3, and was set to be 15 for the The PML filters and 12 for the Gaussian filters. 4 Measured Results 4.1 Sine wave modulator We tested the accuracy of the sine wave modulation signal by applying two constant voltages at test points A and B, such that the sine wave modulation signal is effectively multiplied 696 R. T. EDWARDS, G. CAUWENBERGHS Sine sequence and filtered sine wave output Binary sine sequence Simulated filtered output x Measured output -1.5 L-___ --' ____ -'-____ ........ ____ -'-____ .J...J o 50 100 150 200 250 Time (us) Figure 4: Filtered sine wave output. by a constant. The output of the mUltiplier is filtered and the output taken at test point D, before the Gaussian filter. Figure 4 shows the (idealized) multiplexer output at test point C, which accurately creates the desired binary sequence. Figure 4 also shows the measured sine wave after filtering with the PML filter and the expected output from the simulation model, using a deviating value of 8.0 for the capacitor ratio a, as justified below. FFT analysis of Figure 4 has shown that the resulting sine wave has all harmonics below about -49 dB. This is in good agreement with the simulation model, provided a correction is made for the value of the capacitor ratio a to account for fringe and (large) parasitic capacitances. The best fit for the measured data from the postmultiplication filter is a = 8.0, compared to the desired value of a = 15.0. The transform of the simulated output shown in the figure takes into account the smaller value of a. Because the postmultiplication filter is followed by the Gaussian filter, the bandwidth of the output can be directly controlled by proper clocking ofthe Gaussian filter, so the distortion in the sine wave is ultimately much smaller than that measured at the output of the postmultiplication filter. 4.2 Gaussian filter The Gaussian filter was tested by applying a signal at test point D and measuring the response at test point E. Figure 5 shows the response of the Gaussian filter as compared to expected responses. There are two sets of curves, one for a filter clocked at 64 kHz, and the other clocked at 128 kHz; these curves are normalized by plotting time relative to the clock frequency is. The solid line indicates the best match for an 8th-order lowpass filter, using the capacitor ratio, a, as a fitting parameter. The best-fit value of a is approximately 6.8. This is again much lower than the capacitor area ratio of 12 on the chip. The dotted line is the response of the ideal Gaussian characteristic exp ( _w 2 / (2aw~)) approximated by the cascade of first-order sections with capacitor ratio a. Figure 5 (b) shows the measured phase response of the Gaussian filter for the 128 kHz clock. The phase response is approximately linear throughout the passband region. Analog VLSI Processor Implementing the Continuous Wavelet Transform 697 Gaussian filter response o~~~~--~-=~~~~~~~--~ x x Chip data at 64kHz clock o 0 Chip data at 128kHz clock iii'-10 ~ <lJ ]1 -20 ~ E « -30 "0 <lJ N ~ -40 § o Z -50 0.01 8th-order filter ideal response Gaussian filter ideal response 0.07 0.08 Frequency (units fs) 500 Theoretical 8-stage phase o Measured response 0.0 I 0.02 0.03 0.04 0.05 0.06 0.07 0.08 Frequency (units fs) Figure 5: Gaussianfilter transfer functions: theoretical and actual. (a) Relative amplitude; (b) Phase. 4.3 Wavelet decomposition Figure 6 shows the test chip performing a wavelet transform on a simple sinusoidal input, illustrating the effects of (oversampled) sinusoidal modulation followed by lowpass filtering through the Gaussian filter. The chip multiplier system is clocked at 500 kHz. The input wave is approximately 3.1 kHz, close to the center frequency of the modulator signal, which is the clock rate divided by 128, or about 3.9 kHz (a typical value for the highestfrequency channel in an auditory application). The top trace in the figure shows the filtered and inverted input, taken from test point B. The middle trace shows the output of the multiplexer (test point C), wherein the output is multiplexed between the signal and its inverse. The bottom trace is taken from the system output (labeled Cosine Out in Figure 2) and shows the demodulated signal of frequency 800 Hz (= 3.9 kHz - 3.1 kHz). Not shown is the cosine output, which is 90° out of phase with the one shown. This demonstrates the proper operation of complex demodulation in a single channel configured for wavelet decomposition. In addition, we have tested the full16-channel chip decomposition, and all individual parts function properly. The total power consumption of the 16-channel wavelet chip was measured to be less than 50mW, of which a large fraction can be attributed to external interfacing and buffering circuitry at the periphery of the chip. 5 Conclusions We have demonstrated the full functionality of an analog chip performing the continuous wavelet transform (decomposition). The chip is based on mixed analog/digital signal processing principles, and uses a demodulation scheme which is accurately implemented using oversampling methods. Advantages of the architecture used in the chip are an increased dynamic range and a precise control over lateral synchronization of wavelet components. An additional advantage inherent to the modulation scheme used is the potential to tune the channel bandwidths over a wide range, down to unusually narrow bands, since the cutoff frequency of the Gaussian filter and the center frequency of the modulator are independently adjustable and precisely controllable parameters. References G. Kaiser, A Friendly Guide to Wavelets, Boston, MA: Birkhauser, 1994. T. Edwards and M. Godfrey, "An Analog Wavelet Transform Chip," IEEE Int'l Can! on 698 R. T. EDWARDS, G. CAUWENBERGHS Figure 6: Scope trace of the wavelet transform: filtered input (top), multiplexed signal (middle), and wavelet output (bottom). Neural Networks, vol. III, 1993, pp. 1247-1251. T. Edwards and G. Cauwenberghs, "Oversampling Architecture for Analog Harmonic Modulation," to appear in Electronics Letters, 1996. A Grossmann, R Kronland-Martinet, and J. MorIet, "Reading and understanding continuous wavelet transforms," Wavelets: Time-Frequency Methods and Phase Space. SpringerVerlag, 1989, pp. 2-20. W. Liu, AG. Andreou, and M.G. Goldstein, "Voiced-Speech Representation by an Analog Silicon Model ofthe Auditory Periphery," IEEE T. Neural Networks, vol. 3 (3), pp 477-487, 1992. J. Lin, W.-H. Ki, T. Edwards, and S. Shamma, "Analog VLSI Implementations of Auditory Wavelet Transforms Using Switched-Capacitor Circuits," IEEE Trans. Circuits and Systems-I, vol.41 (9), pp. 572-583, September 1994. A Lu and W. Roberts, ''A High-Quality Analog Oscillator Using Oversampling D/A Conversion Techniques," IEEE Trans. Circuits and Systems-II, vol.41 (7), pp. 437-444, July 1994. RF. Lyon and C.A Mead, "An Analog Electronic Cochlea," IEEE Trans. Acoustics, Speech and Signal Proc., vol. 36, pp 1119-1134, 1988. H.H. Szu, B. Tefter, and S. Kadembe, "Neural Network Adaptive Wavelets for Signal Representation and Classification," Optical Engineering, vol. 31 (9), pp. 1907-1916, September 1992. L. Watts, D.A Kerns, and RF. Lyon, "Improved Implementation of the Silicon Cochlea," IEEE Journal of Solid-State Circuits, vol. 27 (5), pp 692-700,1992.
|
1995
|
12
|
1,021
|
Hierarchical Recurrent Neural Networks for Long-Term Dependencies Salah El Hihi Dept. Informatique et Recherche Operationnelle Universite de Montreal Montreal, Qc H3C-3J7 elhihiGiro.umontreal.ca Yoshua Bengio· Dept. Informatique et Recherche Operationnelle Universite de Montreal Montreal, Qc H3C-3J7 bengioyGiro.umontreal.ca Abstract We have already shown that extracting long-term dependencies from sequential data is difficult, both for determimstic dynamical systems such as recurrent networks, and probabilistic models such as hidden Markov models (HMMs) or input/output hidden Markov models (IOHMMs). In practice, to avoid this problem, researchers have used domain specific a-priori knowledge to give meaning to the hidden or state variables representing past context. In this paper, we propose to use a more general type of a-priori knowledge, namely that the temporal dependencIes are structured hierarchically. This implies that long-term dependencies are represented by variables with a long time scale. This principle is applied to a recurrent network which includes delays and multiple time scales. Experiments confirm the advantages of such structures. A similar approach is proposed for HMMs and IOHMMs. 1 Introduction Learning from examples basically amounts to identifying the relations between random variables of interest. Several learning problems involve sequential data, in which the variables are ordered (e.g., time series). Many learning algorithms take advantage of this sequential structure by assuming some kind of homogeneity or continuity of the model over time, e.g., bX sharing parameters for different times, as in Time-Delay Neural Networks (TDNNs) tLang, WaIbel and Hinton, 1990), recurrent neural networks (Rumelhart, Hinton and Williams, 1986), or hidden Markov models (Rabiner and Juang, 1986). This general a-priori assumption considerably simplifies the learning problem. In previous papers (Bengio, Simard and Frasconi, 1994· Bengio and Frasconi, 1995a), we have shown for recurrent networks and Markovian models that, even with this assumption, dependencies that span longer intervals are significantly harder to learn. In all of the systems we have considered for learning from sequential data, some form of representation of context ( or state) is required (to summarize all "useful" past information). The "hard learning" problem IS to learn to represent context, which involves performing the proper ° also, AT&T Bell Labs, Holmdel, NJ 07733 494 S. E. HIHI. Y. BENGIO credit assignment through time. Indeed, in practice, recurrent networks (e.g., injecting prior knowledge for grammar inference (Giles and Omlin, 1992; Frasconi et al., 1993)) and HMMs (e.g., for speech recognition (Levinson, Rabiner and Sondhi, 1983; Rabiner and Juang, 1986)) work quite well when the representation of context (the meaning of the state variable) is decided a-priori. The hidden variable is not any more completely hidden. Learning becomes much easier. Unfortunately, this requires a very precise knowledge of the appropriate state variables, which is not available in many applications. We have seen that the successes ofTDNNs, recurrent networks and HMMs are based on a general assumption on the sequential nature of the data. In this paper, we propose another, simple, a-priori assumption on the sequences to be analyzed: the temporal dependencies have a hierarchical structure. This implies that dependencies spanning long intervals are "robust" to small local changes in the timing of events, whereas dependencies spanning short intervals are allowed to be more sensitive to the precise timing of events. This yields a multi-resolution representation of state information. This general idea is not new and can be found in various approaches to learning and artificial intelligence. For example, in convolutional neural networks, both for sequential data with TDNNs (Lang, Waibel and Hinton, 1990), and for 2-dimensional data with MLCNNs (LeCun et al., 1989; Bengio, LeCun and Henderson, 1994), the network is organized in layers representing features of increasing temporal or spatial coarseness. Similarly, mostly as a tool for analyzing and preprocessing sequential or spatial data, wavelet transforms (Daubechies, 1990) also represent such information at mUltiple resolutions. Multi-scale representations have also been proposed to improve reinforcement learning systems (Singh, 1992; Dayan and Hinton, 1993; Sutton, 1995) and path planning systems. However, with these algorithms, one generally assumes that the state of the system is observed, whereas, in this paper we concentrate on the difficulty of learning what the state variable should represent. A related idea using a hierarchical structure was presented in (Schmidhuber, 1992). On the HMM side, several researchers (Brugnara et al., 1992; Suaudeau, 1994) have attempted to improve HMMs for speech recognition to better model the different types of var1ables, intrmsically varying at different time scales in speech. In those papers, the focus was on setting an a-priori representation, not on learning how to represent context. In section 2, we attempt to draw a common conclusion from the analyses performed on recurrent networks and HMMs to learn to represent long-term dependencies. This will justify the proposed approach, presented in section 3. In section 4 a specific hierarchical model is proposed for recurrent networks, using different time scales for different layers of the network. EXp'eriments performed with this model are described in section 4. Finally, we discuss a sim1lar scheme for HMMs and IOHMMs in section 5. 2 Too Many Products In this section, we take another look at the analyses of (Bengio, Simard and Frasconi, 1994) and (Bengio and Frasconi, 1995a), for recurrent networks and HMMs respectively. The objective 1S to draw a parallel between the problems encountered with the two approaches, in order to guide us towards some form of solution, and justify the proposals made here. First, let us consider the deterministic dynamical systems (Bengio, Simard and Frasconi, 1994) (such as recurrent networks), which map an input sequence U1 l . .. , UT to an output sequence Y1, ... , ftr· The state or context information is represented at each time t by a variable Xt, for example the activities of all the hidden units of a recurrent network: (1) where Ut is the system input at time t and 1 is a differentiable function (such as tanh(Wxt_1 + ut)). When the sequence of inputs U1, U2, • .. , UT is given, we can write Xt = It(Xt-d = It(/t-1( .. . l1(xo)) . .. ). A learning criterion Ct yields gradients on outputs, and therefore on the state variables Xt. Since parameters are shared across time, learning using a gradient-based algorithm depends on the influence of parameters W on Ct through an time steps before t : aCt _ " aCt OXt oXT oW L...J OXt OX T oW T (2) Hierarchical Recurrent Neural Networks for Long-term Dependencies 495 The Jacobian matrix of derivatives .!!:U{J{Jx can further be factored as follows: Xr (3) Our earlier analysis (Bengio, Simard and Frasconi, 1994) shows that the difficulty revolves around the matrix product in equation 3. In order to reliably "store" informatIOn in the dynamics of the network, the state variable Zt must remain in regions where If:! < 1 (i.e., near enough to a stable attractor representing the stored information). However, the above products then rapidly converge to 0 when t T increases. Consequently, the sum in 2 is dominated by terms corresponding to short-term dependencies (t T is small). Let us now consider the case of Markovian models (including HMMs and IOHMMs (Bengio and Frasconi, 1995b)). These are probabilistic models, either of an "output" sequence P(YI . . . YT) (HMMs) or of an output sequence given an input sequence P(YI ... YT lUI ... UT) (IOHMMs). Introducing a discrete state variable Zt and using Markovian assumptIOns of independence this probability can be factored in terms of transition probabilities P(ZtIZt-d (or P(ZtIZt-b ut}) and output probabilities P(ytlZt) (or P(ytiZt, Ut)) . According to the model, the distribution of the state Zt at time t given the state ZT at an earlier time T is given by the matrix P(ZtlZT) = P(ZtiZt-I)P(Zt-Ilzt-2) . .. P(zT+dzT) (4) where each of the factors is a matrix of transition probabilities (conditioned on inputs in the case of IOHMMs) . Our earlier analysis (Bengio and Frasconi, 1995a) shows that the difficulty in representing and learning to represent context (i.e., learning what Zt should represent) revolves around equation 4. The matrices in the above equations have one eigenvalue equal to 1 (because of the normalization constraint) and the others ~ 1. In the case in which all eIgenvalues are 1 the matrices have only i's and O's, i.e, we obtain deterministic dynamics for IOHMMs or pure cycles for HMMs (which cannot be used to model most interesting sequences). Otherwise the above product converges to a lower rank matrix (some or most of the eigenvalues converge toward 0). Consequently, P(ZtlZT) becomes more and more independent of ZT as t - T increases. Therefore, both representing and learning context becomes more difficult as the span of dependencies increases or when the Markov model is more non-deterministic (transition probabilities not close to 0 or 1). Clearly, a common trait of both analyses lies in taking too many products, too many time steps, or too many transformations to relate the state variable at time T with the state variable at time t > T, as in equations 3 and 4. Therefore the idea presented in the next section is centered on allowing several paths between ZT and Zt, some with few "transformations" and some with many transformations. At least through those with few transformations, we expect context information (forward), and credit assignment (backward) to propagate more easily over longer time spans than through "paths" lDvolving many tralIBformations. 3 Hierarchical Sequential Models Inspired by the above analysis we introduce an assumption about the sequential data to be modeled, although it will be a very simple and general a-priori on the structure of the data. Basically, we will assume that the sequential structure of data can be described hierarchically: long-term dependencies (e.g., between two events remote from each other in time) do not depend on a precise time scale (Le., on the precise timing of these events). Consequently, in order to represent a context variable taking these long-term dependencies into account, we will be able to use a coarse time scale (or a Slowly changing state variable). Therefore, instead of a single homogeneous state variable, we will introduce several levels of state variables, each "working" at a different time scale. To implement in a discretetime system such a multi-resolution representation of context, two basic approaches can be considered. Either the higher level state variables change value less often or they are constrained to change more slowly at each time step. In our ex~eriments, we have considered input and output variables both at the shortest time scale highest frequency), but one of the potential advantages of the approach presented here is t at it becomes very 496 S. E. IDHI, Y. BENOIO Figure 1: Four multi-resolution recurrent architectures used in the experiments. Small sguares represent a discrete delay, and numbers near each neuron represent its time scale. The architectures B to E have respectively 2, 3, 4, and 6 time scales. simple to incorporate input and output variables that operate at different time scales. For example, in speech recognition and synthesis, the variable of interest is not only the speech signal itself (fast) but also slower-varying variables such as prosodic (average energy, pitch, etc ... ) and phonemic (place of articulation, phoneme duration) variables. Another example is in the application of learning algorithms to financial and economic forecasting and decision taking. Some of the variables of interest are given daily, others weekly, monthly, etc ... 4 Hierarchical Recurrent Neural Network: Experiments As in TDNNs (Lang, Waibel and Hinton, 1990) and reverse-TDNNs (Simard and LeCun, 1992), we will use discrete time delays and subsampling (or oversampling) in order to implement the multiple time scales. In the time-unfolded network, paths going through the recurrences in the slow varying units (long time scale) will carry context farther, while paths going through faster varying units (short time scale) will respond faster to changes in input or desired changes in output. Examples of such multi-resolution recurrent neural networks are shown in Figure 1. Two sets of simple experiments were performed to validate some of the ideas presented in this paper. In both cases, we compare a hierarchical recurrent network with a single-scale fully-connected recurrent network. In the first set of experiments, we want to evaluate the performance of a hierarchical recurrent network on a problem already used for studying the difficulty in learning longterm dependencies (Bengio, Simard and Frasconi, 1994; Bengio and Frasconi, 1994). In this 2-class J?roblem, the network has to detect a pattern at the beginning of the sequence, keeping a blt of information in "memory" (while the inputs are noisy) until the end of the sequence (supervision is only a the end of the sequence). As in (Bengio, Simard and Frasconi, 1994; Bengio and Frasconi, 1994) only the first 3 time steps contain information about the class (a 3-number pattern was randomly chosen for each class within [-1,1]3). The length of the sequences is varied to evaluate the effect of the span of input/output dependencies. Uniformly distributed noisy inputs between -.1 and .1 are added to the initial patterns as well as to the remainder of the sequence. For each sequence length, 10 trials were run with different initial weights and noise patterns, with 30 training sequences. Experiments were performed with sequence of lengths 10, 20,40 and 100. Several recurrent network architectures were compared. All were trained with the same algorithm (back-propagation through time) to minimize the sum of squared differences between the final output and a desired value. The simplest architecture (A) is similar to architecture B in Figure 1 but it is not hierarchical: it has a single time scale. Like the Hierarchical Recurrent Neural Networks for Long-term Dependencies eo 50 40 l ~ 30 20 ABCDE ABCDE ABCDE ABCDE seq.1engIh 10 20 40 100 1.4 1.3 1.2 1.1 1.0 0.9 Is 0.8 Ii 0.7 I 0.6 0.5 0.4 ~~ '-----'-I....L.~~~....L.-.W....L.mL.......J....I.ll ABCDE ABCDE ABCDE ABCDE 1IIq. 1engIh 10 20 40 100 497 Figure 2: Average classification error after training for 2-sequence problem (left, classification error) and network-generated data (right, mean squared error), for varying sequence lengths and architectures. Each set of 5 consecutive bars represents the performance of 5 architectures A to E, with respectively 1, 2, 3, 4 and 6 time scales (the architectures B to E are shown in Figure 1). Error bars show the standard deviation over 10 trials. other networks, it has however a theoretically "sufficient" architecture, i.e., there exists a set of weights for which it classifies perfectly the trainin~ sequences. Four of the five architectures that we compared are shown in Figure 1, wIth an increasing number of levels in the hierarchy. The performance of these four architectures (B to E) as well as the architecture with a single time-scale (A) are compared in Figure 2 (left, for the 2sequence problem). Clearly, adding more levels to the hierarchy has significantly helped to reduce the difficulty in learning long-term dependencies. In a second set of experiments, a hierarchical recurrent network with 4 time scales was initialized with random (but large) weights and used to generate a data set. To generate the inputs as well as the outputs, the network has feedback links from hidden to input units. At the initial time step as well as at 5% of the time steps (chosen randomly), the input was clamped with random values to introduce some further variability. It is a regression task, and the mean squared error is shown on Figure 2. Because of the network structure, we expect the data to contain long-term dependencies that can be modeled with a hierarchical structure. 100 training sequences of length 10, 20,40 and 100 were generated by this network. The same 5 network architectures as in the previous experiments were compared (see Figure 1 for architectures B to E), with 10 training trials per network and per sequence length. The results are summarized in Figure 2 (right). More high-level hierarchical structure appears to have improved performance for long-term dependencies. The fact that the simpler I-level network does not achieve a good performance suggests that there were some difficult long-term dependencies in the the artificially generated data set. It is interesting to compare those results with those reported in (Lin et al., 1995) which show that using longer delays in certain recurrent connections helps learning longer-term dependencies. In both cases we find that introducing longer time scales allows to learn dependencies whose span is proportionally longer. 5 Hierarchical HMMs How do we represent multiple time scales with a HMM? Some solutions have already been proposed in the speech recognition literature, motivated by the obvious presence of different time scales in the speech phenomena. In (Brugnara et al., 1992) two Markov chains are coupled in a "master/slave" configuration. For the "master" HMM, the observations are slowly varying features (such as the signal energy), whereas for the "slave" HMM the observations are t.he speech spectra themselves. The two chains are synchronous and operate at the same time scale, therefore the problem of diffusion of credit in HMMs would probably also make difficult the learning of long-term dependencies. Note on the other 498 S. E. HIHI, Y. BENOIO hand that in most applications of HMMs to speech recognition the meaning of states is fixed a-priori rather than learned from the data (see (Bengio and Frasconi, 1995a) for a discussion). In a more recent contribution, Nelly Suaudeau (Suaudeau, 1994) proposes a "two-level HMM" in which the higher level HMM represents "segmental" variables (such as phoneme duration). The two levels operate at different scales: the higher level state varIable represents the phonetic identity and models the distributions of the average energy and the duration within each phoneme. Again, this work is not geared towards learning a representation of context, but rather, given the traditional (phoneme-based) representation of context in speech recognition, towards building a better model of the distribution of "slow" segmental variables such as phoneme duration and energy. Another promising approach was recently proposed in (Saul and Jordan, 1995). Using decimation techniques from statistical mechanics, a polynomial-time algorithm is derived for parallel Boltzmann chains (which are similar to parallel HMMs), which can operate at different time scales. The ideas presented here point toward a HMM or IOHMM in which the (hidden) state variable Xt is represented by the Cartesian product of several state variables Xt, each "working" at a different time scale: Xt = (x;, x~, ... I xf).. To take advantage of the decomposition, we propose to consider that tbe state dIstrIbutions at the different levels are conditionally independent (given the state at the previous time step and at the current and previous levels). Transition probabilities are therefore factored as followed: (5) To force the state variable at a each level to effectively work at a given time scale, selftransition probabilities are constrained as follows (using above independence assumptions): P(x:=i3Ixt_l=iI,.· ., x:_l=i3" .. , xt-l=is) = P(x:=i3Ix:_1 =i3, X::t=i3-d = W3 6 Conclusion Motivated by the analysis of the problem of learning long-term dependencies in sequential data, i.e., of learning to represent context, we have proposed to use a very general assumption on the structure of sequential data to reduce the difficulty of these learning tasks. Following numerous previous work in artificial intelligence we are assuming that context can be represented with a hierarchical structure. More precisely, here, it means that long-term dependencies are insensitive to small timing variations, i.e., they can be represented with a coarse temporal scale. This scheme allows context information and credit information to be respectively propagated forward and backward more easily. Following this intuitive idea, we have proposed to use hierarchical recurrent networks for sequence processing. These networks use multiple-time scales to achieve a multi-resolution representation of context. Series of experiments on artificial data have confirmed the advantages of imposing such structures on the network architecture. Finally we have proposed a similar application of this concept to hidden Markov models (for density estimation) and input/output hidden Markov models (for classification and regression). References Bengio, Y. and Frasconi, P. (1994). Credit assignment through time: Alternatives to backpropagation. In Cowan, J., Tesauro, G., and Alspector, J., editors, Advances in Neural Information Processing Systems 6. Morgan Kaufmann. Bengio, Y. and Frasconi, P. (1995a). Diffusion of context and credit information in markovian models. Journal of Artificial Intelligence Research, 3:223-244. Bengio, Y. and Frasconi, P. (1995b). An input/output HMM architecture. In Tesauro, G., Touretzky, D., and Leen, T., editors, Advances in Neural Information Processmg Systems 7, pages 427-434. MIT Press, Cambridge, MA. Bengio, Y., LeCun, Y., and Henderson, D. (1994). Globally trained handwritten word recognizer using spatial representation, space displacement neural networks and hidden Markov models. In Cowan, J ., Tesauro, G., and Alspector, J., editors, Advances in Neural Information Processing Systems 6, pages 937- 944. Hierarchical Recurrent Neural Networks for Long-term Dependencies 499 Bengio, Y., Simard, P., and Frasconi, P. (1994). Learning long-term dependencies with gradient descent is difficult. IEEE Transactions on Neural Networks, 5(2):157-166. Brugnara, F., DeMori, R, Giuliani, D., and Omologo, M. (1992). A family of parallel hidden markov models. In International Conference on Acoustics, Speech and Signal Processing, pages 377-370, New York, NY, USA. IEEE. Daubechies, I. (1990). The wavelet transform, time-frequency localization and signal analysis. IEEE Transaction on Information Theory, 36(5):961-1005 . Dayan, P. and Hinton, G. (1993). Feudal reinforcement learning. In Hanson, S. J., Cowan, J. D., and Giles, C. L., edItors, Advances in Neural Information Processing Systems 5, San Mateo, CA. Morgan Kaufmann. Frasconi, P., Gori, M., Maggini, M., and Soda, G. (1993). Unified integration of explicit rules and learning by example in recurrent networks. IEEE Transactions on Knowledge and Data Engineering. (in press). Giles, C. 1. and amlin, C. W. (1992). Inserting rules into recurrent neural networks. In Kung, Fallside, Sorenson, and Kamm, editors, Neural Networks for Signal Processing II, Proceedings of the 1992 IEEE workshop, pages 13-22. IEEE Press. Lang, K. J., Waibel, A. H., and Hinton, G. E. (1990). A time-delay neural network architecture for isolated word recognition. Neural Networks, 3:23-43. LeCun, Y., Boser, B., Denker, J., Henderson, D., Howard, R, Hubbard, W., and Jackel, L. (1989) . Backpropagation applied to handwritten zip code recognition. Neural Computation, 1:541-551. Levinson, S., Rabiner, 1., and Sondhi, M. (1983). An introduction to the application ofthe theory of probabilistic functions of a Markov process to automatic speech recognition. Bell System Technical Journal, 64(4):1035-1074. Lin, T ., Horne, B., Tino, P., and Giles, C. (1995). Learning long-term dependencies is not as difficult with NARX recurrent neural networks. Techmcal Report UMICAS-TR95-78, Institute for Advanced Computer Studies, University of Mariland. Rabiner, L. and Juang, B. (1986). An introduction to hidden Markov models. IEEE A SSP Magazine, pages 257-285. Rumelhart, D., Hinton, G., and Williams, R (1986). Learning internal representations by error propagation. In Rumelhart, D. and McClelland, J., editors, Parallel Distributed Processing, volume 1, chapter 8, pages 318-362. MIT Press, Cambridge. Saul, L. and Jordan, M. (1995). Boltzmann chains and hidden markov models. In Tesauro, G., Touretzky, D., and Leen, T., editor~ Advances in Neural Information Processing Systems 7, pages 435--442. MIT Press, vambridge, MA. Schmidhuber, J. (1992). Learning complex, extended sequences using the principle of history compression. Neural Computation, 4(2):234-242. Simard, P. and LeCun, Y. (1992). Reverse TDNN: An architecture for trajectory generation. In Moody, J., Hanson, S., and Lipmann, R, editors, Advances in Neural Information Processing Systems 4, pages 579-588, Denver, CO. Morgan Kaufmann, San Mateo. Singh, S. (1992). Reinforcement learning with a hierarchy of abstract models. In Proceedings of the 10th National Conference on Artificial Intelligence, pages 202-207. MIT / AAAI Press. Suaudeau, N. (1994). Un modele probabiliste pour integrer la dimension temporelle dans un systeme de reconnaissance automatique de la parole. PhD thesis, Universite de Rennes I, France. Sutton, RjI995). TD models: modeling the world at a mixture of time scales. In Proceedings 0 the 12th International Conference on Machine Learning. Morgan Kaufmann.
|
1995
|
120
|
1,022
|
Finite State Automata that Recurrent Cascade-Correlation Cannot Represent Stefan C. Kremer Department of Computing Science University of Alberta Edmonton, Alberta, CANADA T6H 5B5 Abstract This paper relates the computational power of Fahlman' s Recurrent Cascade Correlation (RCC) architecture to that of fInite state automata (FSA). While some recurrent networks are FSA equivalent, RCC is not. The paper presents a theoretical analysis of the RCC architecture in the form of a proof describing a large class of FSA which cannot be realized by RCC. 1 INTRODUCTION Recurrent networks can be considered to be defmed by two components: a network architecture, and a learning rule. The former describes how a network with a given set of weights and topology computes its output values, while the latter describes how the weights (and possibly topology) of the network are updated to fIt a specifIc problem. It is possible to evaluate the computational power of a network architecture by analyzing the types of computations a network could perform assuming appropriate connection weights (and topology). This type of analysis provides an upper bound on what a network can be expected to learn, since no system can learn what it cannot represent. Many recurrent network architectures have been proven to be fInite state automaton or even Turing machine equivalent (see for example [Alon, 1991], [Goudreau, 1994], [Kremer, 1995], and [Siegelmann, 1992]). The existence of such equivalence proofs naturally gives confIdence in the use of the given architectures. This paper relates the computational power of Fahlman's Recurrent Cascade Correlation architecture [Fahlman, 1991] to that of fInite state automata. It is organized as follows: Section 2 reviews the RCC architecture as proposed by Fahlman. Section 3 describes fInite state automata in general and presents some specifIc automata which will play an important role in the discussions which follow. Section 4 describes previous work by other Finite State Automata that Recurrent Cascade-Correlation Cannot Represent 613 authors evaluating RCC' s computational power. Section 5 expands upon the previous work, and presents a new class of automata which cannot be represented by RCC. Section 6 further expands the result of the previous section to identify an infinite number of other unrealizable classes of automata. Section 7 contains some concluding remarks. 2 THE RCC ARCHITECTURE The RCC architecture consists of three types of units: input units, hidden units and output units. After training, a RCC network performs the following computation: First, the activation values of the hidden units are initialized to zero. Second, the input unit activation values are initialized based upon the input signal to the network. Third, each hidden unit computes its new activation value. Fourth, the output units compute their new activations. Then, steps two through four are repeated for each new input signal. The third step of the computation, computing the activation value of a hidden unit, is accomplished according to the formula: a(t+l) = a( t W.a(t+l) + w.a(t)]. J ;=\ 'J' JJ J Here, ai(t) represents the activation value of unit i at time t, a(e) represents a sigmoid squashing function with fInite range (usually from 0 to 1), and Wij represents the weight of the connection from unit ito unitj. That is, each unit computes its activation value by mUltiplying the new activations of all lowered numbered units and its own previous activation by a set of weights, summing these products, and passing the sum through a logistic activation function. The recurrent weight Wjj from a unit to itself functions as a sort of memory by transmitting a modulated version of the unit's old activation value. The output units of the RCC architecture can be viewed as special cases of hidden units which have weights of value zero for all connections originating from other output units. This interpretation implies that any restrictions on the computational powers of general hidden units will also apply to the output units. For this reason, we shall concern ourselves exclusively with hidden units in the discussions which follow. Finally, it should be noted that since this paper is about the representational power of the RCC architecture, its associated learning rule will not be discussed here. The reader wishing to know more about the learning rule, or requiring a more detailed description of the operation of the RCC architecture, is referred to [Fahlman, 1991]. 3 FINITE STATE AUTOMATA A Finite State Automaton (FSA) [Hopcroft, 1979] is a formal computing machine defmed by a 5-tuple M=(Q,r.,8,qo,F), where Q represents a fmite set of states, r. a fmite input alphabet, 8 a state transition function mapping Qxr. to Q, qoEQ the initial state, and FcQ a set of fmal or accepting states. FSA accept or reject strings of input symbols according to the following computation: First, the FSA' s current state is initialized to qo. Second, the next inut symbol of the str ing, selected from r., is presented to the automaton by the outside world. Third, the transition function, 8, is used to compute the FSA' s new state based upon the input symbol, and the FSA's previous state. Fourth, the acceptability of the string is computed by comparing the current FSA state to the set of valid fmal states, F. If the current state is a member of F then the automaton is said to accept the string of input symbols presented so far. Steps two through four are repeated for each input symbol presented by the outside world. Note that the steps of this computation mirror the steps of an RCC network's computation as described above. It is often useful to describe specifIc automata by means of a transition diagram [Hopcroft, 1979]. Figure 1 depicts the transition diagrams of fIve FSA. In each case, the states, Q, 614 S.C. KREMER are depicted by circles, while the transitions defmed by 0 are represented as arrows from the old state to the new state labelled with the appropriate input symbol. The arrow labelled "Start" indicates the initial state, qo; and fmal accepting states are indicated by double circles. We now defme some terms describing particular FSA which we will require for the following proof. The first concerns input signals which oscillate. Intuitively, the input signal to a FSA oscillates if every pm symbol is repeated for p> 1. More formally, a sequence of input symbols, s(t), s(t+ 1), s(t+ 2), ... , oscillates with a period of p if and only if p is the minimum value such that: Vt s(t)=s(t+p). Our second definition concerns oscillations of a FSA's internal state, when the machine is presented a certain sequence of input signals. Intuitively, a FSA' s internal state can oscillate in response to a given input sequence if there is some starting state for which every subsequent <.>th state is repeated. Formally, a FSA' s state can oscillate with a period of <.> in response to a sequence of input symbols, s(t), s(t+ 1), s(t+2), ... , if and only if <.> is the minimum value for which: 3qo S.t. Vt o(qo, s(t» = o( .. , o( o( o(qo, s(t», s(t+ 1», s(t+2», ... , s(t+<.>)) The recursive nature of this formulation is based on the fact a FSA' s state depends on its previous state, which in tum depends on the state before, etc .. We can now apply these two defmitions to the FSA displayed in Figure 1. The automaton labelled "a)" has a state which oscillates with a period of <.>=2 in response to any sequence consisting of Os and Is (e.g. "00000 ... ", "11111.. .. ", "010101. .. ", etc.). Thus, we can say that it has a state cycle of period <.>=2 (Le. qoqtqoqt ... ), when its input cycles with a period of p= 1 (Le. "0000 ... If). Similarly, when automaton b)'s input cycles with period p= 1 (Le. ''000000 ... "), its state will cycle with period <.>=3 (Le. qOqtq2qOqtq2' .. ). For automaton c), things are somewhat more complicated. When the input is the sequence "0000 .. . ", the state sequence will either be qoqo%qo .. ' or fA fA fA fA .. . depending on the initial state. On the other hand, when the input is the sequence "1111 ... ", the state sequence will alternate between qo and qt. Thus, we say that automaton c) has a state cycle of <.> = 2 when its input cycles with period p = 1. But, this automaton can also have larger state cycles. For example, when the input oscillates with a period p=2 (Le. "01010101. .. If), then the state of the automaton will oscillate with a period <.>=4 (Le. qoqoqtqtqoqoqtqt ... ). Thus, we can also say that automaton c) has a state cycle of <.>=4 when its input cycles with period p =2. The remaining automata also have state cycles for various input cycles, but will not be discussed in detail. The importance of the relationship between input period (P) and the state period (<.» will become clear shortly. 4 PREVIOUS RESULTS CONCERNING THE COMPUTATIONAL POWEROFRCC The first investigation into the computational powers of RCC was performed by Giles et. al. [Giles, 1995]. These authors proved that the RCC architecture, regardless of connection weights and number of hidden units, is incapable of representing any FSA which "for the same input has an output period greater than 2" (p. 7). Using our oscillation defmitions above, we can re-express this result as: if a FSA' s input oscillates with a period of p= 1 (Le. input is constant), then its state can oscillate with a period of at most <.>=2. As already noted, Figure Ib) represents a FSA whose state oscillates with a period of <.>=3 in response to an input which oscillates with a period of p=1. Thus, Giles et. al.'s theorem proves that the automaton in Figure Ib) cannot be implemented (and hence learned) by a RCC network. Finite State Automata that Recurrent Cascade-Correlation Cannot Represent 615 a) Start b) Start 0, I c) Start d) Start o o o o o e) Start Figure I: Finite State Automata. Giles et. al. also examined the automata depicted in Figures la) and lc). However, unlike the formal result concerning FSA b), the authors' conclusions about these two automata were of an empirical nature. In particular, the authors noted that while automata which oscillated with a period of 2 under constant input (Le. Figure la» were realizable, the automaton of Ic) appeared not be be realizable by RCC. Giles et. al. could not account for this last observation by a formal proof. 616 S.C.KREMER 5 AUTOMATA WITH CYCLES UNDER ALTERNATING INPUT We now turn our attention to the question: why is a RCC network unable to learn the automaton of lc)? We answer this question by considering what would happen if lc) were realizable. In particular, suppose that the input units of a RCC network which implements automaton lc) are replaced by the hidden units of a RCC network implementing la). In this situation, the hidden units of la) will oscillate with a period of 2 under constant input. But if the inputs to lc) oscillate with a period of 2, then the state of Ic) will oscillate with a period of 4. Thus, the combined network's state would oscillate with a period of four under constant input. Furthermore, the cascaded connectivity scheme of the RCC architecture implies that a network constructed by treating one network's hidden units as the input units of another, would not violate any of the connectivity constraints of RCC. In other words, if RCC could implement the automaton of lc), then it would also be able to implement a network which oscillates with a period of 4 under constant input. Since Giles et. al. proved that the latter cannot be the case, it must also be the case that RCC cannot implement the automaton of lc). The line of reasoning used here to prove that the FSA of Figure lc) is unrealizable can also be applied to many other automata. In fact, any automaton whose state oscillates with a period of more than 2 under input which oscillates with a period 2, could be used to construct one of the automata proven to be illegal by Giles. This implies that RCC cannot implement any automaton whose state oscillates with a period of greater than <.>=2 when its input oscillates with a period of p=2. 6 AUTOMATA WITH CYCLES UNDER OSCILLATING INPUT Giles et. aI.' s theorem can be viewed as defining a class of automata which cannot be implemented by the RCC architecture. The proof in Section 5 adds another class of automata which also cannot be realized. More precisely, the two proofs concern inputs which oscillate with periods of one and two respectively. It is natural to ask whether further proofs for state cycles can be developed when the input oscillates with a period of greater than two. We now present the central theorem of this paper, a unified defmition of unrealizable automata: Theorem: If the input signal to a RCC network oscillates with a period, p, then the network can represent only those FSA whose outputs form cycles of length <.>, where pmod<.>=O if p is even and 2pmod<.> =0 if p is odd. To prove this theorem we will first need to prove a simpler one relating the rate of oscillation of the input signal to one node in an RCC network to the rate of oscillation of that node's output signal. By "the input signal to one node" we mean the weighted sum of all activations of all connected nodes (Le. all input nodes, and all lower numbered hidden nodes), but not the recurrent signal. I. e . : j - I A(t+ 1) == " W .. a .(t+ 1) . L.J IJ I 1=1 Using this defmition, it is possible to rewrite the equation to compute the activation of node j (given in Section 2) as: ap+l) == a( A(t+l)+Wha/t) ) . But if we assume that the input signal oscillates with a period of p, then every value of A(t+ 1) can be replaced by one of a fmite number of input signals (.to, AI, A 2, .,. Ap. I ) . In other words, A(t+ 1) = Atmodp ' Using this substitution, it is possible to repeatedly expand the addend of the previous equation to derive the formula: ap+ 1) = a( Atmodp + '")j . a( A(t-I)modp + Wp . a( A(t-2)modp + '")j .... a( A(t-p+I)modp + ,")/ait-p+ 1) ) ... ) ) ) Finite State Automata that Recurrent Cascade-Correlation Cannot Represent 617 The unravelling of the recursive equation now allows us to examine the relationship between ap+ 1) and t;(t-p+ 1). Specifically, we note that if ~ >0 or if p is even then aj{t+ 1) = ft.ap-p+ 1» implies that/is a monotonically increasing function. Furthermore, since 0' is a function with finite range,f must also have finite range. It is well known that for any monotonically increasing function with [mite range, /, the sequence, ft.x), fif(x» , fift.j{x») , ... , is guaranteed to monotonically approach a fixed point (whereft.x)=x). This implies that the sequence, ap+l), t;(t+p+l), q(t+2p+l), ... , must also monotonically approach a fixed point (where ap+ 1) = q.(t-p+ 1». In other words, the sequence does not oscillate. Since every prh value of ~{t) approaches a fixed point, the sequence ap), ap+ 1), ap+2), '" can have a period of at most p, and must have a period which divides p evenly. We state this as our first lemma: Lemma 1: If A.(t) oscillates with even period, p, or if Wu > 0, then state unit j's activation value must oscillate with a period c..>, where pmodc..> =0. We must now consider the case where '"11 < 0 and p is odd. In this case, ap+ 1) = ft.ap-p+ 1» implies that/is a monotonically decreasing function. But, in this situation the function/ 2(x)=ft.f{x» must be monotonically increasing with finite range. This implies that the sequence: ap+ 1), a;<t+2p+ 1), a;<t+4p+ 1), ... , must monotonically approach a fixed point (where a;<t+ 1)=ap-2p+ 1». This in turn implies that the sequence ap), ap+ 1), ap+2), ... , can have a period of at most 2p, and must have a period which divides 2p evenly. Once again, we state this result in a lemma: Lemma 2: If A.(t) oscillates with odd period p, and if Wii<O, then state unit j must oscillate with a period c..>, where 2pmodc..>=0. Lemmas 1 and 2 relate the rate of oscillation of the weighted sum of input signals and lower numbered unit activations, A.(t) to that of unitj. However, the theorem which we wish to prove relates the rate of oscillation of only the RCC network's input signal to the entire hidden unit activations. To prove the theorem, we use a proof by induction on the unit number, i: Basis: Node i= 1 is connected only to the network inputs. Therefore, if the input signal oscillates with period p, then node i can only oscillate with period c..>, where pmodc..> =0 if P is even and 2pmodc..> =0 if P is odd. (This follows from Lemmas 1 and 2). Assumption: If the input signal to the network oscillates with period p, then node i can only oscillate with period c..>, where pmodc..> =0 if p is even and 2pmodc..>=0 if p is odd. Proof: If the Assumption holds for all nodes i, then Lemmas 1 and 2 imply that it must also hold for node i+ 1.0 This proves the theorem: Theorem: If the input signal to a RCC network oscillates with a period, p, then the network can represent only those FSA whose outputs form cycles of length c..>, where pmodc..>=O ifp is even and 2pmodc..> =0 ifp is odd. 7 CONCLUSIONS It is interesting to note that both Giles et. al. 's original proof and the constructive proof by contradiction described in Section 5 are special cases of the theorem. Specifically, Giles et. al. I S original proof concerns input cycles of length p = 1. Applying the theorem of Section 6 proves that an RCC network can only represent those FSA whose state transitions form cycles of length c..>, where 2(I)modc..>=0, implying that state cannot oscillate with a period of greater than 2. This is exactly what Giles et. al concluded, and proves that (among others) the automaton of Figure Ib) cannot be implemented by RCC. 618 S.C.KREMER Similarly, the proof of Section 5 concerns input cycles of length p=2. Applying our theorem proves that an RCC network can only represent those machines whose state transitions form cycles of length <.>, where (2)modw=O. This again implies that state cannot oscillate with a period greater than 2, which is exactly what was proven in Section 5. This proves that the automaton of Figure lc) (among others) cannot be implemented by RCC. In addition to unifying both the results of Giles et. al. and Section 5, the theorem of Section 6 also accounts for many other FSA which are not representable by RCC. In fact, the theorem identifies an inflnite number of other classes of non-representable FSA (for p = 3, P =4, P = 5, ... ). Each class itself of course contains an infinite number of machines. Careful examination of the automaton illustrated in Figure ld) reveals that it contains a state cycle of length 9 (QcIJ.IQ2QIQ2Q3Q2Q3Q4QcIJ.IQ2Qlq2q3q2q3q4"') in response to an input cycle of length 3 ("001001... "). Since this is not one of the allowable input/state cycle relationships defined by the theorem, it can be concluded that the automaton of Figure Id) (among others) cannot be represented by RCC. Finally, it should be noted that it remains unknown if the classes identified by this paper IS theorem represent the complete extent of RCC's computational limitations. Consider for example the automaton of Figure Ie). This device has no input/state cycles which violate the theorem, thus we cannot conclude that it is unrepresentable by RCC. Of course, the issue of whether or not this particular automaton is representable is of little interest. However, the class of automata to which the theorem does not apply, which includes automaton Ie), requires further investigation. Perhaps all automata in this class are representable; perhaps there are other subclasses (not identified by the theorem) which RCC cannot represent. This issue will be addressed in future work. References N. Alon, A. Dewdney, and T. Ott, Efficient simulation of flnite automata by neural nets, Journal of the Association for Computing Machinery, 38 (2) (1991) 495-514. S. Fahlman, The recurrent cascade-correlation architecture, in: R. Lippmann, J. Moody and D. Touretzky, Eds., Advances in Neural Information Processing Systems 3 (Morgan Kaufmann, San Mateo, CA, 1991) 190-196. C.L. Giles, D. Chen, G.Z. Sun, H.H. Chen, Y.C. Lee, and M.W. Goudreau, Constructive Learning of Recurrent Neural Networks: Limitations of Recurrent Cascade Correlation and a Simple Solution, IEEE Transactions on Neural Networks, 6 (4) (1995) 829-836. M. Goudreau, C. Giles, S. Chakradhar, and D. Chen, First-order v.S. second-order single layer recurrent neural networks, IEEE Transactions on Neural Networks, 5 (3) (1994) 511513. J.E. Hopcroft and J.D. Ullman, Introduction to Automata Theory, Languages and Computation (Addison-Wesley, Reading, MA, 1979). S.C. Kremer, On the Computational Power of Elman-style Recurrent Networks, IEEE Transactions on Neural Networks, 6 (4) (1995) 1000-1004. H.T. Siegelmann and E.D. Sontag, On the Computational Power of Neural Nets, in: Proceedings of the Fifth ACM Workshop on Computational Learning Theory, (ACM, New York, NY, 1992) 440-449.
|
1995
|
121
|
1,023
|
Memory-based Stochastic Optimization Andrew W. Moore and Jeff Schneider School of Computer Science Carnegie-Mellon University Pittsburgh, PA 15213 Abstract In this paper we introduce new algorithms for optimizing noisy plants in which each experiment is very expensive. The algorithms build a global non-linear model of the expected output at the same time as using Bayesian linear regression analysis of locally weighted polynomial models. The local model answers queries about confidence, noise, gradient and Hessians, and use them to make automated decisions similar to those made by a practitioner of Response Surface Methodology. The global and local models are combined naturally as a locally weighted regression. We examine the question of whether the global model can really help optimization, and we extend it to the case of time-varying functions. We compare the new algorithms with a highly tuned higher-order stochastic optimization algorithm on randomly-generated functions and a simulated manufacturing task. We note significant improvements in total regret, time to converge, and final solution quality. 1 INTRODUCTION In a stochastic optimization problem, noisy samples are taken from a plant. A sample consists of a chosen control u (a vector ofreal numbers) and a noisy observed response y. y is drawn from a distribution with mean and variance that depend on u. y is assumed to be independent of previous experiments. Informally the goal is to quickly find control u to maximize the expected output E[y I u). This is different from conventional numerical optimization because the samples can be very noisy, there is no gradient information, and we usually wish to avoid ever performing badly (relative to our start state) even during optimization. Finally and importantly: each experiment is very expensive and there is ample computational time (often many minutes) for deciding on the next experiment. The following questions are both interesting and important: how should this computational time best be used, and how can the data best be used? Stochastic optimization is of real industrial importance, and indeed one of our reasons for investigating it is an association with a U.S. manufacturing company Memory-based Stochastic Optimization 1067 that has many new examples of stochastic optimization problems every year. The discrete version of this problem, in which u is chosen from a discrete set, is the well known k-armed bandit problem. Reinforcement learning researchers have recently applied bandit-like algorithms to efficiently optimize several discrete problems [Kaelbling, 1990, Greiner and Jurisica, 1992, Gratch et al., 1993, Maron and Moore, 1993]. This paper considers extensions to the continuous case in which u is a vector of reals. We anticipate useful applications here too. Continuity implies a formidable number of arms (uncountably infinite) but permits us to assume smoothness of E[y I u] as a function of u. The most popular current techniques are: • Response Surface Methods (RSM). Current RSM practice is described in the classic reference [Box and Draper, 1987]. Optimization proceeds by cautious steepest ascent hill-climbing. A region of interest (ROI) is established at a starting point and experiments are made at positions within the region that can best be used to identify the function properties with low-order polynomial regression. A large portion of the RSM literature concerns experimental design-the decision of where to take data points in order to acquire the lowest variance estimate of the local polynomial coefficients in a fixed number of experiments. When the gradient is estimated with sufficient confidence, the ROI is moved accordingly. Regression of a quadratic locates optima within the ROI and also diagnoses ridge systems and saddle points. The strength of RSM is that it is careful not to change operating conditions based on inadequate evidence, but moves once the data justifies. A weakness of RSM is that human judgment is needed: it is not an algorithm, but a manufacturing methodology . • Stochastic Approximation methods. The algorithm of [Robbins and Monro, 1951] does root finding without the use of derivative estimates. Through the use of successively smaller steps convergence is proven under broad assumptions about noise. Keifer-Wolfowitz (KW) [Kushner and Clark, 1978] is a related algorithm for optimization problems. From an initial point it estimates the gradient by performing an experiment in each direction along each dimension of the input space. Based on the estimate, it moves its experiment center and repeats. Again, use of decreasing step sizes leads to a proof of convergence to a local optimum. The strength of KW is its aggressive exploration, its simplicity, and that it comes with convergence guarantees. However, it has more of a danger of attempting wild experiments in the presence of noise, and effectively discards the data it collects after each gradient estimate is made. In practice, higher order versions of KW are available in which convergence is accelerated by replacing the fixed step size schedule with an adaptive one [Kushner and Clark, 1978]. Later we compare the performance of our algorithms to such a higher-order KW. 2 MEMORY-BASED OPTIMIZATION Neither KW nor RSM uses old data. After a gradient has been identified the control u is moved up the gradient and the data that produced the gradient estimate is discarded. Does this lead to inefficiencies in operation? This paper investigates one way of using old data: build a global non-linear plant model with it. We use locally weighted regression to model the system [Cleveland and Delvin, 1988, Atkeson, 1989, Moore, 1992]. We have adapted the methods to return posterior distributions for their coefficients and noise (and thus, indirectly, their predictions) 1068 A. W. MOORE, J. SCHNEIDER based on very broad priors, following the Bayesian methods for global linear regression described in [DeGroot, 1970]. We estimate the coefficients f3 = {,8I ... ,8m} of a local polynomial model in which the data was generated by the polynomial and corrupted with gaussian noise of variance u2, which we also estimate. Our prior assumption will be that f3 is distributed according to a multivariate gaussian of mean 0 and covariance matrix E. Our prior on u is that 1/u2 has a gamma distribution with parameters a and ,8. Assume we have observed n pieces of data. The jth polynomial term for the ith data point is Xij and the output response of the ith data point is Ii. Assume further that we wish to estimate the model local to the query point X q , in which a data point at distance di from the the query point has weight Wi = exp( -dl! K). K, the kernel width is a fixed parameter that determines the degree of localness in the local regression. Let W = Diag(wl,w2 . .. Wn). The marginal posterior distribution of f3 is' a t distribution with mean 13 = (E- 1 + X T W 2X)-1(XT W 2y) covariance (2,8 + (yT - f3T XT)W2yT)(E-l + X TW 2 X)-l / (2a + I:~=l wi) (1) and a + I:~=l w'f degrees of freedom. We assume a wide, weak, prior E = Diag(202,202, ... 202), a = 0.8,,8 = 0.001, meaning the prior assumes each regression coefficient independently lies with high probability in the range -20 to 20, and the noise lies in the range 0.01 to 0.5. Briefly, we note the following reasons that Bayesian locally weighted polynomial regression is particularly suited to this application: • We can directly obtain meaningful confidence estimates of the joint pdf of the regressed coefficients and predictions. Indirectly, we can compute the probability distribution of the steepest gradient, the location of local optima and the principal components of the local Hessian. • The Bayesian approach allows meaningful regressions even with fewer data points than regression coefficients-the posterior distribution reveals enormous lack of confidence in some aspects of such a model but other useful aspects can still be predicted with confidence. This is crucial in high dimensions, where it may be more effective to head in a known positive gradient without waiting for all the experiments that would be needed for a precise estimate of steepest gradient. • Other pros and cons of locally weighted regression in the context of control can be found in [Moore et ai., 1995]. Given the ability to derive a plant model from data, how should it best be used? The true optimal answer, which requires solving an infinite-dimensional Markov decision process, is intractable. We have developed four approximate algorithms that use the learned model, described briefly below. • AutoRSM. Fully automates the (normally manual) RSM procedure and incorporates weighted data from the model; not only from the current design. It uses online experimental design to pick ROI design points to maximize information about local gradients and optima. Space does not permit description of the linear algebraic formulations of these questions. • PMAX. This is a greedy, simpler approach that uses the global non-linear model from the data to jump immediately to the model optimum. This is similar to the technique described in [Botros, 1994], with two extensions. First, the Bayesian Memory-based Stochastic Optimization 1069 Figure 1: Three examples of 2-d functions used in optimization experiments priors enable useful decisions before the regression becomes full-rank. Second, local quadratic models permit second-order convergence near an optimum. • IEMAX. Applies Kaelbling's IE algorithm [Kaelbling, 1990] in the continuous case using Bayesian confidence intervals. argmax; () llchosen = u J opt U (2) where iopt(u) is the top of the 95th %-ile confidence interval. The intuition here is that we are encouraged to explore more aggressively than PMAX, but will not explore areas that are confidently below the best known optimum . • COMAX. In a real plant we would never want to apply PMAX or IEMAX. Experiments must be cautious for reasons of safety, quality control, and managerial peace of mind. COMAX extends IEMAX thus: argmax A A . llchosen = fopt(u);U E SAFE{=} f,pess(U) > dIsaster threshold (3) u E SAFE Analysis of these algorithms is problematic unless we are prepared to make strong assumptions about the form of E[Y I u]. To examine the general case we rely on Monte Carlo simulations, which we now describe. The experiments used randomly generated nonlinear unimodal (but not necessarily convex) d-dimensional functions from [0, l]d -+ [0,1]. Figure 1 shows three example 2-d functions. Gaussian noise (0- = 0.1) is added to the functions. This is large noise, and means several function evaluations would be needed to achieve a reliable gradient estimate for a system using even a large step size such as 0.2. The following optimization algorithms were tested on a sample of such functions. Vary-KW The best performing KW algorithm we could find varied step size and adapted gradient estimation steps to avoid undue regret at optima. Fixed-KW A version of KW that keeps its gradient-detecting step size fixed. This risks causing extra regret at a true optima, but has less chance of becoming delayed by a non-optimum. Auto-RSM The best performing version thereof. Passlve-RSM Auto-RSM continues to identify the precise location of the optimum when it's arrived at that optimum. When Passive-RSM is confident (greater than 99%) that it knows the location of the optimum to two significant places, it stops experimenting. Linear RSM A linear instead of quadratic model, thus restricted to steepest ascent. CRSM Auto-RSM with conservative parameters, more typical of those recommended in the RSM literature. Pmax, IEmax As described above. and Comax Figures 2a and 2b show the first sixty experiments taken by AutoRSM and KW respectively on their journeys to the goal. 1070 (a) (b) (0) RetroI_d ............ _ A. W. MOORE, J. SCHNEIDER Figure 2a: The path taken (start at (0.8,0.2)) by AutoRSM optimizing the given function with added noise of standard deviation 0.1 at each experiment. Figure 2b: The path taken (start at (0.8,0.2)) by KW. KW's path looks deceptively bad, but remember it is continually buffeted by considerable noise. te) No. of", YntII wllhln 0,05 rtf optimum let) ............ of FINAL ... tepe Figure 3: Comparing nine stochastic optimization algorithms by four criteria: (a) Regret, (b) Disasters, (c) Speed to converge (d) Quality at convergence. The partial order depicted shows which results are significant at the 99% level (using blocked pairwise comparisons). The outputs of the random functions range between 0-1 over the input domain. The numbers in the boxes are means over fifty 5-d functions. (a) Regret is defined as the mean Yopt - Yi-the cost incurred during the optimization compared with performance if we had known the optimum location and used it from the beginning. With the exception of IEMAX, model-based methods perform significantly better than KW, with reduced advantage for cautious and linear methods. (b) The %-age of steps which tried experiments with more than 0.1 units worse performance than at the search start. This matters to a risk averse manager. AutoRSM has fewer than 1% disasters, but COMAX and the modelfree methods do better still. PMAX's aggressive exploration costs it. (c) The number of steps until we reach within 0.05 units of optimal. PMAX's aggressiveness wins. (d) The quality of the "final" solution between steps 50 and 60 of the optimization. Results for 50 trials of each optimization algorithms for five-dimensional randomly generated functions are depicted in Figure 3. Many other experiments were performed in other dimensionalities and for modified versions of the algorithm, but space does not permit detailed discussion here. Finally we performed experiments with the simulated power-plant process in Figure 4. The catalyst controller adjusts the flow rate of the catalyst to achieve the goal chemical A content. Its actions also affect chemical B content. The temperature controller adjusts the reaction chamber temperature to achieve the goal chemical B content. The chemical contents are also affected by the flow rate which is determined externally by demand for the product. The task is to find the optimal values for the six controller parameters that minimize the total squared deviation from desired values of chemical A and chemical B contents. The feedback loops from sensors to controllers have significant delay. The controller gains on product demand are feedforward terms since there is significant delay in the effects of demand on the process. Finally, the performance of the system may also depend on variations over time in the composition of the input chemicals which can not be directly sensed. Memory-based Stochastic Optimization Catalyst Supply Sensor A Raw Input Chemicals Optimize 6 Controller Parameters To Minimize Squared Deviation from Goal Chemical A and B Content Te Catalyst Controller base lenns: Base temperature rccdback term,,; Sen.for B gain Product Demand f'L-_---'----'----~orwvd tem\s: Product demand gain Base input rate Sensor A gain Product demand gain REACTION CHAMBER Chemical A content sensor Pumps governed by demand for product Chemical B content sensor Product output 1071 Figure 4: A Simulated Chemical Process The total summed regrets of the optimization methods on 200 simulated steps were: Stay AtStart 10.86 FixedKW 2.82 AutoRSM 1.32 PMAX 3.30 COMAX 4.50 In this case AutoRSM is best, considerably beating the best KW algorithm we could find. In contrast PM AX and COMAX did poorly: in this plant wild experiments are very costly to PMAX and COMAX is too cautious. Stay AtStart is the regret that would be incurred if all 200 steps were taken at the initial parameter setting. 3 UNOBSERVED DISTURBANCES An apparent danger of learning a model is that if the environment changes, the out of date model will mean poor performance and very slow adaptation. The modelfree methods, which use only recent data, will react more nimbly. A simple but unsatisfactory answer to this is to use a model that implicitly (e.g. a neural net) or explicitly (e.g. local weighted regression of the fifty most recent points) forgets. An interesting possibility is to learn a model in a way that automatically determines whether a disturbance has occurred, and if so, how far back to forget. The following "adaptive forgetting" (AF) algorithm was added to the AutoRSM algorithm: At each step, use all the previous data to generate 99% confidence intervals on the output value at the current step. If the observed output is outside the intervals assume that a large change in the system has occured and forget all previous data. This algorithm is good for recognizing jumps in the plant's operating characteristics and allows AutoRSM to respond to them quickly, but is not suitable for detecting and handling process drift. We tested our algorithm's performance on the simulated plant for 450 steps. Operation began as before, but at step 150 there was an unobserved change in the composition of the raw input chemicals. The total regrets of the optimization methods were: StayAtStart FixedKW AutoRSM PMAX AutoRSM/AF 11.90 5.31 8.37 9.23 2.75 AutoRSM and PMAX do poorly because all their decisions after step 150 are based partially on the invalid data collected before then. The AF addition to AutoRSM solves the problem while beating the best KW by a factor of 2. Furthermore, AutoRSMj AF gets 1.76 on the invariant task, thus demonstrating that it can be used safely in cases where it is not known if the process is time varying. 1072 A. W. MOORE, J. SCHNEIDER 4 DISCUSSION Botros' thesis [Botros, 1994] discusses an algorithm similar to PMAX based on local linear regression. [Salganicoff and Ungar, 1995] uses a decision tree to learn a model. They use Gittins indices to suggest experiments: we believe that the memory-based methods can benefit from them too. They, however, do not use gradient information, and so require many experiments to search a 2D space. IEmax performed badly in these experiments, but optimism-gl1ided exploration may prove important in algorithms which check for potentially superior local optima. A possible extension is self tuning optimization. Part way through an optimization, to estimate the best optimization parameters for an algorithm we can run montecarlo simulations which run on sample functions from the posterior global model given the current data. This paper has examined the question of how much can learning a Bayesian memorybased model accelerate the convergence of stochastic optimization. We have proposed four algorithms for doing this, one based on an autonomous version of RSM; the other three upon greedily jumping to optima of three criteria dependent on predicted output and uncertainty. Empirically the model-based methods provide significant gains over a highly tuned higher order model-free method. References [Atkeson, 1989] C . G . Atkeson. Using Local Models to Control Movement. In Proceedings of Neural Information Processing Systems Conference, November 1989. [Botros, 1994] S. M. Botros. Model-Based Techniques in Motor Learning and Task Optimization. PhD. Thesis, MIT Dept. of Brain and Cognitive Sciences, February 1994. [Box and Draper, 1987] G. E . P. Box and N. R. Draper. Empirical Model-Building and Response Surfaces. Wiley, 1987. [Cleveland and Delvin, 1988] W. S. Cleveland and S. J . Delvin. Locally Weighted Regression: An Approach to Regression Analysis by Local Fitting. Journal of the American Statistical Association, 83(403):596-610, September 1988. [DeGroot, 1970] M. H . DeGroot. Optimal Statistical Decisions. McGraw-Hill, 1970. [Gratch et al., 1993] J. Gratch , S. Chien, and G. DeJong. Learning Search Control Knowledge for Deep Space Network Scheduling. In Proceedings of the 10th International Conference on Machine Learning. Morgan Kaufmann, June 1993. [Greiner and Jurisica, 1992] R. Greiner and I. Jurisica. A statistical approach to solving the EBL utility problem. In Proceedings of the Tenth International Conference on Artificial Intelligence (AAAI92). MIT Press, 1992. [Kaelbling, 1990] L. P. Kaelbling. Learning in Embedded Systems. PhD. Thesis; Technical Report No. TR-90-04, Stanford University, Department of Computer Science, June 1990. [Kushner and Clark, 1978] H. Kushner and D. Clark. Stochastic Approximation Methods for Constrained and Unconstrained Systems. Springer-Verlag, 1978. [Maron and Moore, 1993] O. Maron and A. Moore. Hoeffding Races: Accelerating Model Selection Search for Classification and Function Approximation. In Advances in Neural Information Processing Systems 6. Morgan Kaufmann, December 1993. [Moore et al., 1995] A . W . Moore, C. G . Atkeson, and S. Schaal. Memory-based Learning for Control. Technical report, CMU Robotics Institute, Technical Report CMU-RI-TR-95-18 (Submitted for Publication), 1995. [Moore, 1992] A. W . Moore. Fast, Robust Adaptive Control by Learning only Forward Models. In J . E . Moody, S. J . Hanson , and R . P. Lippman, editors, Advances in Neural Information Processing Systems 4. Morgan Kaufmann, April 1992. [Robbins and Monro, 1951] H . Robbins and S. Monro. A stochastic approximation method. Annals of Mathematical Statist2cs, 22:400-407, 1951. [Salganicoff and Ungar, 1995] M. Salganicoffand L. H. Ungar. Active Exploration and Learning in RealValued Spaces using Multi-Armed Bandit Allocation Indices. In Proceedings of the 12th International Conference on Machine Learning. Morgan Kaufmann, 1995.
|
1995
|
122
|
1,024
|
Handwritten Word Recognition using Contextual Hybrid Radial Basis Function NetworklHidden Markov Models Bernard Lemarie La Poste/SRTP 10, Rue de l'lle-Mabon F-44063 Nantes Cedex France lemarie@srtp.srt-poste.fr Michel Gilloux La Poste/SRTP 10, Rue de l'1le-Mabon F-44063 Nantes Cede x France gilloux@srtp.srt-poste.fr Manuel Leroux La Poste/SRTP 10, Rue de l'lle-Mabon F-44063 Nantes Cedex France leroux@srtp.srt-poste.fr Abstract A hybrid and contextual radial basis function networklhidden Markov model off-line handwritten word recognition system is presented. The task assigned to the radial basis function networks is the estimation of emission probabilities associated to Markov states. The model is contextual because the estimation of emission probabilities takes into account the left context of the current image segment as represented by its predecessor in the sequence. The new system does not outperform the previous system without context but acts differently. 1 INTRODUCTION Hidden Markov models (HMMs) are now commonly used in off-line recognition of handwritten words (Chen et aI., 1994) (Gilloux et aI., 1993) (Gilloux et al. 1995a). In some of these approaches (Gilloux et al. 1993), word images are transformed into sequences of image segments through some explicit segmentation procedure. These segments are passed on to a module which is in charge of estimating the probability for each segment to appear when the corresponding hidden state is some state s (state emission probabilities). Model probabilities are generally optimized for the Maximum Likelihood Estimation (MLE) criterion. MLE training is known to be sub-optimal with respect to discrimination ability when the underlying model is not the true model for the data. Moreover, estimating the emission probabilities in regions where examples are sparse is difficult and estimations may not be accurate. To reduce the risk of over training, images segments consisting of bitmaps are often replaced by feature vector of reasonable length (Chen et aI., 1994) or even discrete symbols (Gilloux et aI., 1993). Handwritten Word Recognition Using HMMlRBF Networks 765 In a previous paper (Gilloux et aI., 1995b) we described a hybrid HMMlradial basis function system in which emission probabilities are computed from full-fledged bitmaps though the use of a radial basis function (RBF) neural network. This system demonstrated better recognition rates than a previous one based on symbolic features (Gilloux et aI., 1995b). Yet, many misclassification examples showed that some of the simplifying assumptions made in HMMs were responsible for a significant part of these errors. In particular, we observed that considering each segment independently from its neighbours would hurt the accuracy of the model. For example, figure 1 shows examples of letter a when it is segmented in two parts. The two parts are obviously correlated. al Figure 1: Examples of segmented a. We propose a new variant of the hybrid HMMIRBF model in which emission probabilities are estimated by taking into account the context of the current segment. The context will be represented by the preceding image segment in the sequence. The RBF model was chosen because it was proven to be an efficient model for recognizing isolated digits or letters (Poggio & Girosi, 1990) (Lemarie, 1993). Interestingly enough, RBFs bear close relationships with gaussian mixtures often used to model emission probabilities in markovian contexts. Their advantage lies in the fact that they do not directly estimate emission probabilities and thus are less prone to errors in this estimation in sparse regions. They are also trained through the Mean Square Error (MSE) criterion which makes them more discriminant. The idea of using a neural net and in particular a RBF in conjunction with a HMM is not new. In (Singer & Lippman, 1992) it was applied to a speech recognition task. The use of context to improve emission probabilities was proposed in (Bourlard & Morgan, 1993) with the use of a discrete set of context events. Several neural networks are there used to estimate various relations between states, context events and current segment. Our point is to propose a different method without discrete context based on a adapted decomposition of the HMM likelihood estimation.This model is next applied to off-line handwritten word recognition. The organization of this paper is as follows. Section 1 is an overview of the architecture of our HMM. Section 2 describes the justification for using RBF outputs in a contextual hidden Markov model. Section 3 describes the radial basis function network recognizer. Section 4 reports on an experiment in which the contextual model is applied to the recognition of handwritten words found on french bank or postal cheques. 2 OVERVIEW OF THE HIDDEN MARKOV MODEL In an HMM model (Bahl et aI., 1983), the recognition scores associated to words ware likelihoods L(wli) ... in) = P(i1 ···inlw)xP(W) in which the first term in the product encodes the probability with which the model of each word w generates some image (some sequence of image segments) ij ... in- In the HMM paradigm, this term is decomposed into a sum on all paths (i.e. sequence of hidden states) of products of the probability of the hidden path by the probability that the path generated the Image sequence: p(i) ... inlw) = 766 B. LEMARIE. M. GILLOUX. M. LEROUX It is often assumed that only one path contributes significantly to this term so that In HMMs, each sequence element is assumed to depend only on its corresponding state: n p(il···i ISI···s ) = ITp(i·ls .) n n } } j=1 Moreover, first-order Markov models assume that paths are generated by a first-order Markov chain so that n P(sl ···s ) = ITp(s . ls. I) n } }j = I We have reported in previous papers (Gilloux et aI., 1993) (Gilloux et aI., 1995a) on several handwriting recognition systems based on this assumption.The hidden Markov model architecture used in all systems has been extensively presented in (Gilloux et aI., 1995a). In that model, word letters are associated to three-states models which are designed to account for the situations where a letter is realized as 0, 1 or 2 segments. Word models are the result of assembling the corresponding letter models. This architecture is depicted on figure 2. We used here transition emission rather than state emission. However, this does not E, 0.05 E,0.05 I a val Figure 2: Outline of the model for "laval". change the previous formulas if we replace states by transitions, i.e. pairs of states. One of these systems was an hybrid RBFIHMM model in which a radial basis function network was used to estimate emission probabilities p (i. Is.) . The RBF outputs are introduced by applying Bayes rule in the expression of p (i I .~. i; I s I ... S n) : n p(s.1 i.) xp(i.) p(il ···i IsI· ··s) = IT }} } n n. p (s.) } = 1 } Since the product of a priori image segments probabilities p (i.) does not depend on the word hypothesis w, we may write: } n p (s. Ii.) p(il···inlsl···sn)oc.IT p~s./ } = 1 } In the above formula, terms of form p (s . Is. _ I) are transition probabilities which may be estimated through the Baum-Welch re-istirhatlOn algorithm. Terms of form p (s.) are a priori probabilities of states. Note that for Bayes rule to apply, these probabilitid have and only have to be estimated consistently with terms of form p (s. Ii.) since p (i. Is.) is independent of the statistical distribution of states. } } } } It has been proven elsewhere (Richard & Lippman, 1992) that systems trained through the MSE criterion tend to approximate Bayes probabilities in the sense that Bayes probaHandwritten Word Recognition Using HMMlRBF Networks 767 bilities are optimal for the MSE criterion. In practice, the way in which a given system comes close to Bayes optimum is not easily predictable due to various biases of the trained system (initial parameters, local optimum, architecture of the net, etc.). Thus real output scores are generally not equal to Bayes probabilities. However, there exist different procedures which act as a post-processor for outputs of a system trained with the MSE and make them closer to Bayes probabilities (Singer & Lippman, 1992). Provided that such a postprocessor is used, we will assume that terms p (s. Ii.) are well estimated by the post-processed outputs of the recognition system. Then, u~~ p (s .) are just the a priori probabilities of states on the set used to train the system or post-prbcess the system outputs. This hybrid handwritten word recognition system demonstrated better performances than previous systems in which word images were represented through sequences of symbolic features instead of full-fledged bitmaps (Gilloux et aI., 1995b). However, some recognition errors remained, many of which could be explained by the simplifying assumptions made in the model. In particular, the fact that emission probabilities depend only on the state corresponding to the current bitmap appeared to be a poor choice. For example, on figure 3 the third and fourth segment are classified as two halves of the letter i. For letters Figure 3: An image of trente classified as mille. segmented in two parts, the second half is naturally correlated to the first (see figure 1). Yet, our Markov model architecture is designed so that both halves are assumed uncorrelated. This has two effects. Two consecutive bitmaps which cannot be the two parts of a unique letter are sometimes recognized as such like on figure 3. Also, the emission probability of the second part of a segmented letter is lower than if the first part has been considered for estimating this probability. The contextual model described in the next section is designed so has to make a different assumption on emission probabilities. 3 THE HYBRID CONTEXTUAL RBFIHMM MODEL The exact decomposition of the emission part of word likelihoods is the following: n p(i1 ···inls1···sn) = P(il ls 1··· sn) x ITp(ijlsl ... sn,il ... ij_l) j=2 We assume now that bitmaps are conditioned by their state and the previous image in the sequence: n P(il··· in I sl· ·· sn) ==p(i11 sl) x IT p (ij I sj'ij _ l ) j=2 The RBF is again introduced by applying Bayes rule in the following way: P(s1 1 il ) xp(i l ) n p(s . 1 i ., i . 1) xp (i . I i . 1) p (i 1··· in lSI·· · s n) == () x IT }}} (I · /) P sl . P s. !. 1 J=2 J JSince terms of form p (i . Ii . _ 1) do not contribute to the discrimination of word hypotheses, we may write: J J p (s 1 IiI) n p (s . Ii., i . 1 ) ( . . I) IT } J JP 11· · ·ln sl··· sn oc () x I.) pSI . P (s . I. 1 J=2 J J768 B. LEMARIE, M. GILLOUX, M. LEROUX The RBF has now to estimate not only terms of form p (s. Ii ., i. _ 1) but also terms like p (s . Ii. 1) which are no longer computed by mere countind. 'two radial basis function netJoris-are then used to estimate these probabilities. Their common architecture is described in the next section. 4 THE RADIAL BASIS FUNCTION MODEL The radial basis function model has been described in (Lemarie, 1993). RBF networks are inspired from the theory of regularization (Poggio & Girosi, 1990). This theory studies how multivariate real functions known on a finite set of points may be approximated at these points in a family of parametric functions under some bias of regularity. It has been shown that when this bias tends to select smooth functions in the sense that some linear combination of their derivatives is minimum, there exist an analytical solution which is a linear combination of gaussians centred on the points where the function is known (Poggio & Girosi, 1990). It is straightforward to transpose this paradigm to the problem of learning probability distributions given a set of examples. In practice, the theory is not tractable since it requires one gaussian per example in the training set. Empirical methods (Lemarie, 1993) have been developed which reduce the number of gaussian centres. Since the theory is no longer applicable when the number of centres is reduced, the parameters of the model (centres and covariance matrices for gaussians, weights for the linear combination) have to be trained by another method, in that case the gradient descent method and the MSE criterion. Finally, the resulting RBF model may be looked at like a particular neural network with three layers. The first is the input layer. The second layer is completely connected to the input layer through connections with unit weights. The transfer functions of cells in the second layer are gaussians applied to the weighed distance between the corresponding centres and the weighed input to the cell. The weight of the distance are analogous to the parameters of a diagonal covariance matrix. Finally, the last layer is completely connected to the second one through weighted connections. Cells in this layer just output the sum of their input. In our experiments, inputs to the RBF are feature vectors of length 138 computed from the bitmaps of a word segment (Lemarie, 1993). The RBF that estimates terms of form p (s. Ii., i. 1) uses to such vectors as input whereas the second RBF (terms p (/ I/_il) ) is only fed with the vector associated to ij _l . These vectors are inspired from "cha'rac{eristic loci" methods (Gluksman, 1967) and encode the proportion of white pixels from which a bitmap border can be reached without meeting any black pixel in various of directions. 5 EXPERIMENTS The model has been assessed by applying it to the recognition of words appearing in legal amounts of french postal or bank cheques. The size of the vocabulary is 30 and its perplexity is only 14.3 (Bahl et aI., 1983). The training and test bases are made of images of amount words written by unknown writers on real cheques. We used 7 191 images during training and 2 879 different images for test. The image resolution was 300 dpi. The amounts were manually segmented into words and an automatic procedure was used to separate the words from the preprinted lines of the cheque form. The training was conducted by using the results of the former hybrid system. The segmentation module was kept unchanged. There are 48 140 segments in the training set and 19577 in the test set. We assumed that the base system is almost always correct when aligning segments onto letter models. We thus used this alignment to label all the segments in the training set and took these labels as the desired outputs for the RBF. We used a set of 63 different labels since 21 letters appear in the amount vocabulary and 3 types of segments are possible for each letter. The outputs of the RBF are directly interpreted as Bayes probHandwritten Word Recognition Using HMMJRBF Networks 769 abilities without further post-processing. First of all, we assessed the quality of the system by evaluating its ability to recognize the class of a segment through the value of p (s . Ii., i. 1) and compared it with that of the previous hybrid system. The results of this e'xpdrirhent are reported on table 1 for the test set. They demonstrate the importance of the context and thus its potential interest for a Table 1: Recognition and confusion rates for segment classifiers Recognition rate Confusion rate Mean square error RBF system without context 32.6% 67.4% 0.828 RBF system with context 41.7% 58.3% 0.739 word recognition system. We next compare the performance on word recognition on the data base of 2878 images of words. Results are shown in table 2. The first remark is that the system without context Table 2: Recognition and confusion rates for the word recognition systems Recognition rate Confusion rate # Confusions RBF system without context 81,3% 16,7% 536 RBF system with context 76,3% 23,7% 683 present better results than the contextual system. Some of the difference between the systems with and without context are shown below in figures 4 and 5 and may explain why the contextual system remains at a lower level of performance. The word "huit" and "deux" of figure 4 are well recognized by the system without context but badly identified by the contextual system respectively as "trente" and "franc". The image of the word "huit", for example, is segmented into eight segments and each of the four letters of the word is thus necessarily considered as separated in two parts. The fifth and sixth segments are thus recognized as two halves of the letter "i" by the standard system while the contextual system avoids this decomposition of the letter "i". On the next image, the contextual system proposes "ra" for the second and third segments mainly because of the absence of information on the relative position ofthese segments. On the other hand, figure 5 shows examples where the contextual system outperforms the system without context. In the first case the latter proposed the class "trois" with two halves on the letter "i" on the fifth and sixth segments. In the second case the context is clearly useful for the recognition on the first letter of the word. Forthcoming experiments will try to combine the two systems so as to benefit of their respective characteritics. Figure 4 : some new confusions produced by the contextual system. Experiments have also revealed that the contextual system remains very sensible to the numerical output values for the network which estimates p (s. Ii. _ 1) . Several approaches for solving this problem are currently under investigation. Ffrst'results have yet been obtained by trying to approximate the network which estimates p (Sj I ij _ 1) from the network which estimates p (Sj I ij' ij _ 1) . 770 B. LEMARIE, M. GILLOUX, M. LEROUX 6 CONCLUSION We have described a new application of a hybrid radial basis function/hidden Markov model architecture to the recognition of off-line handwritten words. In this architecture, the estimation of emission probabilities is assigned to a discriminant classifier. The estimation of emission probabilities is enhanced by taking into account the context as represented by the previous bitmap in the sequence to be classified. A formula have been derived introducing this context in the estimation of the likelihood of word scores. The ratio of the output values of two networks are now used so as to estimate the likelihood. The reported experiments reveal that the use of context, if profitable at the segment recognition level, is not yet useful at the word recognition level. Nevertheless, the new system acts differently from the previous system without context and future applications will try to exploit this difference. The dynamic of the ratio networks output values is also very unstable and some solutions to stabilize it which will be deeply tested in the forthcoming experiences. References Bahl L, Jelinek F, Mercer R, (1983). A maximum likelihood approach to speech recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence 5(2): 179-190. Bahl LR, Brown PF, de Souza PV, Mercer RL, (1986). Maximum mutual information estimation of hidden Markov model parameters for speech recognition. In: Proc of the Int Conf on Acoustics, Speech, and Signal Processing (ICASSP'86):49-52. Bourlard, H., Morgan, N., (1993). Continuous speech recognition by connectionist statistical methods, IEEE Trans. on Neural Networks, vol. 4, no. 6, pp. 893-909, 1993. Chen, M.-Y., Kundu, A., Zhou, J., (1994). Off-line handwritten word recognition using a hidden Markov model type stochastic network, IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 16, no. 5:481-496. Gilloux, M., Leroux, M., Bertille, J.-M., (1993). Strategies for handwritten words recognition using hidden Markov models, Proc. of the 2nd Int. Conf. on Document Analysis and Recognition:299-304. Gilloux, M., Leroux, M., Bertille, J.-M., (1995a). "Strategies for Cursive Script Recognition Using Hidden Markov Models", Machine Vision & Applications, Special issue on Handwriting recognition, R Plamondon ed., accepted for publication. Gilloux, M., Lemarie, B., Leroux, M., (l995b). "A Hybrid Radial Basis Function Network! Hidden Markov Model Handwritten Word Recognition System", Proc. of the 3rd Int. Conf. on Document Analysis and Recognition:394-397. Gluksman, H.A., (1967). Classification of mixed font alphabetics by characteristic loci, 1 st Annual IEEE Computer Conf.: 138-141. Lemarie, B., (1993). Practical implementation of a radial basis function network for handwritten digit recognition, Proc. of the 2nd Int. Conf. on Document Analysis and Recognition:412-415. Poggio, T., Girosi, F., (1990). Networks for approximation and learning, Proc. of the IEEE, vol 78, no 9. Richard, M.D., Lippmann, RP., (1991). "Neural network classifiers estimate bayesian a posteriori probabilities", Neural Computation, 3:461-483. Singer, E, Lippmann, RP., (1992). A speech recognizer using radial basis function networks in an HMM framework, Proc. of the Int. Conf. on acoustics, Speech, and Signal Processing.
|
1995
|
123
|
1,025
|
Universal Approximation and Learning of Trajectories Using Oscillators Pierre Baldi* Division of Biology California Institute of Technology Pasadena, CA 91125 pfbaldi@juliet.caltech.edu Kurt Hornik Technische Universitat Wien Wiedner Hauptstra8e 8-10/1071 A-1040 Wien, Austria Kurt.Hornik@tuwien.ac.at Abstract Natural and artificial neural circuits must be capable of traversing specific state space trajectories. A natural approach to this problem is to learn the relevant trajectories from examples. Unfortunately, gradient descent learning of complex trajectories in amorphous networks is unsuccessful. We suggest a possible approach where trajectories are realized by combining simple oscillators, in various modular ways. We contrast two regimes of fast and slow oscillations. In all cases, we show that banks of oscillators with bounded frequencies have universal approximation properties. Open questions are also discussed briefly. 1 INTRODUCTION: TRAJECTORY LEARNING The design of artificial neural systems, in robotics applications and others, often leads to the problem of constructing a recurrent neural network capable of producing a particular trajectory, in the state space of its visible units. Throughout evolution, biological neural systems, such as central pattern generators, have also been faced with similar challenges. A natural approach to tackle this problem is to try to "learn" the desired trajectory, for instance through a process of trial and error and subsequent optimization. Unfortunately, gradient descent learning of complex trajectories in amorphous networks is unsuccessful. Here, we suggest a possible approach where trajectories are realized, in a modular and hierarchical fashion, by combining simple oscillators. In particular, we show that banks of oscillators have universal approximation properties. * Also with the Jet Propulsion Laboratory, California Institute of Technology. 452 P. BALDI, K. HORNIK To begin with, we can restrict ourselves to the simple case of a network with one! visible linear unit and consider the problem of adjusting the network parameters in a way that the output unit activity u(t) is equal to a target function I(t), over an interval of time [0, T]. The hidden units of the network may be non-linear and satisfy, for instance, one of the usual neural network charging equations such as dUi Ui ~ dt = - Ti + L..JjWij/jUj(t Tij), (1) where Ti is the time constant of the unit, the Tij represent interaction delays, and the functions Ij are non-linear input/output functions, sigmoidal or other. In the next section, we briefly review three possible approaches for solving this problem, and some of their limitations. In particular, we suggest that complex trajectories can be synthesized by proper combination of simple oscillatory components. 2 THREE DIFFERENT APPROACHES TO TRAJECTORY LEARNING 2.1 GRADIENT DESCENT APPROACHES One obvious approach is to use a form of gradient descent for recurrent networks (see [2] for a review), such as back-propagation through time, in order to modify any adjustable parameters of the networks (time constants, delays, synaptic weights and/or gains) to reduce a certain error measure, constructed by comparing the output u(t) with its target I(t). While conceptually simple, gradient descent applied to amorphous networks is not a successful approach, except on the most simple trajectories. Although intuitively clear, the exact reasons for this are not entirely understood, and overlap in part with the problems that can be encountered with gradient descent in simple feed-forward networks on regression or classification tasks. There is an additional set of difficulties with gradient descent learning offixed points or trajectories, that is specific to recurrent networks, and that has to do with the bifurcations of the system being considered. In the case of a recurrent2 network, as the parameters are varied, the system mayor may not undergo a series of bifurcations, i.e., of abrupt changes in the structure of its trajectories and, in particular, of its at tractors (fixed points, limit cycles, ... ). This in turn may translate into abrupt discontinuities, oscillations or non-convergence in the corresponding learning curve. At each bifurcation, the error function is usually discontinuous, and therefore the gradient is not defined. Learning can be disrupted in two ways: when unwanted abrupt changes occur in the flow of the dynamical system, or when desirable bifurcations are prevented from occurring. A classical example of the second type is the case of a neural network with very small initial weights being trained to oscillate, in a symmetric and stable fashion, around the origin. With small initial weights, the network in general converges to its unique fixed point at the origin, with a large error. If we slightly perturb the weights, remaining away from any bifurcation, the network continues to converge to its unique fixed point which now may be slightly displaced from the origin, and yield an even greater error, so that learning by gradient descent becomes impossible (the starting configuration of zero weights is a local minimum of the error function). 1 All the results to be derived can be extended immediately to the case of higherdimensional trajectories. 2In a feed-forward network, where the transfer functions of the units are continuous, the output is a continuous function of the parameters and therefore there are no bifurcations. Universal Approximation and Learning of Trajectories Using Oscillators 453 8 o Figure 1: A schematic representation of a 3 layer oscillator network for double figure eight. Oscillators with period T in a given layer gate the corresponding oscillators, with period T /2, in the previous layer. 2.2 DYNAMICAL SYSTEM APPROACH In the dynamical system approach, the function /(t) is approximated in time, over [0, T] by a sequence of points Yo, Yl, .... These points are associated with the iterates of a dynamical system, i.e., Yn+l = F(Yn) = Fn(yo), for some function F. Thus the network implementation requires mainly a feed-forward circuit that computes the function F. It has a simple overall recursive structure where, at time n, the output F(Yn) is calculated, and fed back into the input for the next iteration. While this approach is entirely general, it leaves open the problem of constructing the function F. Of course, F can be learned from examples in a usual feed-forward connectionist network. But, as usual, the complexity and architecture of such a network are difficult to determine in general. Another interesting issue in trajectory learning is how time is represented in the network, and whether some sort of clock is needed. Although occasionally in the literature certain authors have advocated the introduction of an input unit whose output is the time t, this explicit representation is clearly not a suitable representation, since the problem of trajectory learning reduces then entirely to a regression problem. The dynamical system approach relies on one basic clock to calculate F and recycle it to the input layer. In the next approach, an implicit representation of time is provided by the periods of the oscillators. 2.3 OSCILLATOR APPROACH A different approach was suggested in [1] where, loosely speaking, complex trajectories are realized using weakly pre-structured networks, consisting of shallow hierarchical combinations of simple oscillatory modules. The oscillatory modules can consist, for instance, of simple oscillator rings of units satisfying Eq. 1, with two or three high-gain neurons, and an odd number of inhibitory connections ([3]). To fix the ideas, consider the typical test problem of constructing a network capable of producing a trajectory associated with a double figure eight curve (i.e., a set of four loops joined at one point), see Fig. 1. In this example, the first level of the hierarchy could contain four oscillator rings, one for each loop of the target trajectory. The parameters in each one of these four modules can be adjusted, for instance by gradient descent, to match each of the loops in the target trajectory. 454 P. BALDI, K. HORNIK The second level of the pyramid should contain two control modules. Each of these modules controls a distinct pair of oscillator networks from the first level, so that each control network in the second level ends up producing a simple figure eight. Again, the control networks in level two can be oscillator rings and their parameters can be adjusted. In particular, after the learning process is completed, they should be operating in their high-gain regimes and have a period equal to the sum of the periods of the circuits each one controls. Finally, the third layer consists of another oscillatory and adjustable module which controls the two modules in the second level, so as to produce a double figure eight. The third layer module must also end up operating in its high-gain regime with a period equal to four times the period of the oscillators in the first layer. In general, the final output trajectory is also a limit cycle because it is obtained by superposition of limit cycles in the various modules. If the various oscillators relax to their limit cycles independently of one another, it is essential to provide for adjustable delays between the various modules in order to get the proper phase adjustments. In this way, a sparse network with 20 units or so can be constructed that can successfully execute a double figure eight. There are actually different possible neural network realizations depending on how the action of the control modules is implemented. For instance, if the control units are gating the connections between corresponding layers, this amounts to using higher order units in the network. If one high-gain oscillatory unit, with activity c(t) always close to 0 or 1, gates the oscillatory activities of two units Ul(t) and U2(t) in the previous layer, then the overall output can be written as out(t) = C(t)Ul (t) + (1 - C(t))U2(t) . (2) The number of layers in the network then becomes a function of the order of the units one is willing to use. This approach could also be described in terms of a dynamic mixture of experts architecture, in its high gain regime. Alternatively, one could assume the existence of a fast weight dynamics on certain connections governed by a corresponding set of differential equations. Although we believe that oscillators with limit cycles present several attractive properties (stability, short transients, biological relevance, . . . ), one can conceivably use completely different circuits as building blocks in each module. 3 GENERALIZATION AND UNIVERSAL APPROXIMATION We have just described an approach that combines a modular hierarchical architecture, together with some simple form of learning, enabling the synthesis of a neural circuit suitable for the production of a double figure eight trajectory. It is clear that the same approach can be extended to triple figure eight or, for that matter, to any trajectory curve consisting of an arbitrary number of simple loops with a common period and one common point. In fact it can be extended to any arbitrary trajectory. To see this, we can subdivide the time interval [0, T] into n equal intervals of duration f = Tin . Given a certain level of required precision, we can always find n oscillator networks with period T (or a fraction of T) and visible trajectory Ui(t), such that for each i, the i-th portion of the trajectory u(t) with if ~ t ~ (i + l)f can be well approximated by a portion of Ui(t) , the trajectory of the i-th oscillator. The target trajectory can then be approximated as (3) Universal Approximation and Learning of Trajectories Using Oscillators 455 As usual, the control coefficient Cj(t) must have also period T and be equal to 1 for i{ :5 t :5 (i + 1){, and 0 otherwise. The control can be realized with one large high-gain oscillator, or as in the case described above, by a hierarchy of control oscillators arranged, for instance, as a binary tree of depth m if n = 2m , with the corresponding multiple frequencies. We can now turn to a slightly different oscillator approach, where trajectories are to be approximated with linear combinations of oscillators, with constant coefficients. What we would like to show again is that oscillators are universal approximators for trajectories. In a sense, this is already a well-known result of Fourier theory since, for instance, any reasonable function f with period T can be expanded in the form3 A.k = kiT. (4) For sufficiently smooth target functions, without high frequencies in their spectrum, it is well known that the series in Eq. 4 can be truncated. Notice, however, that both Eqs. 3 and 4 require having component oscillators with relatively high frequencies, compared to the final trajectory. This is not implausible in biological motor control, where trajectories have typical time scales of a fraction of a second, and single control neurons operate in the millisecond range. A rather different situation arises if the component oscillators are "slow" with respect to the final product. The Fourier representation requires in principle oscillations with arbitrarily large frequencies (0, liT, 2IT, .. . , niT, .. . ). Most likely, relatively small variations in the parameters (for instance gains, delays andlor synaptic weights) of an oscillator circuit can only lead to relatively small but continuous variations of the overall frequency. For instance, in [3] it is shown that the period T of an oscillator ring with n units obeying Eq. 1 must satisfy Thus, we need to show that a decomposition similar in flavor to Eq. 4 is possible, but using oscillators with frequencies in a bounded interval. Notice that by varying the parameters of a basic oscillator, any frequency in the allowable frequency range can be realized, see [3]. Such a linear combination is slightly different in spirit from Eq. 2, since the coefficients are independent of time, and can be seen as a soft mixture of experts. We have the following result. Theorem 1 Let a < b be two arbitrary real numbers and let f be a continuous function on [0, T]. Then for any error level { > 0, there exist n and a function 9n of the form such that the uniform distance Ilf - 9n 1100 is less than {. In fact, it is not even necessary to vary the frequencies A. over a continuous band [a, b]. We have the following. Theorem 2 Let {A.k} be an infinite sequence with a finite accumulation point, and let f be a continuous function on [0,7]. Then for any error level { > 0, there exist n and a function 9n(t) = 2:~=10:'ke27rjAkt such that Ilf - 9nll00 < {. 3In what follows, we use the complex form for notational convenience. 456 P. BALDI, K. HORNIK Thus, we may even fix the oscillator frequencies as e.g. Ak = l/k without losing universal approximation capabilities. Similar statements can be made about meansquare approximation or, more generally, approximation in p-norm LP(Il), where 1 ~ p < 00 and Il is a finite measure on [0, T]: Theorem 3 For all p and f in LP(Il) and for all { > 0, we can always find nand gn as above such that Ilf - gn IILP{Jl) < {. The proof of these results is surprisingly simple. Following the proofs in [4], if one of the above statements was not true, there would exist a nonzero, signed finite measure (T with support in [0, T] such that hO,T] e21fi>.t d(T(t) = ° for all "allowed" frequencies A. Now the function z t-+ !rO,T] e21fizt d(T(t) is clearly analytic on the whole complex plane. Hence, by a well-known result from complex variables, if it vanishes along an infinite sequence with a finite accumulation point, it is identically zero. But then in particular the Fourier transform of (T vanishes, which in turn implies that (T is identically zero by the uniqueness theorem on Fourier transforms, contradicting the initial assumption. Notice that the above results do not imply that f can exactly be represented as e.g. f(t) = f: e21fi>.t dV(A) for some signed finite measure v-such functions are not only band-limited, but also extremely smooth (they have an analytic extension to the whole complex plane). Hence, one might even conjecture that the above approximations are rather poor in the sense that unrealistically many terms are needed for the approximation. However, this is not true-one can easily show that the rates of approximation cannot be worse that those for approximation with polynomials. Let us briefly sketch the argument, because it also shows how bounded-frequency oscillators could be constructed. Following an idea essentially due to Stinchcombe & White [5], let, more generally, 9 be an analytic function in a neighborhood of the real line for which no derivative vanishes at the origin (above, we had g(t) = e21fit ). Pick a nonnegative integer n and a polynomial p of degree not greater than n - 1 arbitrarily. Let us show that for any { > 0, we can always find a gn of the form gn(t) = E~=l Cl'kg(Akt) with Ak arbitrarily small such that lip - gn 1100 < {. To do so, note that we can write L n - l p(t) = is,t', 1=0 where rn(At) is of the order of An, as A -t 0, uniformly for t in [0, T] . Hence, L:=l Cl'kg(Akt) L:=l Cl'k (L~=-ol fil (At)l + rn (At)) = L~=~l (L:: 1 Cl'kAi) filtl + L:=l Cl'krn (Akt). Now fix n distinct numbers el, ... ,en, let Ak = Ak(p) = pek, and choose the Cl'k = Cl'k(p) such that E;=lCl'k(p)Ak(p)' = iSl/fil for I = 0, ... , n 1. (This is possible because, by assumption, all fil are non-zero.) It is readily seen that Cl'k (p) is of the order of pl-n as p -t ° (in fact, the j-th row of the inverse of the coefficient matrix of the linear system is given by the coefficients of the polynomial nktj (A Ak)/(Aj -Ak)). Hence, as p -t 0, the remainder term EZ=lCl'k(p)rn(Ak(p)t) is ofthe order of p, and thus E~=lCl'k(p)g(Adp)t) -t E~=-oliS,t' = p(t) uniformly on [0, T]. Note that using the above method, the coefficients in the approximation grow quite rapidly when the approximation error tends to 0. In some sense, this was to be Universal Approximation and Learning of Trajectories Using Oscillators 457 expected from the observation that the classes of small-band-limited functions are rather "small". There is a fundamental tradeoff between the size of the frequencies, and the size of the mixing coefficients. How exactly the coefficients scale with the width of the allowed frequency band is currently being investigated. 4 CONCLUSION The modular oscillator approach leads to trajectory architectures which are more structured than fully interconnected networks, with a general feed-forward flow of information and sparse recurrent connections to achieve dynamical effects. The sparsity of units and connections are attractive features for hardware design; and so is also the modular organization and the fact that learning is much more circumscribed than in fully interconnected systems. We have shown in different ways that such architectures have universal approximation properties. In these architectures, however, some form of learning remains essential, for instance to fine tune each one of the modules. This, in itself, is a much easier task than the one a fully interconnected and random network would have been faced with. It can be solved by gradient or random descent or other methods. Yet, fundamental open problems remain in the overall organization of learning across modules, and in the origin of the decomposition. In particular, can the modular architecture be the outcome of a simple internal organizational process rather than an external imposition and how should learning be coordinated in time and across modules (other than the obvious: modules in the first level learn first, modules in the second level second, .. . )? How successful is a global gradient descent strategy applied across modules? How can the same modular architecture be used for different trajectories, with short switching times between trajectories and proper phases along each trajectory? Acknowledgments The work of PB is in part supported by grants from the ONR and the AFOSR. References [1] Pierre Baldi. A modular hierarchical approach to learning. In Proceedings of the 2nd International Conference on Fuzzy Logic and Neural Networks, volume II, pages 985-988, IIzuka, Japan, 1992. [2] Pierre F. Baldi. Gradient descent learning algorithm overview: a general dynamic systems perspective. IEEE Transactions on Neural Networks, 6(1}:182195, January 1995. [3] Pierre F. Baldi and Amir F. Atiya. How delays affect neural dynamics and learning. IEEE Transactions on Neural Networks, 5(4):612-621, July 1994. [4] Kurt Hornik. Some new results on neural network approximation. Neural Networks, 6:1069-1072,1993. [5] Maxwell B. Stinchcombe and Halbert White. Approximating and learning unknown mappings using multilayer feedforward networks with bounded weights. In International Joint Conference on Neural Networks, volume III, pages 7-16, Washington, 1990. Lawrence Earlbaum, Hillsdale.
|
1995
|
124
|
1,026
|
Correlated Neuronal Response: Time Scales and Mechanisms Wyeth Bair Howard Hughes Medical Inst. NYU Center for Neural Science 4 Washington PI., Room 809 New York, NY 10003 Ehud Zohary Dept. of Neurobiology Institute of Life Sciences The Hebrew University, Givat Ram Jerusalem, 91904 ISRAEL Christof Koch Computation and Neural Systems Caltech, 139-74 Pasadena, CA 91125 Abstract We have analyzed the relationship between correlated spike count and the peak in the cross-correlation of spike trains for pairs of simultaneously recorded neurons from a previous study of area MT in the macaque monkey (Zohary et al., 1994). We conclude that common input, responsible for creating peaks on the order of ten milliseconds wide in the spike train cross-correlograms (CCGs), is also responsible for creating the correlation in spike count observed at the two second time scale of the trial. We argue that both common excitation and inhibition may play significant roles in establishing this correlation. 1 INTRODUCTION In a previous study of pairs of MT neurons recorded using a single extracellular electrode, it was found that the spike count during two seconds of visual motion stimulation had an average correlation coefficient of r = 0.12 and that this correlation could significantly limit the usefulness of pooling across increasingly large populations of neurons (Zohary et aI., 1994). However, correlated spike count between two neurons could in principle occur at several time-scales. Correlated drifts Correlated Neuronal Response: Time Scales and Mechanisms 69 in the excitability of the cells, for example due to normal biological changes or electrode induced changes, could cause correlation at a time scale of many minutes. Alternatively, attentional or priming effects from higher areas could change the responsivity of the cells at the time scale of an experimental trial. Or, as suggested here, common input that changes on the order of milliseconds could cause correlation in spike count. The first section determines the time scale at which the neurons are correlated by analyzing the relationship between the peak in the spike train cross-correlograms (CCGs) and the correlation between the spike counts using a construct we call the trial CCG. The second section examines temporal structure that is indicative of correlated suppression of firing, perhaps due to inhibition, which may also contribute to the spike count correlation. 2 THE TIME SCALE OF CORRELATION At the time scale of the single trial, the correlation, r se, of spike counts x and y from two neurons recorded during nominally identical two second stimuli was computed using Pearson's correlation coefficient, E[xy] - ExEy rse = , uxuy (1) where E is expected value and u2 is variance. If spike counts are converted to z-scores, i.e., zero mean and unity variance, then rse = E[xy], and rse may be interpreted as the zero-lag value of the cross-correlation of the z-scored spike counts. The trial CCGs resulting from this procedure are shown for two pairs of neurons in Fig. l. To distinguish between cases like the two shown in Fig. 1, the correlation was broken into a long-term component, rlt, the average value (computed using a Gaussian window of standard deviation 4 trials) surrounding the zero-lag value, and a shortterm component, rst, the difference between the zero-lag value and rlt. Across 92 pairs of neurons from three monkeys, the average rst was 0.10 (s.d. 0.17) while rlt was not significantly different from zero (mean 0.01, s.d. 0.11). The mean of rst was similar to the overall correlation of 0.12 reported by Zohary et al. (1994). Under certain assumptions, including that the time scale of correlation is less than the trial duration, rst can be estimated from the area under the spike train CCG and the areas under the autocorrelations (derivation omitted). Under the additional assumption that the spike trains are individually Poisson and have no peak in the autocorrelation except that which occurs by definition at lag zero, the correlation coefficient for spike count can be estimated by rpeak ~ j.AA.ABArea, (2) where .AA and .AB are the mean firing rates of neurons A and B, and Area is the area under the spike train CCG peak, like that shown in Fig. 2 for one pair of neurons. Taking Area to be the area under the CCG between ±32 msec gives a good estimate of short-term rst, as shown in Fig. 3. In addition to the strong correlation (r = 0.71) between rpeak and rst, rpeak is a less noisy measure, having standard deviation (not shown) on average one fourth as large as those of rst. We conclude that the common input that causes the peaks in the spike train CCGs is also responsible for the correlation in spike count that has been previously reported. 70 0.3 0.2 d U U 0.1 ";3 o 80 160 240 320 0 Trial Number .~ 0 ~ ~~------------------~~ -0.1 +'-r-~..:...-,:......-~-.-,-~~---,-,~,.--,--, W. BAIR. E. ZOHARY. C. KOCH 400 800 1200 Trial Number ernu090 -100 -50 0 50 100 -50 -25 0 25 50 Lag (Trials) Lag (Trials) Figure 1: Normalized responses for two pairs of neurons and their trial crosscorrelograms (CCGs). The upper traces show the z-scored spike counts for all trials in the order they occurred. Spikes were counted during the 2 sec stimulus, but trials occurred on average 5 sec apart, so 100 trials represents about 2.5 minutes. The lower traces show the trial CCGs. For the pair of cells in the left panel, responsivity drifts during the experiment. The CCG (lower left) shows that the drift is correlated between the two neurons over nearly 100 trials. For the pair of cells in the right panel, the trial CCG shows a strong correlation only for simultaneous trials. Thus, the measured correlation coefficient (trial CCG at zero lag) seems to occur at a long time scale on the left but a short time scale (less than or equal to one trial) on the right. The zero-lag value can be broken into two components, T st and Tlt (short term and long term, respectively, see text). The short-term component, T st, is the value at zero lag minus the weighted average value at surrounding lag times. On the left, Tst ~ 0, while on the right, Tlt ~ O. Correlated Neuronal Response: Time Scales and Mechanisms 71 5 o 8 16 24 32 Width at HH (msec) 1 o emu064P -100 -50 o 50 100 Time Lag (msec) Figure 2: A spike train CCG with central peak. The frequency histogram of widths at half-height is shown (inset) for 92 cell pairs from three monkeys. The area of the central peak measured between ±32 msec is used to predict the correlation coefficients, rpeak. plotted in Fig. 3. The y-axis indicates the probability of a coincidence relative to that expected for Poisson processes at the measured firing rates. 0.8 0.6 ~ 0.4 ~ (1) ~ 0.2 • '-" ~ • 0 • • -0.2 -0.2 • • • • • • • • • ..... • • • • • • • • • o 0.2 0.4 0.6 0.8 r (Short Term) Figure 3: The area of the peak of the spike train CCG yields a prediction, rpeak (see Eqn. 2), that is strongly correlated (r = 0.71, p < 0.00001), with the short-term spike count correlation coefficient, rst . The absence of points in the lower right corner of the plot indicates that there are no cases of a pair of cells being strongly correlated without having a peak in the spike train CCG. 72 w. BAIR, E. ZOHARY, C. KOCH In Fig. 3, there are no pairs of neurons that have a short-term correlation and yet do not have a peak in the ±32 msec range of the spike train CCG. 3 CORRELATED SUPPRESSION There is little doubt that common excitatory input causes peaks like the one shown in Fig. 2 and therefore results in the correlated spike count at the time scale of the trial. However, we have also observed correlated periods of suppressed firing that may point to inhibition as another contribution to the CCG peaks and consequently to the correlated spike count. Fig. 4 A and B show the response of one neuron to coherent preferred and null direction motion, respectively. Excessively long inter-spike intervals (ISIs), or gaps, appear in the response to preferred motion, while bursts appear in the response to null motion. Across a database of 84 single neurons from a previous study (Britten et aI., 1992), the occurrence of the gaps and bursts has a symmetrical time course-both are most prominent on average from 600-900 msec post-stimulus onset, although there are substantial variations from cell to cell (Bair, 1995). The gaps, roughly 100 msec long, are not consistent with the slow, steady adaptation (presumably due to potassium currents) which is observed under current injection in neocortical pyramidal neurons, e.g., the RS 1 and RS2 neurons of Agmon and Connors (1992). Fig. 4 C shows spike trains from two simultaneously recorded neurons stimulated with preferred direction motion. The longest gaps appear to occur at about the same time. To assess the correlation with a cross-correlogram, we first transform the spike trains to interval trains, shown in Fig. 4 D for the spike trains in C. This emphasizes the presence of long ISIs and removes some of the information regarding the precise occurrence times of action potentials. The interval crosscorrelation (ICC) between each pair of interval trains is computed and averaged over all trials, and the average shift predictor is subtracted. Fig. 4 E and F show ICCs (thick lines) for two different pairs of neurons. In 17 of 31 pairs (55%), there were peaks in the raw ICC that were at least 4 standard errors above the level of the shift predictor. The peaks were on average centered (mean 4.3 msec, SD 54 msec) and had mean width at half-height of 139 msec (SD 59 msec). To isolate the cause of the peaks, the long intervals in the trains were set to the mean of the short intervals. Long intervals were defined as those that accounted for 30% of the duration of the data and were longer than all short intervals. Note that this is only a small fraction of the number of ISIs in the spike train (typically less than about 10%), since a few long intervals consume the same amount of time as many short intervals. Data from 300-1950 msec was processed, avoiding the on-transient and the lack of final interval. With the longest intervals neutralized, the peaks were pushed down to the level of the noise in the ICC (thin lines, Fig. 4 E, F). Thus, 90% of the action potentials may serve to set a mean rate, while a few periods of long ISIs dominate the ICC peaks. The correlated gaps are consistent with common inhibition to neurons in a local region of cortex, and this inhibition adds area to the spike train CCG peaks in the form of a broader base (not shown). The data analyzed here is from behaving animals, so the gaps may be related to small saccades (within the 0.5 degree Correlated Neuronal Response: Time Scales and Mechanisms 73 1 2 I I 0 II II I I 1111 """111111111111"'"" 11111'""111111111"11 ""11"1'"''''''"' III ""'11111"'"'' 1111111 '"' ""II.!IIIIIIIIIIII "11"""1111111110111111111111111111111""111'1 II IlIg" "' 111111111111111"11111 111"'."111111111111 11111111'"11111111111111111111 111111111'"'"111 IIItIlI! 111111 1111111111'"1111111 1111111111 '"IIIIIII!!I1I11'"1I1I 1I1"'"tll ""UIU ""'"'" '"""" 111'"""""1111 1111"111'"'"1111111 '"""""'"11'"111111' 11111111'"1111111111'"'"""'"' 11""'"1111111 II' ""1111111111.11. IIUII IIIIIU"'"IIIIIIIIIII.II' "'''''1111111111 11,.""""'" 111111111111111' 111'''.,11111111111111111 1111111111111111111'"'"11111"11111111'11"11"11111111111"1 '"'''''" I I I 1111111'111111"""11 1111 1111111111111"'""1111111 "'"111 '"'"1111111'"111111111111111' "11'"" 11111111"1111'"111 111"" 1111 "11"11 """'"111 "" II 111111""" 1111' 1111 111'11'111111111 I III 1M II , '''' I" ! III '''.1'111111111111111,,11111111111111111111111111111111111111111111''"1111" I 11111 I l"lIll! III 1111111 II! I 1111 III II! I!!" I I! "'" '" III III. II 11.11 III II "I 111M"" tI III. " 111"111 I • "I ""I I I 11111 .11111111111111 II 11111111111111'", """""""'"111'"1111111111"1"1 "11'11111'1111111111 1111 1111.1.11111111 '"'1 II!!' I I ,,'''''HIII'''!I1''! "" 11"" 11111111111"1 "'"111'"'" III'"!II " I II! 1111111.111111111111111111111' 11111111""'"1111111111111'1111111111 '1IIIII'"IIIIIIII""""""YlI.""111I ,""'',''1''','' ,,'.', '11I'1~I'M'i"IlI'III"'I,IIII"II'"III,III' t""'IIIIII "II'"'~IIIIII,II!'"'''',,\ I III I "' II "' II I I II 1111 ,. 'I " I I "" ta' II! .111 I'" I I "' • 1111 III " "11 I "" III II .1 I' N 1 I I II • '"~ '111' I! II "' 11 II 11111 , I II. I" I III III III I' 'I! • II II i I A B 500 1000 1500 2000 11111111111111 11.1111 1111 1 E msec 1111111111111111 11111111 II 01 n 11111111 11111111111111111111111111111111111111111111111111118 11111 c 1111 1 11111111 111111 I 1I11 III I 11I1111I111I111I111I111111I11Im1l11111 II D 1000 Time (msec) 2000 F -1000 -500 o 500 1000 -1000 -500 Time Lag (msec) o 500 1000 Figure 4: (A) The brisk response to coherent preferred direction motion is interrupted by occasional excessively long inter-spike intervals, i.e., gaps. (B) The suppressed response to null direction motion is interrupted by bursts of spikes. (C) Simultaneous spike trains from two neurons show correlated gaps in the preferred direction response. (D) The interval representation for the spike trains in C. (E,F) Interval cross-correlograms have peaks indicating that the gaps are correlated (see text). 74 w. BAIR. E. ZOHARY. C. KOCH fixation window) or eyelid blink. It has been hypothesized that blink suppression and saccadic visual suppression may operate through the same pathways and are of neuronal origin (Ridder and Tomlinson, 1993). An alternative hypothesis is that the gaps and bursts arise in cortex from intrinsic circuitry arranged in an opponent fashion. 4 CONCLUSION Common input that causes central peaks on the order of tens of milliseconds wide in spike train CCGs is also responsible for causing the correlation in spike count at the time scale of two second long trials. Long-term correlation due to drifts in responsivity exists but is zero on average across all cell pairs and may represent a source of noise which complicates the accurate measurement of cell-to-cell correlation. The area of the peak of the spike train CCG within a window of ±32 msec is the basis of a good prediction of the spike count correlation coefficient and provides a less noisy measure of correlation between neurons. Correlated gaps observed in the response to coherent preferred direction motion is consistent with common inhibition and contributes to the area of the spike train CCG peak, and thus to the correlation between spike count. Correlation in spike count is an important factor that can limit the useful pool-size of neuronal ensembles (Zohary et al., 1994; Gawne and Richmond, 1993). Acknowledgements We thank William T. Newsome, Kenneth H. Britten, Michael N. Shadlen, and J. Anthony Movshon for kindly providing data that was recorded in previous studies and for helpful discussion. This work was funded by the Office of Naval Research and the Air Force Office of Scientific Research. W. B. was supported by the L. A. Hanson Foundation and the Howard Hughes Medical Institute. References Agmon A, Connors BW (1992) Correlation between intrinsic firing patterns and thalamocortical synaptic responses of neurons in mouse barrel cortex. J N eurosci 12:319-329. Bair W (1995) Analysis of Temporal Structure in Spike Trains of Visual Cortical Area MT. Ph.D. thesis, California Institute of Technology. Britten KH, Shadlen MN, Newsome WT, Movshon JA (1992) The analysis of visual motion: a comparison of neuronal and psychophysical performance. J Neurosci 12:4745-4765. Gawne T J, Richmond BJ (1993) How independent are the messages carried by adjacent inferior temporal cortical neurons? J Neurosci 13:2758-2771. Ridder WH, Tomlinson A (1993) Suppression of contrasts sensitivity during eyelid blinks. Vision Res 33: 1795- 1802. Zohary E, Shadlen MN, Newsome WT (1994) Correlated neuronal discharge rate and its implications for psychophysical performance. Nature 370:140-143.
|
1995
|
125
|
1,027
|
Worst-case Loss Bounds for Single Neurons David P. Helmbold Department of Computer Science University of California, Santa Cruz Santa Cruz, CA 95064 USA Jyrki Kivinen Department of Computer Science P.O. Box 26 (Teollisuuskatu 23) FIN-00014 University of Helsinki Finland Manfred K. Warmuth Department of Computer Science University of California, Santa Cruz Santa Cruz, CA 95064 USA Abstract We analyze and compare the well-known Gradient Descent algorithm and a new algorithm, called the Exponentiated Gradient algorithm, for training a single neuron with an arbitrary transfer function . Both algorithms are easily generalized to larger neural networks, and the generalization of Gradient Descent is the standard back-propagation algorithm. In this paper we prove worstcase loss bounds for both algorithms in the single neuron case. Since local minima make it difficult to prove worst-case bounds for gradient-based algorithms, we must use a loss function that prevents the formation of spurious local minima. We define such a matching loss function for any strictly increasing differentiable transfer function and prove worst-case loss bound for any such transfer function and its corresponding matching loss. For example, the matching loss for the identity function is the square loss and the matching loss for the logistic sigmoid is the entropic loss. The different structure of the bounds for the two algorithms indicates that the new algorithm out-performs Gradient Descent when the inputs contain a large number of irrelevant components. 310 D. P. HELMBOLD, J. KIVINEN, M. K. WARMUTH 1 INTRODUCTION The basic element of a neural network, a neuron, takes in a number of real-valued input variables and produces a real-valued output. The input-output mapping of a neuron is defined by a weight vector W E RN, where N is the number of input variables, and a transfer function ¢. When presented with input given by a vector x E RN, the neuron produces the output y = ¢(w . x). Thus, the weight vector regulates the influence of each input variable on the output, and the transfer function can produce nonlinearities into the input-output mapping. In particular, when the transfer function is the commonly used logistic function, ¢(p) = 1/(1 + e-P), the outputs are bounded between 0 and 1. On the other hand, if the outputs should be unbounded, it is often convenient to use the identity function as the transfer function, in which case the neuron simply computes a linear mapping. In this paper we consider a large class of transfer functions that includes both the logistic function and the identity function, but not discontinuous (e.g. step) functions. The goal of learning is to come up with a weight vector w that produces a desirable input-output mapping. This is achieved by considering a sequence S = ((X1,yt}, ... ,(Xl,Yl» of examples, where for t = 1, ... ,i the value Yt E R is the desired output for the input vector Xt, possibly distorted by noise or other errors. We call Xt the tth instance and Yt the tth outcome. In what is often called batch learning, alli examples are given at once and are available during the whole training session. As noise and other problems often make it impossible to find a weight vector w that would satisfy ¢(w· Xt) = Yt for all t, one instead introduces a loss function L, such as the square loss given by L(y, y) = (y - y)2/2, and finds a weight vector w that minimizes the empirical loss (or training error) l Loss(w,S) = LL(Yt,¢(w . xt}) . (1) t=l With the square loss and identity transfer function ¢(p) = p, this is the well-known linear regression problem. When ¢ is the logistic function and L is the entropic loss given by L(y, y) = Y In(yJY) + (1 - y) In((l - y)/(l - y)), this can be seen as a special case oflogistic regression. (With the entropic loss, we assume 0 ~ Yt, Yt ~ 1 for all t, and use the convention OlnO = Oln(O/O) = 0.) In this paper we use an on-line prediction (or life-long learning) approach to the learning problem. It is well known that on-line performance is closely related to batch learning performance (Littlestone, 1989; Kivinen and Warmuth, 1994). Instead of receiving all the examples at once, the training algorithm begins with some fixed start vector W1, and produces a sequence W1, ... , w l+1 of weight vectors. The new weight vector Wt+1 is obtained by applying a simple update rule to the previous weight vector Wt and the single example (Xt, Yt). In the on-line prediction model, the algorithm uses its tth weight vector, or hypothesis, to make the prediction Yt = ¢(Wt . xt). The training algorithm is then charged a loss L(Yt, Yt) for this tth trial. The performance of a training algorithm A that produces the weight vectors Wt on an example sequence S is measured by its total (cumulative) loss l Loss(A, S) = L L(Yt, ¢(Wt . Xt» . (2) t=l Our main results are bounds on the cumulative losses for two on-line prediction algorithms. One of these is the standard Gradient Descent (GO) algorithm. The other one, which we call EG±, is also based on the gradient but uses it in a different Worst-case Loss Bounds for Single Neurons 311 manner than GD. The bounds are derived in a worst-case setting: we make no assumptions about how the instances are distributed or the relationship between each instance Xt and its corresponding outcome Yt. Obviously, some assumptions are needed in order to obtain meaningful bounds. The approach we take is to compare the total losses, Loss(GD,5) and Loss(EG±, 5), to the least achievable empirical loss, infw Loss( w, 5). If the least achievable empirical loss is high, the dependence between the instances and outcomes in 5 cannot be tracked by any neuron using the transfer function, so it is reasonable that the losses of the algorithms are also high. More interestingly, if some weight vector achieves a low empirical loss, we also require that the losses of the algorithms are low. Hence, although the algorithms always predict based on an initial segment of the example sequence, they must perform almost as well as the best fixed weight vector for the whole sequence. The choice of loss function is crucial for the results that we prove. In particular, since we are using gradient-based algorithms, the empirical loss should not have spurious local minima. This can be achieved for any differentiable increasing transfer function ¢ by using the loss function L¢ defined by r1(y) L¢(y, fj) = f (¢(z) - y) dz . J ¢-l(y) (3) For y < fj the value L¢(y, fj) is the area in the z X ¢(z) plane below the function ¢(z), above the line ¢(z) = y, and to the left of the line z = ¢-l(fj). We call L¢ the matching loss function for transfer function ¢, and will show that for any example sequence 5, if L = L¢ then the mapping from w to Loss(w, 5) is conveX. For example, if the transfer function is the logistic function, the matching loss function is the entropic loss, and ifthe transfer function is the identity function, the matching loss function is the square loss. Note that using the logistic activation function with the square loss can lead to a very large number of local minima (Auer et al., 1996). Even in the batch setting there are reasons to use the entropic loss with the logistic transfer function (see, for example, Solla et al. , 1988). How much our bounds on the losses of the two algorithms exceed the least empirical loss depends on the maximum slope of the transfer function we use. More importantly, they depend on various norms of the instances and the vector w for which the least empirical loss is achieved. As one might expect, neither of the algorithms is uniformly better than the other. Interestingly, the new EG± algorithm is better when most of the input variables are irrelevant, i.e., when some weight vector w with Wi = 0 for most indices i has a low empirical loss. On the other hand, the GD algorithm is better when the weight vectors with low empirical loss have many nonzero components, but the instances contain many zero components. The bounds we derive concern only single neurons, and one often combines a number of neurons into a multilayer feedforward neural network. In particular, applying the Gradient Descent algorithm in the multilayer setting gives the famous back propagation algorithm. Also the EG± algorithm, being gradient-based, can easily be generalized for multilayer feedforward networks. Although it seems unlikely that our loss bounds will generalize to multilayer networks, we believe that the intuition gained from the single neuron case will provide useful insight into the relative performance of the two algorithms in the multilayer case. Furthermore, the EG± algorithm is less sensitive to large numbers of irrelevant attributes. Thus it might be possible to avoid multilayer networks by introducing many new inputs, each of which is a non-linear function of the original inputs. Multilayer networks remain an interesting area for future study. Our work follows the path opened by Littlestone (1988) with his work on learning 312 D. P. HELMBOLD, J. KIVINEN, M. K. WARMUTH thresholded neurons with sparse weight vectors. More immediately, this paper is preceded by results on linear neurons using the identity transfer function (CesaBianchi et aI., 1996; Kivinen and Warmuth, 1994). 2 THE ALGORITHMS This section describes how the Gradient Descent training algorithm and the new Exponentiated Gradient training algorithm update the neuron's weight vector. For the remainder of this paper, we assume that the transfer function </J is increasing and differentiable, and Z is a constant such that </J'(p) ~ Z holds for all pER. For the loss function LcjJ defined by (3) we have aLcjJ(Y, </J(w . x» = (</J(w . x) - Y)Xi . aWi (4) Treating LcjJ(Y, </J(w·x» for fixed x and Y as a function ofw, we see that the Hessian H of the function is given by Hij = </J'(W·X)XiXj. Then v T Hv = </J'(w·x)(v.x)2, so H is positive definite. Hence, for an arbitrary fixed 5, the empirical loss Loss( w, 5) defined in (1) as a function of W is convex and thus has no spurious local minima. We first describe the Gradient Descent (GO) algorithm, which for multilayer networks leads to the back-propagation algorithm. Recall that the algorithm's prediction at trial t is Yt = </J(Wt . Xt), where Wt is the current weight vector and Xt is the input vector. By (4), performing gradient descent in weight space on the loss incurred in a single trial leads to the update rule Wt+l = Wt - TJ(Yt Yt)Xt . The parameter TJ is a positive learning rate that multiplies the gradient of the loss function with respect to the weight vector Wt. In order to obtain worst-case loss bounds, we must carefully choose the learning rate TJ. Note that the weight vector Wt of GO always satisfies Wt = Wi + E!:; aixi for some scalar coefficients ai. Typically, one uses the zero initial vector Wi = O. A more recent training algorithm, called the Exponentiated Gradient (EG) algorithm (Kivinen and Warmuth, 1994), uses the same gradient in a different way. This algorithm makes multiplicative (rather than additive) changes to the weight vector, and the gradient appears in the exponent. The basic version of the EG algorithm also normalizes the weight vector, so the update is given by N Wt+i,i = Wt,ie-IJ(Yt-Yt)Xt" / L Wt,je-IJ(Yt-Y,)Xt,i j=i The start vector is usually chosen to be uniform, Wi = (1/ N, ... ,1/ N). Notice that it is the logarithms of the weights produced by the EG training algorithm (rather than the weights themselves) that are essentially linear combinations of the past examples. As can be seen from the update, the EG algorithm maintains the constraints Wt,i > 0 and Ei Wt,i = 1. In general, of course, we do not expect that such constraints are useful. Hence, we introduce a modified algorithm EG± by employinj a linear transformation of the inputs. In addition to the learning rate TJ, the EG algorithm has a scaling factor U > 0 as a parameter. We define the behavior of EG± on a sequence of examples 5 = ((Xi,Yi), .. . ,(Xl,Yl» in terms of the EG algorithm's behavior on a transformed example sequence 5' = ((xi, yd, .. . , (x~, Yl» Worst-case Loss Bounds for Single Neurons 313 where x' = (U Xl , ... , U XN , -U Xl, ... , -U XN) ' The EG algorithm uses the uniform start vector (1/(2N), . .. , 1/(2N» and learning rate supplied by the EG± algorithm. At each time time t the N-dimensional weight vector w of EG± is defined in terms of the 2N -dimensional weight vector Wi of EG as Wt,i = U(W~ , i W~ ,N+i ) ' Thus EG± with scaling factor U can learn any weight vector w E RN with Ilwlll < U by having the embedded EG algorithm learn the appropriate 2N-dimensional (nonnegative and normalized) weight vector Wi. 3 MAIN RESULTS The loss bounds for the GO and EG± algorithms can be written in similar forms that emphasize how different algorithms work well for different problems. When L = L¢n we write Loss¢(w, S) and Loss¢(A, S) for the empirical loss of a weight vector wand the total loss of an algorithm A, as defined in (1) and (2). We give the upper bounds in terms of various norms. For x E RN, the 2-norm Ilxl b is the Euclidean length of the vector x, the I-norm Ilxlll the sum of the absolute values of the components of x , and the (X)-norm Ilxlloo the maximum absolute value of any component of x . For the purposes of setting the learning rates, we assume that before training begins the algorithm gets an upper bound for the norms of instances. The GO algorithm gets a parameter X2 and EG a parameter Xoo such that IIxtl12 ~ X 2 and Ilxtl loo ~ X oo hold for all t. Finally, recall that Z is an upper bound on ¢/(p). We can take Z = 1 when ¢ is the identity function and Z = 1/4 when ¢ is the logistic function. Our first upper bound is for GO. For any sequence of examples S and any weight vector u ERN, when the learning rate is TJ = 1/(2X?Z) we have Loss¢(GO,S) ~ 2Loss¢(u,S) + 2(llulbX2)2Z . Our upper bounds on the EG± algorithm require that we restrict the one-norm of the comparison class: the set of weight vectors competed against. The comparison class contains all weight vectors u such that Ilulh is at most the scaling factor, U. For any scaling factor U , any sequence of examples S, and any weight vector u ERN with Ilulll ~ U, we have 4 16 Loss¢(EG± , S) ~ 3Loss¢(u,S)+ 3(UXoo )2Z1n(2N) when the learning rate is TJ = 1/(4(UXoo )2Z). Note that these bounds depend on both the unknown weight vector u and some norms of the input vectors. If the algorithms have some further prior information on the sequence S they can make a more informed choice of TJ. This leads to bounds with a constant of 1 before the the Loss¢(u, S) term at the cost of an additional square-root term (for details see the full paper, Helmbold et al. , 1996). It is important to realize that we bound the total loss of the algorithms over any adversarially chosen sequence of examples where the input vectors satisfy the norm bound. Although we state the bounds in terms of loss on the data, they imply that the algorithms must also perform well on new unseen examples, since the bounds still hold when an adversary adds these additional examples to the end of the sequence. A formal treatment of this appears in several places (Littlestone, 1989; 314 D. P. HELMBOLD, J. KIVINEN, M. K. WARMUTH Kivinen and Warmuth, 1994). Furthermore, in contrast to standard convergence proofs (e.g. Luenberger, 1984), we bound the loss on the entire sequence of examples instead of studying the convergence behavior of the algorithm when it is arbitrarily close to the best weight vector. Comparing these loss bounds we see that the bound for the EG± algorithm grows with the maximum component of the input vectors and the one-norm of the best weight vector from the comparison class. On the other hand, the loss bound for the GD algorithm grows with the tWo-norm (Euclidean length) of both vectors. Thus when the best weight vector is sparse, having few significant components, and the input vectors are dense, with several similarly-sized components, the bound for the EG± algorithm is better than the bound for the GD algorithm. More formally, consider the noise-free situation where Lossr/>(u, S) = 0 for some u. Assume Xt E { -1, I}N and U E {-I, 0, I}N with only k nonzero components in u. We can then take X 2 = ..,(N, Xoo = 1, IIuI12 = Vk, and U = k. The loss bounds become (16/3)k 2Z1n(2N) for EG± and 2kZN for GD, so for N ~ k the EG± algorithm clearly wins this comparison. On the other hand, the GD algorithm has the advantage over the EG algorithm when each input vector is sparse and the best weight vector is dense, having its weight distributed evenly over its components. For example, if the inputs Xt are the rows of an N x N unit matrix and U E { -1, 1 } N , then X2 = Xoo = 1, IIuI12 = ..,(N, and U = N. Thus the upper bounds become (16/3)N 2Z1n(2N) for EG± and 2NZ for GD, so here GD wins the comparison. Of course, a comparison of the upper bounds is meaningless unless the bounds are known to be reasonably tight. Our experiments with artificial random data suggest that the upper bounds are not tight. However, the experimental evidence also indicates that EG± is much better than G D when the best weight vector is sparse. Thus the upper bounds do predict the relative behaviors of the algorithms. The bounds we give in this paper are very similar to the bounds Kivinen and Warmuth (1994) obtained for the comparison class of linear functions and the square loss. They observed how the relative performances of the GD and EG± algorithms relate to the norms of the input vectors and the best weight vector in the linear case. Our methods are direct generalizations of those applied for the linear case (Kivinen and Warmuth, 1994). The key notion here is a distance function d for measuring the distance d( u, w) between two weight vectors U and w. Our main distance measures are the Squared Euclidean distance ~ II u - w II ~ and the Relative Entropy distance (or Kullback-Leibler divergence) L~l Ui In(ui/wi). The analysis exploits an invariant over t and u of the form aLr/>(Yt, Wt . Xt) - bLr/>(Yt, U· Xt) ~ d(u, Wt) - d(u, Wt+l) , where a and b are suitably chosen constants. This invariant implies that at each trial, if the loss of the algorithm is much larger than that of an arbitrary vector u, then the algorithm updates its weight vector so that it gets closer to u. By summing the invariant over all trials we can bound the total loss of the algorithms in terms of Lossr/>(u, S) and d(u, wI). Full details will be contained in a technical report (Helmbold et al., 1996). 4 OPEN PROBLEMS Although the presence of local minima in multilayer networks makes it difficult to obtain worst case bounds for gradient-based algorithms, it may be possible to Worst-case Loss Bounds for Single Neurons 315 analyze slightly more complicated settings than just a single neuron. One likely candidate is to generalize the analysis to logistic regression with more than two classes. In this case each class would be represented by one neuron. As noted above, the matching loss for the logistic transfer function is the entropic loss, so this pair does not create local minima. No bounded transfer function matches the square loss in this sense (Auer et aI., 1996), and thus it seems impossible to get the same kind of strong loss bounds for a bounded transfer function and the square loss as we have for any (increasing and differentiable) transfer function and its matching loss function . As the bounds for EG± depend only logarithmically on the input dimension, the following approach may be feasible. Instead of using a multilayer net, use a single (linear or sigmoided) neuron on top of a large set of basis functions. The logarithmic growth of the loss bounds in the number of such basis functions mean that large numbers of basis functions can be tried. Note that the bounds of this paper are only worst-case bounds and our experiments on artificial data indicate that the bounds may not be tight when the input values and best weights are large. However, we feel that the bounds do indicate the relative merits of the algorithms in different situations. Further research needs to be done to tighten the bounds. Nevertheless, this paper gives the first worst-case upper bounds for neurons with nonlinear transfer functions. References P. Auer, M. Herbster, and M. K. Warmuth (1996). Exponentially many local minima for single neurons. In Advances in Neural Information Processing Systems 8. N. Cesa-Bianchi, P. Long, and M. K. Warmuth (1996). Worst-case quadratic loss bounds for on-line prediction of linear functions by gradient descent. IEEE Transactions on Neural Networks. To appear. An extended abstract appeared in COLT '93, pp. 429-438. D. P. Helmbold, J . Kivinen, and M. K. Warmuth (1996). Worst-case loss bounds for single neurons. Technical Report UCSC-CRL-96-2, Univ. of Calif. Computer Research Lab, Santa Cruz, CA, 1996. In preparation. J . Kivinen and M. K. Warmuth (1994). Exponentiated gradient versus gradient descent for linear predictors. Technical Report UCSC-CRL-94-16, Univ. of Calif. Computer Research Lab, Santa Cruz, CA, 1994. An extended abstract appeared in STOC '95, pp. 209-218. N. Littlestone (1988). Learning when irrelevant attributes abound: A new linearthreshold algorithm. Machine Learning, 2:285-318. N. Littlestone (1989). From on-line to batch learning. In Proc. 2nd Annual Workshop on Computational Learning Theory, pages 269-284. Morgan Kaufmann, San Mateo, CA. D. G. Luenberger (1984). Linear and Nonlinear Programming. Addison-Wesley, Reading, MA. S. A. Solla, E. Levin, and M. Fleisher (1988). Accelerated learning in layered neural networks. Complex Systems, 2:625- 639 .
|
1995
|
126
|
1,028
|
Stock Selection via Nonlinear Multi-Factor Models Asriel U. Levin BZW Barclays Global Investors Advanced Strategies and Research Group 45 Fremont Street San Francisco CA 94105 email: asriel.levin@bglobal.com Abstract This paper discusses the use of multilayer feed forward neural networks for predicting a stock's excess return based on its exposure to various technical and fundamental factors. To demonstrate the effectiveness of the approach a hedged portfolio which consists of equally capitalized long and short positions is constructed and its historical returns are benchmarked against T-bill returns and the S&P500 index. 1 Introduction Traditional investment approaches (Elton and Gruber, 1991) assume that the return of a security can be described by a multifactor linear model: (1) where Hi denotes the return on security i, Fl are a set of factor values and Uil are security i exposure to factor I, ai is an intercept term (which under the CAPM framework is assumed to be equal to the risk free rate of return (Sharpe, 1984)) and ei is a random term with mean zero which is assumed to be uncorrelated across securities. The factors may consist of any set of variables deemed to have explanatory power for security returns. These could be aspects of macroeconomics, fundamental security analysis, technical attributes or a combination of the above. The value of a factor is the expected excess return above risk free rate of a security with unit exposure to the factor and zero exposure to all other factors. The choice offactors can be viewed as a proxy for the" state of the world" and their selection defines a metric imposed on the universe of securities: Once the factors are set, the model assumption is that, Stock Selection via Nonlinear Multi-factor Models 967 on average, two securities with similar factor loadings (Uil) will behave in a similar manner. The factor model (1) was not originally developed as a predictive model, but rather as an explanatory model, with the returns It; and the factor values Pi assumed to be contemporaneous. To utilize (1) in a predictive manner, each factor value must be replaced by an estimate, resulting in the model A A A It; = ai + UilFl + Ui2 F 2 + ... + UiLFL + ei (2) where Ri is a security's future return and F/ is an estimate of the future value of factor 1, based on currently available information. The estimation of Fl can be approached with varying degree of sophistication ranging from a simple use of the historical mean to estimate the factor value (setting Fl(t) = Fi), to more elaborate approaches attempting to construct a time series model for predicting the factor values. Factor models of the form (2) can be employed both to control risk and to enhance return. In the first case, by capturing the major sources of correlation among security returns, one can construct a well balanced portfolio which diversifies specific risk away. For the latter, if one is able to predict the likely future value of a factor, higher return can be achieved by constructing a portfolio that tilts toward "good" factors and away from "bad" ones. While linear factor models have proven to be very useful tools for portfolio analysis and investment management, the assumption of linear relationship between factor values and expected return is quite restrictive. Specifically, the use of linear models assumes that each factor affects the return independently and hence, they ignore the possible interaction between different factors. Furthermore, with a linear model, the expected return of a security can grow without bound as its exposure to a factor increases. To overcome these shortcomings of linear models, one would have to consider more general models that allow for nonlinear relationship among factor values, security exposures and expected returns. Generalizing (2), while maintaining the basic premise that the state of the world can be described by a vector of factor values and that the expected return of a security is determined through its coordinates in this factor world, leads to the nonlinear model: It; = j(Uil' Ui2,···, UiL, Fl , F2, ... , FL ) + ei (3) where JO is a nonlinear function and ei is the noise unexplained by the model, or "security specific risk" . The prediction task for the nonlinear model (3) is substantially more complex than in the linear case since it requires both the estimation of future factor values as well as a determination of the unknown function j. The task can be somewhat simplified if factor estimates are replaced with their historical means: It; J(Uil, Ui2, ... , UiL, lA, F2, ... , FL) + ei (4) where now Uil are the security's factor exposure at the beginning of the period over which we wish to predict. To estimate the unknown function t(-), a family of models needs to be selected, from which a model is to be identified. In the following we propose modeling the relationship between factor exposures and future returns using the class of multilayer feedforward neural networks (Hertz et al., 1991). Their universal approximation 968 A. U. LEVIN capabilities (Cybenko, 1989; Hornik et al., 1989), as well as the existence of an effective parameter tuning method (the backpropagation algorithm (Rumelhart et al., 1986)) makes this family of models a powerful tool for the identification of nonlinear mappings and hence a natural choice for modeling (4). 2 The stock selection problem Our objective in this paper is to test the ability of neural network based models of the form (4) to differentiate between attractive and unattractive stocks. Rather than trying to predict the total return of a security, the objective is to predict its performance relative to the market, hence eliminating the need to predict market directions and movements. The data set consists of monthly historical records (1989 through 1995) for the largest 1200-1300 US companies as defined by the BARRA HiCap universe. Each data record (::::::1300 per month) consists of an input vector composed of a security's factor exposures recorded at the beginning of the month and the corresponding output is the security's return over the month. The factors used to build the model include Earning/Price, Book/Price, past price performance, consensus of analyst sentiments etc, which have been suggested in the financial literature as having explanatory power for security returns (e.g. (Fama and French, 1992)). To minimize risk, exposure to other unwarranted factors is controlled using a quadratic optimizer. 3 Model construction and testing Potentially, changes in a price of a security are a function of a very large number of forces and events, of which only a small subset can be included in the factor model (4). All other sources of return play the role of noise whose magnitude is probably much larger than any signal that can be explained by the factor exposures. When this information is used to train a neural network, the network attempts to replicate the examples it sees and hence much of what it tries to learn will be the particular realizations of noise that appeared in the training set. To minimize this effect, both a validation set and regularization are used in the training. The validation set is used to monitor the performance of the model with data on which it has not been trained on. By stopping the learning process when validation set error starts to increase, the learning of noise is minimized. Regularization further limits the complexity of the function realized by the network and, through the reduction of model variance, improves generalization (Levin et al., 1994). The stock selection model is built using a rolling train/test window. First, M "two layer" feedforward networks are built for each month of data (result is rather insensitive to the particular choice of M). Each network is trained using stochastic gradient descent with one quarter of the monthly data (randomly selected) used as a validation set. Regularization is done using principal component pruning (Levin et al., 1994). Once training is completed, the models constructed over N consecutive month of data (again, result is insensitive to particular choice of N) are combined (thus increasing the robustness of the model (Breiman, 1994)) to predict the returns in the following month. Thus the predicted (out of sample) return of stock i in month k is given by (5) Stock Selection via Nonlinear Multi-factor Models 0.4 0.35 0.3 0.25 c 0 :;:::; 0.2 ctI Q) .... .... 0 0.15 () 0.1 ,-, , , 0.05 , ! i 0 i 0 5 Nonlinear Linear ~ -., I : i ! r---: ! ! 1! i r---l b -I I ! r- -I i ~mf ----_]'-T_-I ': : -rl -l-.+-r-+-,-...L-J.-l- -L J. , I L _J 10 15 20 Cell 969 Figure 1 : Average correlation between predicted alphas and realized returns for linear and nonlinear models where k(k) is stock's i predicted return, N Nk-j(·) denoted the neural network model built in month k - j and u71 are stock's i factor exposures as measured at the beginning of month k. 4 Benchmarking to linear As a first step in evaluating the added value of the nonlinear model, its performance was benchmarked against a generalized least squares linear model. Each model was run over three universes: all securities in the HiCap universe, the extreme 200 stocks (top 100, bottom 100 as defined by each model), and the extreme 100 stocks. As a comparative performance measure we use the Sharpe ratio (Elton and Gruber, 1991). As shown in Table 4, while the performance of the two models is quite comparable over the whole universe of stocks, the neural network based model performs better at the extremes, resulting in a substantially larger Sharpe ratio (and of course, when constructing a portfolio, it is the extreme alphas that have the most impact on performance). I Portfolio\Model II Linear Nonlinear II All HiCap 6.43 6.92 100 long/100 short 4.07 5.49 50 long/50 short 3.07 4.23 Table 1: Ex ante Sharpe ratios: Neural network vs. linear While the numbers in the above table look quite impressive, it should be emphasised that they do not represent returns of a practical strategy: turnover is huge and the figures do not take transaction costs into account. The main purpose of the table 970 A. U. LEVIN is to compare the information that can be captured by the different models and specifically to show the added value of the neural network at the extremes. A practical implementation scheme and the associated performance will be discussed in the next section. Finally, some insight as to the reason for the improved performance can be gained by looking at the correlation between model predictions and realized returns for different values of model predictions (commonly referred to as alphas). For that, the alpha range was divided to 20 cells, 5% of observations in each and correlations were calculated separately for each cell. As is shown in figure 1, while both neural network and linear model seem to have more predictive power at the extremes, the network's correlations are substantially larger for both positive and negative alphas. 5 Portfolio construction Given the superior predictive ability of the nonlinear model at the extremes, a natural way of translating its predictions into an investment strategy is through the use of a long/short construct which fully captures the model information on both the positive as well as the negative side. The long/short portfolio (Jacobs and Levy, 1993) is constructed by allocating equal capital to long and short positions. By monitoring and controlling the risk characteristics on both sides, one is able to construct a portfolio that has zero correlation with the market ((3 = 0) - a "market neutral" portfolio. By construction, the return of a market neutral portfolio is insensitive to the market up or down swings and its only source of return is the performance spread between the long and short positions, which in turn is a direct function of the model (5) discernment ability. Specifically, the translation of the model predictions into a realistically implementable strategy is done using a quadratic optimizer. Using the model predicted returns and incorporating volatility information about the various stocks, the optimizer is utilized to construct a portfolio with the following characteristics: • Market neutral (equal long and short capitalization). • Total number of assets in the portfolio <= 200. • Average (one sided) monthly turnover ~ 15%. • Annual active risk ~ 5%. In the following, all results are test set results (out of sample), net of estimated transaction costs (assumed to be 1.5% round trip). The standard benchmark for a market neutral portfolio is the return on 3 month T-bill and as can be seen in Table 2, over the test period the market neutral portfolio has consistently and decisively outperformed its benchmark. Furthermore, the results reported for 1995 were recorded in real-time (simulated paper portfolio). An interesting feature of the long/short construct is its ease of transportability (Jacobs and Levy, 1993). Thus, while the base construction is insensitive to market movement, if one wishes, full exposure to a desired market can be achieved through the use of futures or swaps (Hull, 1993). As an example, by adding a permanent S&P500 futures overlay in an amount equal to the invested capital, one is fully exposed to the equity market at all time, and returns are the sum of the long/short performance spread and the profits or losses resulting from the market price movements. This form of a long/short strategy is referred to as an "equitized" strategy and the appropriate benchmark will be overlayed index. The relative performance Stock Selection via Nonlinear Multi-factor Models 971 I Statistics II T-Bill I Neutral II S&P500 I Equitized II Total Return~%) 27.8 131.5 102.0 264.5 Annual total(Yr%) 4.6 16.8 10.4 27.0 Active Return(%) 103.7 162.5 Annual active(Yr%) 12.2 16.6 Active risk(Yr%) 4.8 4.8 Max draw down(%) 3.2 13.9 10.0 Turnover(Y r%) 198.4 198.4 Table 2: Comparative summary of ex ante portfolio performance (net of transaction costs) 8/90 - 12/95 4 3.5 Equitized --SP500 -+--3 Neutral .-0 . .. . T-bill ...... _. (I) ;:) 2.5 «i > .2 ~ 2 0 0.. C]) > ~ 1.5 "S E ::I () 91 92 93 94 95 96 Year Figure 2: Cumulative portfolio value 8/90 - 12/95 (net of estimated transaction costs) of the equitized strategy with an S&P500 futures overlay is presented in Table 2. Summary of the accumulated returns over the test period for the market neutral and equitized portfolios compared to T-bill and S&P500 are given in Figure 2. Finally, even though the performance of the model is quite good, it is very difficult to convince an investor to put his money on a "black box". A rather simple way to overcome this problem of neural networks is to utilize a CART tree (Breiman et aI., 1984) to explain the model's structure. While the performance of the tree on the raw data in substantially inferior to the network's, it can serve as a very effective tool for analyzing and interpreting the information that is driving the model. 6 Conclusion We presented a methodology by which neural network based models can be used for security selection and portfolio construction. In spite of the very low signal to noise ratio of the raw data, the model was able to extract meaningful relationship 972 A. U. LEVIN between factor exposures and expected returns. When utilized to construct hedged portfolios, these predictions achieved persistent returns with very favorable risk characteristics. The model is currently being tested in real time and given its continued consistent performance, is expected to go live soon. References Anderson, J. and Rosenfeld, E., editors (1988) . Neurocomputing: Foundations of Research. MIT Press, Cambridge. Breiman, L. (1994). Bagging predictors. Technical Report 416, Department of Statistics, VCB, Berkeley, CA. Breiman, L., Friedman, J ., Olshen, R., and Stone, C. (1984). Classification and Regression Trees. Chapman & Hall. Cybenko, G. (1989) . Approximation by superpositions of a sigmoidal function. Mathematics of Control, Signals, and Systems, 2:303-314. Elton, E. and Gruber, M. (1991). Modern Portfolio Theory and Investment Analysis. John Wiley. Fama, E. and French, K. (1992). The cross section of expected stock returns. Journal of Finance, 47:427- 465 . Hertz, J., Krogh, A., and Palmer, R. (1991). Introduction to the theory of neural computation, volume 1 of Santa Fe Institute studies in the sciences ofcomplexity. Addison Wesley Pub. Co. Hornik, K. , Stinchcombe, M., and White, H. (1989). Multilayer feedforward networks are universal approximators. Neural Networks, 2:359-366. Hull, J . (1993). Options, Futures and Other Derivative Securities. Prentice-Hall. Jacobs, B. and Levy, K. (1993). Long/short equity investing. Journal of Portfolio Management, pages 52-63. Levin, A. V., Leen, T. K., and Moody, J . E. (1994). Fast pruning using principal components. In Cowan, J . D., Tesauro, G., and Alspector, J., editors, Advances in Neural Information Processing Systems, volume 6. Morgan Kaufmann. to apear. Rumelhart, D., Hinton, G., and Williams, R. (1986) . Learning representations by back-propagating errors. Nature, 323:533- 536. Reprinted in (Anderson and Rosenfeld, 1988). Sharpe, W. (1984) . Factor models, CAPMs and the APT. Journal of Portfolio Management, pages 21-25.
|
1995
|
127
|
1,029
|
The Geometry of Eye Rotations and Listing's Law Amir A. Handzel* Tamar Flasht Department of Applied Mathematics and Computer Science Weizmann Institute of Science Rehovot, 76100 Israel Abstract We analyse the geometry of eye rotations, and in particular saccades, using basic Lie group theory and differential geometry. Various parameterizations of rotations are related through a unifying mathematical treatment, and transformations between co-ordinate systems are computed using the Campbell-BakerHausdorff formula. Next, we describe Listing's law by means of the Lie algebra so(3). This enables us to demonstrate a direct connection to Donders' law, by showing that eye orientations are restricted to the quotient space 80(3)/80(2). The latter is equivalent to the sphere S2, which is exactly the space of gaze directions. Our analysis provides a mathematical framework for studying the oculomotor system and could also be extended to investigate the geometry of mUlti-joint arm movements. 1 INTRODUCTION 1.1 SACCADES AND LISTING'S LAW Saccades are fast eye movements, bringing objects of interest into the center of the visual field. It is known that eye positions are restricted to a subset of those which are anatomically possible, both during saccades and fixation (Tweed & Vilis, 1990). According to Donders' law, the eye's gaze direction determines its orientation uniquely, and moreover, the orientation does not depend on the history of eye motion which has led to the given gaze direction. A precise specification of the "allowed" subspace of position is given by Listing's law: the observed orientations of the eye are those which can be reached from the distinguished orientation called primary *hand@wisdom.weizmann.ac.il t tamar@wisdom.weizmann.ac.il 118 A. A. HANDZEL, T. FLASH position through a single rotation about an axis which lies in the plane perpendicular to the gaze direction at the primary position (Listing's plane). We say then that the orientation of the eye has zero torsion. Recently, the domain of validity of Listing's law has been extended to include eye vergence by employing a suitable mathematical treatment (Van Rijn & Van Den Berg, 1993). Tweed and Vilis used quaternion calculus to demonstrate, in addition, that in order to move from one allowed position to another in a single rotation, the rotation axis itself lies outside Listing's plane (Tweed & Vilis, 1987). Indeed, normal saccades are performed approximately about a single axis. However, the validity of Listing's law does not depend on the rotation having a single axis, as was shown in double-step target displacement experiments (Minken, Van Opstal & Van Gisbergen, 1993): even when the axis of rotation itself changes during the saccade, Listing's law is obeyed at each and every point along the trajectory which is traced by the eye. Previous analyses of eye rotations (and in particular of Listing's law) have been based on various representations of rotations: quaternions (Westheimer, 1957), rotation vectors (Hepp, 1990), spinors (Hestenes, 1994) and 3 x 3 rotation matrices; however, they are all related through the same underlying mathematical object the three dimensional (3D) rotation group. In this work we analyse the geometry of saccades using the Lie algebra of the rotation group and the group structure. Next, we briefly describe the basic mathematical notions which will be needed later. This is followed by Section 2 in which we analyse various parameterizations of rotations from the point of view of group theory; Section 3 contains a detailed mathematical analysis of Listing's law and its connection to Donders' law based on the group structure; in Section 4 we briefly discuss the issue of angular velocity vectors or axes of rotation ending with a short conclusion. 1.2 THE ROTATION GROUP AND ITS LIE ALGEBRA The group of rotations in three dimensions, G = 80(3), (where '80' stands for special orthogonal transformations) is used both to describe actual rotations and to denote eye positions by means of a unique virtual rotation from the primary position. The identity operation leaves the eye at the primary position, therefore, we identify this position with the unit element of the group e E 80(3). A rotation can be parameterized by a 3D axis and the angle of rotation about it. Each axis "generates" a continuous set of rotations through increasing angles. Formally, if n is a unit axis of rotation, then EXP(O· n) (1) is a continuous one-parameter subgroup (in G) of rotations through angles () in the plane that is perpendicular to n. Such a subgroup is denoted as 80(2) C 80(3). We can take an explicit representation of n as a matrix and the exponent can be calculated as a Taylor series expansion. Let us look, for example, at the one parameter subgroup of rotations in the y- z plane, i.e. rotations about the x axis which is represented in this case by the matrix o o -1 A direct computation of this rotation by an angle () gives o cos () - sin () o sin () cos () (2) ) (3) The Geometry of Eye Rotations and Listing's Law 119 where I is the identity matrix. Thus, the rotation matrix R( 0) can be constructed from the axis and angle of rotation. The same rotation, however, could also be achieved using A Lx instead of Lx, where A is any scalar, while rescaling the angle to 0/ A. The collection of matrices ALx is a one dimensional linear space whose elements are the generators of rotations in the y-z plane. The set of all the generators constitutes the Lie algebra of a group. For the full space of 3D rotations, the Lie algebra is the three dimensional vector space that is spanned by the standard orthonormal basis comprising the three direction vectors of the principal axes: (4) Every axis n can be expressed as a linear combination of this basis. Elements of the Lie algebra can also be represented in matrix form and the corresponding basis for the matrix space is L.= 0 0 D L, = ( 0 0 n L, = ( ~1 1 D; 0 0 0 0 -1 -1 0 0 (5) hence we have the isomorphism ( -~, Oz 8, ) U: ) 0 Ox +-------t -Oy -Ox 0 (6) Thanks to its linear structure, the Lie algebra is often more convenient for analysis than the group itself. In addition to the linear structure, the Lie algebra has a bilinear antisymmetric operation defined between its elements which is called the bracket or commutator. The bracket operation between vectors in g is the usual vector cross product. When the elements of the Lie algebra are written as matrices, the bracket operation becomes a commutation relation, i.e. [A,B] == AB - BA. (7) As expected, the commutation relations of the basis matrices of the Lie algebra (of the 3D rotation group) are equivalent to the vector product: (8) Finally, in accordance with (1), every rotation matrix is obtained by exponentiation: R(8) = EXP(OxLx +OyLy +OzLz). where 8 stands for the three component angles. 2 CO-ORDINATE SYSTEMS FOR ROTATIONS (9) In linear spaces the "position" of a point is simply parameterized by the co-ordinates w.r.t. the principal axes (a chosen orthonormal basis). For a non-linear space (such as the rotation group) we define local co-ordinate charts that look like pieces of a vector space ~ n. Several co-ordinate systems for rotations are based on the fact that group elements can be written as exponents of elements of the Lie algebra (1). The angles 8 appearing in the exponent serve as the co-ordinates. The underlying property which is essential for comparing these systems is the noncommutativity of rotations. For usual real numbers, e.g. Cl and C2, commutativity implies expCI expC2 = expCI +C2. A corresponding equation for non-commuting elements is the Campbell-Baker-Hausdorff formula (CBH) which is a Taylor series 120 A. A. HANDZEL. T. FLASH expansion using repeated commutators between the elements of the Lie algebra. The expansion to third order is (Choquet-Bruhat et al., 1982): EXP(Xl)EXP(X2) = EXP (Xl + X2 + ~[Xl' X2] + 112 [Xl - X2, [Xl, X2]]) (10) where Xl, X2 are variables that stand for elements of the Lie algebra. One natural parameterization uses the representation of a rotation by the axis and the angle of rotation. The angles which appear in (9) are then called canonical co-ordinates of the first kind (Varadarajan, 1974). Gimbal systems constitute a second type of parameterization where the overall rotation is obtained by a series of consecutive rotations about the principal axes. The component angles are then called canonical co-ordinates of the second kind. In the present context, the first type of co-ordinates are advantageous because they correspond to single axis rotations which in turn represent natural eye movements. For convenience, we will use the name canonical co-ordinates for those of the first kind, whereas those of the second type will simply be called gimbals. The gimbals of Fick and Helmholtz are commonly used in the study of oculomotor control (Van Opstal, 1993). A rotation matrix in Fick gimbals is RF(Bx,Oy,Oz) = EXP(OzLz) . EXP(ByLy) . EXP(OxLx), (11) and in Helmholtz gimbals the order of rotations is different: RH(Ox, By,Oz) = EXP(ByLy) . EXP(OzLz) . EXP(OxLx). (12) The CBH formula (10) can be used as a general tool for obtaining transformations between various co-ordinate systems (Gilmore, 1974) such as (9,11,12). In particular, we apply (10) to the product of the two right-most terms in (11) and then again to the product of the result with the third term. We thus arrive at an expression whose form is the same as the right hand side of (10). By equating it with the expression for canonical angles (9) and then taking the log of the exponents on both sides of the equation, we obtain the transformation formula from Fick angles to canonical angles. Repeating this calculation for (12) gives the equivalent formula for Helmholtz anglesl . Both transformations are given by the following three equations where OF,H stands for an angle either in Fick or in Helmholtz co-ordinates; for Helmholtz angles there is a plus sign in front of the last term of the first equation and a minus sign in the case of Fick angles: Be - OF,H (1 _ ...L ((BF,H)2 + (OF,H)2)) ± lOF,H OF,H x x 12 Y z 2 Y z Of = O:,H ( 1 - /2 (( O;,H)2 + (O:,H)2) ) + ~O;,H O:,H Of = O;,H ( 1 - /2 (( B;,H? + (B:,H)2)) - !O;,H O:,H (13) The error caused by the above approximation is smaller than 0.1 degree within most of the oculomotor range. We mention in closing two additional parameterizations, namely quaternions and rotation vectors. Unit quaternions lie on the 3D sphere S3 (embedded in lR 4) which constitutes the same manifold as the group of unitary rotations SU(2). The latter is the double covering group of SO(3) having the same local structure. This enables to use quaternions to parameterize rotations. The popular rotation vectors (written as tan(Oj2)n, n being the axis of rotation and B its angle) are closely related to 1 In contrast to this third order expansion, second order approximations usually appear in the literature; see for example equation B2 in (Van Rijn & Van Den Berg, 1993). The Geometry of Eye Rotations and Listing's Law 121 quaternions because they are central (gnomonic) projections of a hemisphere of S3 onto the 3D affine space tangent to the quaternion qe = (1,0,0,0) E ]R4. 2 3 LISTING'S LAW AND DONDERS' LAW A customary choice of a head fixed coordinate system is the following: ex IS III the straight ahead direction in the horizontal plane, ey is in the lateral direction and ez points upwards in the vertical direction. ex and ez thus define the midsagittal plane; ey and ez define the coronal plane. The principal axes of rotations (Lx, Ly, Lz) are set parallel to the head fixed co-ordinate system. A reference eye orientation called the primary position is chosen with the gaze direction being (1,0,0) in the above co-ordinates. How is Listing's law expressed in terms of the Lie algebra of SO(3)? The allowed positions are generated by linear combinations of Lz and Ly only. This 2D subspace of the Lie algebra, 1 = Span{Ly, Lz }, (14) is Listing's plane. Denoting Span{ Lx} by h, we have a decomposition of the Lie algebra so(3) into a direct sum of two linear subspaces: 9 = 1 EB h. (15) Every vector v E 9 can be projected onto its component which is in I: proj. V = VI + Vh ----t VI. (16) Until now, only the linear structure has been considered. In addition, h is closed under the bracket operation: (17) and because h is closed both under vector addition and the Lie bracket, it is a sub algebra of g. In contrast, I is not a sub algebra because it is not closed under commutation (8) . The fact that h stands as an algebra on its own implies that it has a corresponding group H, just as 9 = so(3) corresponds to G = SO(3). The subalgebra h generates rotations about the x axis, and therefore H is SO(2), the group of rotations in a plane. The group G = SO(3) does not have a linear structure. We may still ask whether some kind of decomposition and projection can be achieved in G in analogy to (15,16). The answer is positive and the projection is performed as follows: take any element of the group, a E G, and multiply it by all the elements of the subgroup H. This gives a subset in G which is considered as a single object a called a coset: a = {ab I bEH} . (18) The set of all cosets constitutes the quotient space. It is written as S == G / H = SO(3)/ SO(2) (19) because mapping the group to the quotient space can be understood as dividing G by H. The quotient space is not a group , and this corresponds to the fact that the subspace I above (14) is not a sub algebra. The quotient space has been constructed algebraically but is difficult to visualize; however, it is mathematically equivalent 2 Geometrically, each point q E S3 can be connected to the center of the sphere by a line. Another line runs from qe in the direction parallel to the vector part of q within the tangent space. The intersection of the two lines is the projected point. Numerically, one simply takes the vector part of q divided by its scalar part. 122 A. A.HANDZEL,T. FLASH Table 1: Summary table of biological notions and the corresponding mathematical representation, both in terms of the rotation group and its Lie algebra. Biological notion Lie Algebra Rotation Group general eye position 9 = so(3) = h El71 G = SO(3) primary position O.q E 9 eE G eye torsion h = Span{Lx} H = SO(2) "allowed" eye 1= Span{ Ly, LzJ S = ~/H = SO(3)/SO(2) positions (Listing's plane) ~ S2 (Donders' sphere of gaze directions) to another space the unit sphere S2 (embedded in ~3). This equivalence can be seen in the following way: a unit vector in ~3, e.g. e = (1,0,0), can be rotated so that its head reaches every point on the unit sphere S2; however, for any such point there are infinitely many rotations by which the point can be reached. Moreover, all the rotations around the x axis leave the vector e above invariant. We therefore have to "factor out" these rotations (of H =SO(2» in order to eliminate the above degeneracy and to obtain a one-to-one correspondence between the required subset of rotations and the sphere. This is achieved by going to the quotient space. The matrix of a torsion less rotation (generated by elements in Listing's plane) is obtained by setting Ox = 0 in (9): ( cosO R = - sin 0 sin ljJ - sin 0 cos ljJ sin 0 sin ljJ cos 0 + (1 - cos 0) cos2 ljJ cos ljJ sin ljJ(l - cos 0) cos ljJ sin ljJ(l - cos 0) ,(20) sin 0 cos ljJ ) cos 0 + (1 - cos 0) sin 2 ljJ where 0 = .)0;+0; is the total angle of rotation and ljJ is the angle between 0 and the y axis in the Oy -Oz plane, i.e. (0, ljJ) are polar co-ordinates in Listing's plane. Notice that the first column on the left constitutes the Cartesian co-ordinates of a point on a sphere of unit radius (Gilmore, 1974). As we have just seen, there is an exact correspondence between the group level and the Lie algebra level. In fact, the two describe the same reality, the former in a global manner and the latter in an infinitesimal one. Table 1 summarizes the important biological notions concerning Listing's law together with their corresponding mathematical representations. The connection between Donders' law and Listing's law can now be seen in a clear and intuitive way. The sphere, which was obtained by eliminating torsion, is the space of gaze directions. Recall that Donders' law states that the orientation of the eye is determined uniquely by its gaze direction. Listing's law implies that we need only take into consideration the gaze direction and disregard torsion. In order to emphasize this point, we use the fact that locally, SO(3) looks like a product of topological spaces: 3 p = u x SO(2) where (21) U parameterizes gaze direction and SO(2) torsion. Donders' law restricts eye orientation to an unknown 2D submanifold of the product space P. Listing's law shows that the submanifold is U, a piece of the sphere. This representation is advantageous for biological modelling, because it mathematically sets apart the degrees of freedom of gaze orientation from torsion, which also differ functionally. 350(3) is a principal bundle over S2 with fiber 50(2). The Geometry of Eye Rotations and Listing's Law 123 4 AXES OF ROTATION FOR LISTING'S LAW As mentioned in the introduction, moving between two (non-primary) positions requires a rotation whose axis (i.e. angular velocity vector) lies outside Listing's plane. This is a result of the group structure of SO(3). Had the axis of rotation been contained within Listing's plane, the matrices of the quotient space (20) should have been closed under multiplication so as to form a subgroup of SO(3). In other words, if ri and rJ are matrices representing the current and target orientations of the eye corresponding to axes in Listing's plane, then rJ . r;l should have been a matrix of the same form (20); however, as explained in Section 3, this condition is not fulfilled. Finally, since normal saccades involve rotations about a single axis, they are oneparameter subgroups generated by a single element of the Lie algebra (1). In addition, they have the property of being geodesic curves in the group manifold under the natural metric which is given by the bilinear Cartan-Killing form of the group (Choquet-Bruhat et al., 1982). 5 CONCLUSION We have analysed the geometry of eye rotations using basic Lie group theory and differential geometry. The unifying view presented here can serve to improve the understanding of the oculomotor system. It may also be extended to study the three dimensional rotations of the joints of the upper limb. Acknowledgements We would like to thank Stephen Gelbart, Dragana Todoric and Yosef Yomdin for instructive conversations on the mathematical background and Dario Liebermann for fruitful discussions. Special thanks go to Stan Gielen for conversations which initiated this work. References Choquet-Bruhat Y., De Witt-Morette C. & Dillard-Bleick M., Analysis, Manifolds and Physics, North-Holland (1982). Gilmore R.,LieGroups, Lie Algebras, and Some of Their Applications, Wiley (1974). Hepp K., Commun. Math. Phys. 132 (1990) 285-292. Hestenes D., Neural Networks 7, No.1 (1994) 65-77. Minken A.W.H. Van Opstal A.J. & Van Gisbergen J.A.M., Exp. Brain Research 93 (1993) 521-533. Tweed, D. & Vilis T., J. Neurophysiology 58 (1987) 832-849. Tweed D. & Vilis T., Vision Research 30 (1990) 111-127. Van Opstal J., "Representations of Eye Positions in Three Dimensions", in Multisensory Control of Movement, ed. Berthoz A., (1993) 27-4l. Van Rijn L.J. & Van Den Berg A.V., Vision Research 33, No. 5/6 (1993) 691-708. Varadarajan V.S., Lie Groups, Lie Algebras, and Their Reps., Prentice-Hall (1974). Westheimer G., Journal of the Optical Society of America 47 (1957) 967-974.
|
1995
|
128
|
1,030
|
When is an Integrate-and-fire Neuron like a Poisson Neuron? Charles F. Stevens Salk Institute MNL/S La Jolla, CA 92037 cfs@salk.edu Anthony Zador Salk Institute MNL/S La Jolla, CA 92037 zador@salk.edu Abstract In the Poisson neuron model, the output is a rate-modulated Poisson process (Snyder and Miller, 1991); the time varying rate parameter ret) is an instantaneous function G[.] of the stimulus, ret) = G[s(t)]. In a Poisson neuron, then, ret) gives the instantaneous firing rate-the instantaneous probability of firing at any instant t-and the output is a stochastic function of the input. In part because of its great simplicity, this model is widely used (usually with the addition of a refractory period), especially in in vivo single unit electrophysiological studies, where set) is usually taken to be the value of some sensory stimulus. In the integrate-and-fire neuron model, by contrast, the output is a filtered and thresholded function of the input: the input is passed through a low-pass filter (determined by the membrane time constant T) and integrated until the membrane potential vet) reaches threshold 8, at which point vet) is reset to its initial value. By contrast with the Poisson model, in the integrate-and-fire model the ouput is a deterministic function of the input. Although the integrate-and-fire model is a caricature of real neural dynamics, it captures many of the qualitative features, and is often used as a starting point for conceptualizing the biophysical behavior of single neurons. Here we show how a slightly modified Poisson model can be derived from the integrate-and-fire model with noisy inputs yet) = set) + net). In the modified model, the transfer function G[.] is a sigmoid (erf) whose shape is determined by the noise variance /T~. Understanding the equivalence between the dominant in vivo and in vitro simple neuron models may help forge links between the two levels. 104 c. F. STEVENS. A. ZADOR 1 Introduction In the Poisson neuron model, the output is a rate-modulated Poisson process; the time varying rate parameter ret) is an instantaneous function G[.] of the stimulus, ret) = G[s(t)]. In a Poisson neuron, then, ret) gives the instantaneous firing rate-the instantaneous probability of firing at any instant t-and the output is a stochastic function of the input. In part because of its great simplicity, this model is widely used (usually with the addition of a refractory period), especially in in vivo single unit electrophysiological studies, where set) is usually taken to be the value of some sensory stimulus. In the integrate-and-fire neuron model, by contrast, the output is a filtered and thresholded function of the input: the input is passed through a low-pass filter (determined by the membrane time constant T) and integrated until the membrane potential vet) reaches threshold 0, at which point vet) is reset to its initial value. By contrast with the Poisson model, in the integrate-and-fire model the ouput is a deterministic function of the input. Although the integrate-and-fire model is a caricature of real neural dynamics, it captures many of the qualitative features, and is often used as a starting point for conceptualizing the biophysical behavior of single neurons (Softky and Koch, 1993; Amit and Tsodyks, 1991; Shadlen and Newsome, 1995; Shadlen and Newsome, 1994; Softky, 1995; DeWeese, 1995; DeWeese, 1996; Zador and Pearlmutter, 1996). Here we show how a slightly modified Poisson model can be derived from the integrate-and-fire model with noisy inputs yet) = set) + net). In the modified model, the transfer function G[.] is a sigmoid (erf) whose shape is determined by the noise variance (j~ . Understanding the equivalence between the dominant in vivo and in vitro simple neuron models may help forge links between the two levels. 2 The integrate-and-fire model Here we describe the the forgetful leaky integrate-and-fire model. Suppose we add a signal set) to some noise net), yet) = net) + set), and threshold the sum to produce a spike train z(t) = F[s(t) + net)], where F is the thresholding functional and z(t) is a list of firing times generated by the input. Specifically, suppose the voltage vet) of the neuron obeys vet) = - vet) + yet) (1) T where T is the membrane time constant. We assume that the noise net) has O-mean and is white with variance (j~. Thus yet) can be thought of as a Gaussian white process with variance (j~ and a time-varying mean set). If the voltage reaches the threshold 00 at some time t, the neuron emits a spike at that time and resets to the initial condition Vo. This is therefore a 5 parameter model: the membrane time constant T, the mean input signal Il, the variance of the input signal 172 , the threshold 0, and the reset value Vo. Of course, if net) = 0, we recover a purely deterministic integrate-and-fire model. When Is an Integrate-and-fire Neuron like a Poisson Neuron? 105 In order to forge the link between the integrate-and-fire neuron dynamics and the Poisson model, we will treat the firing times T probabilistically. That is, we will express the output of the neuron to some particular input set) as a conditional distribution p(Tls(t», i.e. the probability of obtaining any firing time T given some particular input set). Under these assumptions, peT) is given by the first passage time distribution (FPTD) of the Ornstein-Uhlenbeck process (Uhlenbeck and Ornstein, 1930; Tuckwell, 1988). This means that the time evolution of the voltage prior to reaching threshold is given by the Fokker-Planck equation (FPE), 8 u; 82 8 vet) 8t g(t, v) = 2 8v2 get, v) - av [(set) - --;:- )g(t, v)], (2) where uy = Un and get, v) is the distribution at time t of voltage -00 < v ::; (}o. Then the first passage time distribution is related to g( v, t) by 81 90 peT) = - at -00 get, v)dv. (3) The integrand is the fraction of all paths that p.ave not yet crossed threshold. peT) is therefore just the interspike interval (lSI) distribution for a given signal set). A general eigenfunction expansion solution for the lSI distribution is known, but it converges slowly and its terms offer little insight into the behavior (at least to us). We now derive an expression for the probability of crossing threshold in some very short interval ~t, starting at some v. We begin with the "free" distribution of g (Tuckwell, 1988): the probability of the voltage jumping to v' at time t' = t + ~t, given that it was at v at time t, assuming von Neumann boundary conditions at plus and minus infinity, get', v'lt, v) = exp y, 1 [ (v' - m( ~t; u »)2] J27r q(~t;Uy) 2 q(~t;Uy) (4) with and m(~t) = ve-at/ T + set) * T(l _ e-at/ T ), where * denotes convolution. The free distribution is a Gaussian with a timedependent mean m(~t) and variance q(~t; uy). This expression is valid for all ~t. The probability of making a jump ~v = v' - v in a short interval ~t ~ T depends only on ~v and ~t, ga(~t, ~v; uy) = 1 exp [_ ~~2 )]. ..j27r qa(uy) 2 qa uy (5) For small ~t, we expand to get qa(uy) :::::: 2u;~t, which is independent of T, showing that the leak can be neglected for short times. 106 c. F. STEVENS, A. ZADOR Now the probability Pt>, that the voltage exceeds threshold in some short Ilt, given that it started at v, depends on how far v is from threshold; it is Thus Pr[v + Ilv ~ 0] = Pr[llv ~ 0 - v]. (Xl dvgt>,(llt, v; O"y) J9-v -erfc 1 (o-v) 2 J2qt>,(O"y) -erfc 1 (o-v) 2 O"yJ21lt (6) where erfc(x) = 1 - -j; I; e-t~ dt goes from [2 : 0]. This then is the key result: it gives the instantaneous probability of firing as a function of the instantaneous voltage v. erfc is sigmoidal with a slope determined by O"y, so a smaller noise yields a steeper (more deterministic) transfer function; in the limit of 0 noise, the transfer function is a step and we recover a completely deterministic neuron. Note that Pt>, is actually an instantaneous function of v(t), not the stimulus itself s(t). If the noise is large compared with s(t) we must consider the distribution g$ (v, t; O"y) of voltages reached in response to the input s(t): Py(t) (7) 3 Ensemble of Signals What if the inputs s(t) are themselves drawn from an ensemble? If their distribution is also Gaussian and white with mean Jl and variance 0";, and if the firing rate is low (E[T] ~ T), then the output spike train is Poisson. Why is firing Poisson only in the slow firing limit? The reason is that, by assumption, immediately following a spike the membrane potential resets to 0; it must then rise (assuming Jl > 0) to some asymptotic level that is independent of the initial conditions. During this rise the firing rate is lower than the asymptotic rate, because on average the membrane is farther from threshold, and its variance is lower. The rate at which the asymptote is achieved depends on T. In the limit as t ~ T, some asymptotic distribution of voltage qoo(v), is attained. Note that if we make the reset Vo stochastic, with a distribution given by qoo (v), then the firing probability would be the same even immediately after spiking, and firing would be Poisson for all firing rates. A Poisson process is characterized by its mean alone. We therefore solve the FPE (eq. 2) for the steady-state by setting °tg(t, v) = 0 (we consider only threshold crossings from initial values t ~ T; negYecting the early events results in only a small error, since we have assumed E{T} ~ T). Thus with the absorbing boundary When Is an Integrate-and-fire Neuron like a Poisson Neuron? 107 at 0 the distribution at time t ~ T (given here for JJ = 0) is g~(Vj uy) = kl (1 - k2erfi [uyfi]) exp [~i:] , (8) where u; = u; + u~, erfi(z) = -ierf(iz), kl determines the normalization (the sign of kl determines whether the solution extends to positive or negative infinity) and k2 = l/erfi(O/(uy Vr)) is determined by the boundary. The instantaneous Poisson rate parameter is then obtained through eq. (7), (9) Fig. 1 tests the validity of the exponential approximation. The top graph shows the lSI distribution near the "balance point" , when the excitation is in balance with the inhibition and the membrane potential hovers just subthreshold. The bottom curves show the lSI distribution far below the balance point. In both cases, the exponential distribution provides a good approximation for t ~ T. 4 Discussion The main point of this paper is to make explicit the relation between the Poisson and integrate-and-fire models of neuronal acitivity. The key difference between them is that the former is stochastic while the latter is deterministic. That is, given exactly the same stimulus, the Poisson neuron produces different spike trains on different trials, while the integrate-and-fire neuron produces exactly the same spike train each time. It is therefore clear that if some degree of stochasticity is to be obtained in the integrate-and-fire model, it must arise from noise in the stimulus itself. The relation we have derived here is purely formalj we have intentionally remained agnostic about the deep issues of what is signal and what is noise in the inputs to a neuron. We observe nevertheless that although we derive a limit (eq. 9) where the spike train of an integrate-and-fire neuron is a Poisson process-i.e. the probability of obtaining a spike in any interval is independent of obtaining a spike in any other interval (except for very short intervals )-from the point of view of information processing it is a very different process from the purely stochastic rate-modulated Poisson neuron. In fact, in this limit the spike train is deterministically Poisson if u y = u., i. e. when n( t) = OJ in this case the output is a purely deterministic function of the input, but the lSI distribution is exponential. 108 C. F. STEVENS, A. ZADOR References Amit, D. and Tsodyks, M. (1991). Quantitative study of attractor neural network retrieving at low spike rates. i. substrate-spikes, rates and neuronal gain. Network: Computation in Neural Systems, 2:259-273. DeWeese, M. (1995). Optimization principles for the neural code. PhD thesis, Dept of Physics, Princeton University. DeWeese, M. (1996). Optimization principles for the neural code. In Hasselmo, M., editor, Advances in Neural Information Processing Systems, vol. 8. MIT Press, Cambridge, MA. Shadlen, M. and Newsome, W. (1994). Noise, neural codes and cortical organization. Current Opinion in Neurobiology, 4:569-579. Shadlen, M. and Newsome, W. (1995). Is there a signal in the noise? [comment]. Current Opinion in Neurobiology, 5:248-250. Snyder, D. and Miller, M. (1991). Random Point Processes in Time and Space, 2nd edition. Springer-Verlag. Softky, W. (1995). Simple codes versus efficient codes. Current Opinion in Neurobiology, 5:239-247. Softky, W. and Koch, C. (1993). The highly irregular firing of cortical cells is inconsistent with temporal integration of random epsps. J. Neuroscience., 13:334-350. Tuckwell, H. (1988). Introduction to theoretical neurobiology (2 vols.). Cambridge. Uhlenbeck, G. and Ornstein, L. (1930). On the theory of brownian motion. Phys. Rev., 36:823-84l. Zador, A. M. and Pearlmutter, B. A. (1996). VC dimension of an integrate and fire neuron model. Neural Computation, 8(3). In press. When Is an Integrate-and-fire Neuron like a Poisson Neuron? 109 lSI distributions at balance point and the exponential limit 0.02 0.015 .~ 15 0.01 .8 e a. 0.005 50 100 150 200 250 300 350 400 450 500 Time (msec) 2 x 10-3 1.5 ~ ~ 1 .0 0 ... a. 0.5 0 0 200 400 600 800 1000 1200 1400 1600 1800 2000 lSI (msec) Figure 1: lSI distributions. (A; top) lSI distribution for leaky integrate-and-fire model at the balance point, where the asymptotic membrane potential is just subthreshold, for two values of the signal variance (1'2 . Increasing (1'2 shifts the distribution to the left. For the left curve, the parameters were chosen so that E{T} ~ T, giving a nearly exponential distribution; for the right curve, the distribution would be hard to distinguish experimentally from an exponential distribution with a refractory period. (T = 50 msec; left: E{T} = 166 msec; right: E{T} = 57 msec). (B; bottom) In the subthreshold regime, the lSI distribution (solid} is nearly exponential (dashed) for intervals greater than the membrane time constant. (T = 50 msec; E{T} = 500 msec)
|
1995
|
129
|
1,031
|
A Novel Channel Selection System in Cochlear Implants Using Artificial Neural Network Marwan A. Jabri & Raymond J. Wang Systems Engineering and Design Automation Laboratory Department of Electrical Engineering The University of Sydney NSW 2006, Australia {marwan,jwwang}Osedal.usyd.edu.au Abstract State-of-the-art speech processors in cochlear implants perform channel selection using a spectral maxima strategy. This strategy can lead to confusions when high frequency features are needed to discriminate between sounds. We present in this paper a novel channel selection strategy based upon pattern recognition which allows "smart" channel selections to be made. The proposed strategy is implemented using multi-layer perceptrons trained on a multispeaker labelled speech database. The input to the network are the energy coefficients of N energy channels. The output of the system are the indices of the M selected channels. We compare the performance of our proposed system to that of spectral maxima strategy, and show that our strategy can produce significantly better results. 1 INTRODUCTION A cochlear implant is a device used to provide the sensation of sound to those who are profoundly deaf by means of electrical stimulation of residual auditory neurons. It generally consists of a directional microphone, a wearable speech processor, a head-set transmitter and an implanted receiver-stimulator module with an electrode A Novel Channel Selection System in Cochlear Implants 911 array which all together provide an electrical representation of the speech signal to the residual nerve fibres of the peripheral auditory system (Clark et ai, 1990). Electrode Array Figure 1: A simplified schematic diagram ofthe cochlear implants Brain A simplified schematic diagram of the cochlear implants is shown in Figure 1. Speech sounds are picked up by the directional microphone and sent to the speech processor. The speech processor amplifies, filters and digitizes these signals, and then selects and codes the appropriate sound information. The coded signal contains information as to which electrode to stimulate and the intensity level required to generate the appropriate sound sensations. The signal is then sent to the receiver/stimulator via the transmitter coil. The receiver/stimulator delivers electrical impulses to the appropriate electrodes in the cochlea. These stimulated electrodes then directly activate the hearing nerve in the inner ear, creating the sensation of sound, which is then forwarded to the brain for interpretation. The entire process happens in milliseconds. For multi-channel cochlear implants, the task of the speech processor is to compute the spectral energy of the electrical signals it receives, and to quantise them into different levels. The energy spectrum is commonly divided into separate bands using a filter bank of N (typically 20) bandpass filters with centre frequencies ranging from 250 Hz to 10 KHz. The bands of energy are allocated t.o electrodes in the patient's implant on a one-to-one basis. Usually the most-apical bipolar electrode pairs are allocated to the channels in tonotopic order. The limitations of implant systems usually require only a selected number of the quantised energy levels to be fed to the implanted electrode array (Abbas, 1993; Schouten, 1992). The state-of-the-art speech processor for multi-channel implants performs channel selection using spectral maxima strategy (McDermott et ai, 1992; Seligman & McDermott, 1994). The maxima strategy selects the M (about 6) largest spectral energy of the frequency spectrum as stimulation channels from a filter bank of N (typically 20) bandpass. It is believed that compared to other channel selection techniques (FOF2, FOF1F2, MPEAK ... ), the maxima strategy increases the amount of spectral information and improves the speech perception and recognition performance. However, maxima strategy relies heavily on the highest energies. This often leads to the same levels being selected for different sounds, as the energy levels that distinguish them are not high enough to be selected. For some speech signals, 912 M. A. JABRI, R. J. WANG it does not cater for confusions and cannot discriminate between high frequency features. We present in this paper Artificial Neural Networks (ANN) techniques for implementing "smart" channel selection for cochlear implant systems. The input to the proposed channel selection system consists of the energy coefficients (18 in our experiments) and the output the indices of the selected channels (6 in our experiments). The neural network based selection system is trained on a multi-speaker labelled speech and has been evaluated on a separated multi-speaker database not used in the training phase. The most important feature of our ANN based channel selection system is its ability to select the channels for stimulation on the basis of the overall morphology of the energy spectrum and not only on the basis of the maximal energy values. 2 THE PATTERN RECOGNITION BASED CHANNEL SELECTION STRATEGY Speech is the most natural form of human communication. The speech information signal can be divided into phonemes, which share some common acoustic properties with one another for a short interval of time. The phonemes are typically divided into two broad classes: (a) vowels, which allow unrestricted airflow in the vocal tract, and (b) consonants, which restrict airflow at some point and are weaker than vowels. Different phonemes have different morphology in the energy spectrum. Moreover, for different speakers and different speech sentences, the same phonemes have different energy spectrum morphologies (Kent & Read, 1992). Therefore, simple methods to select some of the most important channels for all the phoneme patterns will not perform as good as the method that considers the spectrum in its entirety. The existing maxima strategy only refers to the spectrum amplitudes found in the entire estimated spectrum without considering the morphology. Typically several of the maxima results can be obtained from a single spectral peak. Therefore, for some phoneme patterns, the selection result is good enough to represent the original phoneme. But for some others, some important features of the phoneme are lost. This usually happens to those phonemes with important features in the high frequency region. Due to the low amplitude of the high frequency in the spectrum morphology, maxima methods are not capable to extract those high frequency features. The relationship between the desired M output channels and the energy spectrum patterns is complex, and depending on the conditions, may be influenced by many factors. As mentioned in the Introduction, channel selection methods that make use of local information only in the energy spectrum are bound to produce channel sub-sets where sounds may be confused. The confusions can be reduced if "global" information of the energy spectrum is used in the selection process. The channel selection approach we are proposing makes use of the overall energy spectrum. This is achieved by turning the selection problem into that of a spectrum morphology pattern recognition one and hence, we call our approach Pattern Recognition based Channel Selection (PRCS). A Novel Channel Selection System in Cochlear Implants 913 2.1 PRCS STRATEGY The PRCS strategy is implemented using two cascaded neural networks shown in Figure 2: • Spectral morphological classifier: Its inputs are the spectrum energy amplitudes of all the channels and its outputs all the transformations of the inputs. The transformation between input and out.put can be seen as a recognition, emphasis, and/or decaying of the inputs. The consequence is that some inputs are amplified and some decayed, depending on the morphology of the spectrum. The classifier performs a non-linear mapping . • M strongest of N classifier: It receives the output of morphological classifier and applies a M strongest selection rule. C21 .. 'IR. ---Spectral Morphological CI ..... .., • • • ----- Labeia MStrongaat ofN CIanIf.., Figure 2: The pattern recognition based channel selection architecture 2.2 TRAINING AND TESTING DATA The most difficult task in developing the proposed PRCS is to set up the labelled training and testing data for the spectral morphological classifier. The training and testing data sets have been constructed using the process shown in Figure 3. r " Hlmmlng 18Ch8nnela Training Window + r-- Quantlsatlor Chlinnel r-&Teatlng 128 FFT & scaling labelling Sets 'Figure 3: The process of generating training and testing sets The sounds in the data sets are speech extracted from the DARPA TIMIT multispeaker speech corpus (Fisher et ai, 1987) which contains a total of 6300 sentences, 10 sentences spoken by each of 630 speakers. The speech signal is sampled at 16KHz rate with 16 bit precision. As the speech is nonstationary, to produce the energy spectrum versus channel numbers, a short-time speech analysis method is used. The Fast Fourier Transform with 8ms smooth Hamming window technique is applied to yield the energy spectrum. The hamming window has the shape of a raised 914 cosine pulse: h( n) = { ~.54 - 0.46 cos (J~n. ) M. A. JABRI, R. J. WANG for 0 ~ n ~ N-l otherwise The time frame on which the speech analysis is performed is 4ms long and the successive time frame windows overlap by 50%. Using frequency allocations similar to that used in commercial cochlear implant speech processors, the frequency range in the spectrum is divided into 18 channels with each channel having the center frequencies of 250, 450, 650, 850 1050, 1250, 1450, 1650, 1895, 2177, 2500, 2873, 3300, 3866, 4580, 5307, 6218 and 7285Hz respectively. Each energy spectrum from a time frame is quantised into these 18 frequency bands. The energy amplitude for each level is the sum of the amplitude value of the energy for all the frequency components in the level. The quantised energy spectrum is then labelled using a graphics based tool, called LABEL, developed specially for this application. LABEL displays the spectrum pattern including the unquantised spectrum, the signal source, speaker's name, speech sentence, phoneme, signal pre-processing method and FFT results. All these information assists labelling experts to allocate a score (1 to 18) to each channel. The score reflects the importance of the information provided by each of the bands. Hence, if six channels are only to be selected, the channels with the score 1 to 6 can be used and are highlighted. The labelling is necessary as a supervised neural network training method is being used. A total of 5000 energy spectrum patterns have been labelled. They are from 20 different speakers and different. spoken sentences. Of the 5000 example patterns, 4000 patterns are allocated for training and 1000 patterns for testing. 3 EXPERIMENTAL RESULTS We have implemented and tested the PH.CS system as described above and our experiments show that it has better performance than channel selection systems used in present cochlear implant processors. The PRCS system is effectively constructed as a multi-module neural network using MUME (Jabri et ai, 1994). The back-propagation algorithm in an on-line mode is used to train the MLP. The training patterns input components are the energy amplitudes of the 18 channels and the teacher component consists of a "I" for a channel to be selected and "0" for all others. The MLP is trained for up to 2000 epochs or when a minimum total mean squared error is reached. A learning rate 7J of 0.01 is used (no weight decay). We show the average performance of our PRCS in Table 1 where we also show the performance of a leading commercial spectral maxima strategy called SPEAK on the same test set. In the first column of this table we show the number of channels that matched out of the 6 desired channels. For example, the first row corresponds to the case where all 6 channels match the desired 6 channels in the test data base, and so on. As Table 1 shows, the PRCS produces a significantly better performance than the commercial strategy on the speech test set. The selection performance to different phonemes is listed in Table 2. It clearly A Novel Channel Selection System in Cochlear Implants 915 Table 1: The comparison of average performance between commercial and PRCS system II II The Channel Selections from the two different methods II PRCS results Commercial technique results Fully matched 22 % 4% 5 matched 80 % 25 % 4 matched 98 % 57 % 3 matched 100 % 93 % 2 matched 100 % 99 % 1 matched 100 % 100 % Table 2: PRCS channel selecting performance on different phoneme patterns The P RCS results for different phoneme patterns Phoneme Fully matched 5 matched 4 matched 3 matched Stops 19 % 69 % 96 % 100 % Fricatives 18 % 66 % 92 % 100 % Nasals 14 % 66 % 96 % 100 % Semivowels & Glides 14 % 79 % 95 % 100% Vowels 25 % 84 % 98 % 100 % \I shows that the PRCS strategy can cater for the features of all the speech spectrum patterns. To compare the practical performance of the PRes with the maxima strategies we have developed a direct performance test system which allows us to play the synthesized speech of the selected channels through post-speech synthesizer. Our test shows that the PRCS produces more intelligible speech to the normal ears. Sixteen different sentences spoken by sixteen people are tested using both maxima and PRCS methods. It is found that the synthesized speech from PRCS has much more high frequency features than that of the speech produced by the maxima strategy. All listeners who were asked to take the test agreed that the quality of the speech sound from PRCS is much better than those from the commercial maxima channel selection system. The tape recording of the synthesized speech will be available at the conference. 4 CONCLUSION A pattern recognition based channel selection strategy for Cochlear Implants has been presented. The strategy is based on a 18-72-18 MLP strongest selector. The proposed channel selection strategy has been compared to a leading commercial technique. Our simulation and play back results show that our machine learning based technique produces significantly better channel selections. 916 M. A. JABRI, R. J. WANG Reference Abbas, P. J. (1993) Electrophysiology. "Cochlear Implants: Audiological Foundations" edited by R. S. Tyler, Singular Publishing Group, pp.317-355. Clark, G. M., Tong, Y. C.& Patrick, J. F. (1990) Cochlear Prosthesis. Edi n borough: Churchill Living stone. Fisher, W. M., Zue, V., Bernstein, J. & Pallett, D. (1987) An Acoustic-Phonetic Data Base. In 113th Meeting of Acoust Soc Am, May 1987 Jabri, M. A., Tinker, E. A. & Leerink, L. (1994) MUME A Multi-Net MultiArchitecture Neural Simulation Environment. "Neural Network Simulation Environments", J. Skrzypek ed., Kluwer Academic Publishers. Kent, R. D. & Read, C. (1992) The Acoustic Analysis of Speech. Whurr Publishers. McDermott, H. J., McKay, C. M. & Vandali, A. E. (1992) A new portable sound processor for the University of Melbourne / Nucleus Limited multielectrode cochlear implant. J. Acoust. Soc. Am. 91(6), June 1992, pp.3367-3371 Schouten, M. E. H edited (1992) The Auditory Processing of Speech From Sounds to Words. Speech Research 10, Mouton de Groyter. Seligman, P. & McDermott, H. (1994) Architecture of the SPECTRA 22 Speech Processor. International Cochlear Implant, Speech and Hearing Symposium, Melbourne, October, 1994, p.254.
|
1995
|
13
|
1,032
|
Examples of learning curves from a modified VC-formalism. A. Kowalczyk & J. Szymanski Telstra Research Laboratories 770 Blackbtun Road, Clayton, Vic. 3168, Australia {akowalczyk,j.szymanski }@trl.oz.au) P.L. Bartlett & R.C. Williamson Department of Systems Engineering Australian National University Canberra, ACT 0200, Australia {bartlett, williams }@syseng.anu.edu.au Abstract We examine the issue of evaluation of model specific parameters in a modified VC-formalism. Two examples are analyzed: the 2-dimensional homogeneous perceptron and the I-dimensional higher order neuron. Both models are solved theoretically, and their learning curves are compared against true learning curves. It is shown that the formalism has the potential to generate a variety of learning curves, including ones displaying ''phase transitions." 1 Introduction One of the main criticisms of the Vapnik-Chervonenkis theory of learning [15] is that the results of the theory appear very loose when compared with empirical data. In contrast, theory based on statistical physics ideas [1] provides tighter numerical results as well as qualitatively distinct predictions (such as "phase transitions" to perfect generalization). (See [5, 14] for a fuller discussion.) A question arises as to whether the VC-theory can be modified to give these improvements. The general direction of such a modification is obvious: one needs to sacrifice the universality of the VC-bounds and introduce model (e.g. distribution) dependent parameters. This obviously can be done in a variety of ways. Some specific examples are VC-entropy [15], empirical VC-dimensions [16], efficient complexity [17] or (p., C)-uniformity [8, 9] in a VC-formalism with error shells. An extension of the last formalism is of central interest to this paper. It is based on a refinement of the "fundamental theorem of computational learning" [2] and its main innovation is to split the set of partitions of a training sample into separate "error shells", each composed of error vectors corresponding to the different error values. Such a split introduces a whole range of new parameters (the average number of patterns in each of a series of error shells) in addition to the VC dimension. The difficulty of determining these parameters then arises. There are some crude, "obvious" upper bounds Examples of Learning Curves from a Modified VC-fonnalism 345 on them which lead to both the VC-based estimates [2, 3, 15] and the statistical physics based formalism (with phase transitions) [5] as specific cases of this novel theory. Thus there is an obvious potential for improvement of the theory with tighter bounds. In particular we find that the introduction of a single parameter (order of uniformity), which in a sense determines shifts in relative sizes of error shells, leads to a full family of shapes of learning curves continuously ranging in behavior from decay proportional to the inverse of the training sample size to "phase transitions" (sudden drops) to perfect generalization in small training sample sizes. We present initial comparison of the learning curves from this new formalism with "true" learning curves for two simple neral networks. 2 Overview of the formalism The presentation is set in the typical PAC-style; the notation follows [2]. We consider a space X of samples with a probability measure J1., a subspace H of binary functions X -+ {O, 1} (dichotomies) (called the hypothesis space) and a target hypothesis t E H. Foreachh E H andeachm-samplez = (:el, ... , :em) E xm (m E {1, 2, ... }),wedenoteby €h,z d;j ~ E::llt-hl(:ei)theempiricalerrorofhonz,andbY€h d;j fx It- h l(:e)J1.(d:e) the expected error of h E H. For each m E {1, 2, ... } let us consider the random variable maa: (-) de! { O} € H :l: = max €h j €h z = hEH ' (1) defined as the maximal expected error of an hypothesis h E H consistent with t on z. The learning curve of H, defined as the expected value of tJiaa: , €j{(m) d;j Exm.[€Jiaa:] = f €Jiaa: (z)Jr (dz) (z E xm) (2) Jx = is of central interest to us. Upper bounds on it can be derived from basic PAC-estimates as de! follows. For € ~ ° we denote by HE = {h E H j €h ~ €} the subset of €-bad hypotheses and by Q;! d;j {z E Xm j 3hE H. €h,ri = O} = {z E Xm j 3hEH €h,ri = ° & €h ~ €} (3) the subset of m-samples for which there exists an €-bad hypothesis consistent with the target t. Lemmal IfJ1.m(Q;!) ~ 1J!(€,m), then€j{(m) ~ folmin(l,1J!(€,m))J1.(d€), and equality in the assumption implies equality in the conclusion. 0 Proof outline. If the assumption holds, then 'lr(€, m) d~ 1 - min(l, 1J!( €, m)) is a lower bound on the cumulative distribution of the random variable (1). Thus E x= [€Jiaa:] ~ f01 € tE 'lr( €, m)d€ and integration by parts yields the conclusion. o Givenz = (:el, ... ,:em) E Xm,letusintroducethetransformation(projection)1rt,ri: H-+ {O, l}m allocating to each h E H the vector 1rt,:i!(h) d;j (Ih(:el) - t(:el)l, ... , Ih(:em) - t(:em)l) called the error pattern of h on z. For a subset G C H, let 1rt,:i!(G) = {1rt,:i!(h) : hE G}. The space {o,l}m is the disjoint union of error shells £i d~ {(el, ... ,em) E {O,l}m j el + ... + em = i} for i = 0,1, ... , m, and l1rt,ri(HE) n £il is the number 346 A. KOWALCZYK, J. SZYMANSKI, P. L. BARTLETT, R. C. WILLIAMSON of different error patterns with i errors which can be obtained for h E HE' We shall emplOy the following notation for its average: IHEli d~ Ex ... [l1I't,z(HE) n t:in = r l'II't,z(HE) n t:ilJ.£m(dz). (4) Jx ... The central result of this paper, which gives a bOlUld on the probability of the set Qr;' as in Lemma 1 in terms of I HE Ii, will be given now. It is obtained by modification of the proof of [8, Theorem 1] which is a refinement of the proof of the ''ftmdamental theorem of computational learning" in [2]. It is a simplified version (to the consistent learning case) of the basic estimate discussed in [9, 7]. Theorem 2 For any integer Ie ~ 0 and 0 ::; E, 'Y ::; 1 I-'m(Q';")::; A f ,k,7 t (~) (m:- 1e)-lIHElj+A:, j~7k J J (5) whereA E,k,7 d~ (1- E}~~ O)Ej(l-E)k-j) -l,forle > OandA E,o,7 d~ 1.0 Since error shells are disjoint we have the following relation: PH(m) d~ 2-m i_I".(H)I!r(dZ) = 2-m t.IHli ~ IIH(m)/2m (6) where 1I'z(h) d~ 1I'0,z(h), IHli d~ IHoli and IIH(m) d~ maxz E x'" I 'll'z (H) I is the growth function [2] of H. (Note that assuming that the target t == 0 does not affect the cardinality of 1I't,z(H).) If the VC-dimension of H, d = dvc(H), is finite, we have the well-known estimate [2] IIH(m)::; ~(d,m) d~ t (rr:) ::; (em/d)d. j=O J (7) Corollary 3 (i) If the VC-dimension d of H is finite and m > 8/E, then J.£m(Qr;') ::; 22- mE/ 2(2em/ d)d. (ii) If H has finite cardinality, then J.£m (Qr;') ::; EhEH. (1 - Eh)m. Proof. (i) Use the estimate AE,k,E/2 ::; 2 for Ie ~ 8/E resulting from the Chernoff bound and set'Y = E /2 and Ie = m in (5). (ii) Substitute the following crude estimate: m m IHEli ::; L IHEli ::; L IHli ::; PH ::; (em/d)d, i=O i=O into the previous estimate. (iii) Set Ie = 0 into (i) and use the estimate IHli::; L Prx ... (Eh,z = i/m) = L (1- Eh)m-iEhi. 0 The inequality in Corollary 3.i (ignoring the factor of 2) is the basic estimate of the VCformalism (c.f. [2]); the inequality in Corollary 3.ii is the union bound which is the starting point for the statistical physics based formalism developed in [5]. In this sense both of these theories are unified in estimate (5) and all their conclusions (including the prediction Examples of Learning Curves from a Modified VC-formalism 347 100 (a) (b) , \ \ \ I I I I I 10- 1 I I I I I I I I I I I \ CJ = 3 : chain line 10-2 ...... _-' .... CJ = 3 and COl. = 1 : broken line 10-2 '-'-"'--'-.L...l.....L...l.....L...l....~~.L...l.....L...l.....L...l.....L...l.....L...l.....L...J 4 5 6 7 8 9 o 10 20 30 40 50 mid mi d Figure 1: (a) Examples of upper bounds on the learning curves for the case of finite VCdimension d = dvc(H) implied by Corollary 4.ii for Cw,m == const. They split into five distinct "bands" of four curves each, according to the values of the order of uniformity w = 2, 3,4,5, 10 (in the top-down order). Each band contains a solid line (Cw,m == 1, d = 100), a dotted line (Cw,m == 100, d = 100), a chain line (Cw,m == 1, d = 1000) and a broken line (Cw,m == 100, d = 1000). (b) Various learning curves for the 2-dimensional homogeneous perceptron. Solid lines (top to bottom): (i) - for the VC-theory bound (Corollary 3.ii) with VC-dimension d = 2; (ii) - for the bound (for Eqn, 5 and Lemma 1) with'Y = f, k = m and the upper bounds IHElr ~ IHlr = 2 for i = 1, " " m - 1 and IHElr ~ IHlr = 1 for i = 0, m ; (iii) - as in (ii) but with the exact values for IH Elr as in (11); (iv) - true learning curve (Eqn. 13). The w-uniformity bound for w = 2 (with the minimal C w,m satisfying (9), which turn out to be = const = 1) is shown by dotted line; for w = 3 the chain line gives the result for minimal Cw m and the broken line for Cw m set to 1. , , of phase transitions to perfect generalization for the Ising perceptron for a = mj d < 1.448 in the thermodynamic limit [5]) can be derived from this estimate, and possibly improved with the use of tighter estimates on IH E Ir. We now formally introduce a family of estimates on IHElr in order to discuss a potential of our formalism. For any m, f and w ~ 1.0 there exists Cw,m > 0 such that IH.lr s: IHlr s: Cw,m (7) PH(m)l-ll-2i/ml'" (for 0 s: i ~ m), (8) We shall call such an estimate an w-uniformity bound. Corollary 4 (i) If an w -lllliformity bolllld (8) holds, then ILm(Qm) < A C ~ (m)PH (2m)l-ll-j/ml"', rE _ Elm • .., W,m ~ . , j~",m J (9) (ii) if additionallyd = dvc(H) < 00, then m ( ) m m m 2m d l-Il-j/ml'" J1- (Q.) s: A.,m,,,,Cw,m L . (T (2emjd)) . 0 j~",m J (10) 3 Examples of learning curves In this section we evaluate the above formalism on two examples of simple neural networks. 348 A. KOWALCZYK, J. SZYMANSKI, P. L. BARTLETT, R. C. WILLIAMSON 20 I I I I (b) -'~ -' 15 f_.-r;-"= 3 u 2 10 ." <III " ------------. ..2 / C'oj = 2 / 10- 1 / ", 5r/ '/ J~ ~ C'oj = 2 : dolled line I,' Col = 3: chain line ,/ ' 0 I' I I I I 0 10 20 30 40 50 0 100 200 300 400 500 m/(d+l) m Figure 2: (a) Different learning curves for the higher order neuron (analogous to Fig. l.b). Solid lines (top to bottom)( i) - forthe VC-theory bound (Corollary 3.ii) with VC-dimension d + 1 = 21; (ii) - for the bound (5) with 'Y = € and the upper bounds I H E Ii ~ I H Ii with IHli given by (15); (iii) - true learning curve (the upper bound given by (18)). The wuniformity bound/approximation are plotted as chain and dotted lines for the minimal C w,m satisfying (8), and as broken (long broken) line for C w,m = const = 1 with w = 2 (w = 3). (b) Plots of the minimal value of Cw,m satisfying condition of w-uniformity bound (8) for higher order neuron and selected values of w. 3.1 2-dimensional homogeneous perceptron We consider X d.~ R2 and H defined as the family of all functions (el, 6) ~ 8(el Wl + 6W2)' where (Wl, W2) E R2 and 8(r) is defined as 1 if r ~ 0 and 0, otherwise, and the probability measure jJ. on R2 has rotational symmetry with respect to the origin. Fix an arbitrary target t E H . In such a case IH I~= 1 { 2(1 - €)m - (1 - 2€)m E, . () 22:;=0 j €i (1- €)m-; (for i = 0 and 0 ~ € ~ 1/2), (fori = m), ( otherwise). In particular we find that IHli = 1 for i = 0, m and IHli = 2, otherwise, and m (11) PH(m) = L IHli /2m = (1 + 2 + ... + 2 + 1)/2m = m/2m - l . (12) and the true learning curve is €j{ (m) = 1.5(m + 1)-1. (13) The latter expression results from Lemma 1 and the equality m(Qm) _ { 2(1 - €)m - (1 - 2€)m (for 0 ~ € ~ 1/2), jJ. f 2(1 - €)m (for 1/2 < € ~ 1), (14) Different learning curves (bounds and approximations) for homogeneous perceptron are plotted in Figure 1.b. 3.2 I-dimensional higher order neuron We consider X d.~ [0,1] c R with a continuous probability distribution jJ., Define the hypothesis space H C {O, l}X as the set of all functions of the form 8op(z) where p is a Examples of Learning Curves from a Modified VC-formalism 349 polynomial of degree :::; d on R. Let the target be constant, t == 1. It is easy to see that H restricted to a finite subset of [0,1] is exactly the restriction of the family of all fimctions iI c {O, 1 }[O,lj with up to d "jumps" from a to lor 1 toO and thus dvc(H) = d+ 1. With probability 1 an m-sample Z = (Zl' "" zm) from xm is such that Zi #- Zj for i #- j. For such a generic Z, l7rt,z(H) n t:il = const = IHli. This observation was used to derive the following relations for the computation of I H Ii: min(d,m-l) IHli = L liI(6)li + liI(6)1:_i, (15) 6=0 for ° :::; i :::; m, where liI(6)li, for 0 = 0,1, ... ,d, is defined as follows. We initialize liI(O)lo = liI(l)li d~ 1 fori = 1, .. " m-1, liI(1) 10 = liI(l)l~ d~ ° and liI(6)li d~ ° for i = 0, 1, ... , m, 0 = 2,3, .. " d, and then, recurrently, for 0 ~ 2 we set liI(6) Ii d~ ~m-l . liI(6-l)l~ if 0 is odd and liI(6)1~ d~ ~m-l liI(6-l)l~ ifo is even L.Jk=max(6,m-~) ~-m+k ~ L.Jk=6 ~ . (Here liI(6)li is defined by the relation (4) with the target t == 1 for the hypothesis space H(6) C iI composed of functions having the value 1 near a and exactly 0 jumps in (0,1), exactly at entries of z; similarly as for H, IH(6)li = l7rl,zH(6) n t:il for a generic m-sample z E (0, l)m.) Analyzing an embedding of R into Rd, and using an argument based on the Vandermonde determinant as in [6,13], itcan be proved that the partition function IIH is given by Cover's counting function [4], and that (16) For the uniform distribution on [0, 1] and a generic z E [0, l]m letAk(z) denote the sum of Ie largest segments of the partition of [0, 1] into m + 1 segments by the entries of Z. Then Ald/:lJ(Z):::; e'J;arz:(z):::; Ald/:lJ+l(Z), (17) An explicit expression for the expected value of Ak is known [11], thus a very tight bound on the true learning curve eH (m) defined by (2) can be obtained: ~/2J1 (1 + E ~):::; eH(m):::; Ld~2J : 1 (1 + E ~), (18) + i=ld/:lJ+1 J + i=ld/:lJ+:l J Numerical results are shown in Figure 2. 4 Discussion and conclusions The basic estimate (5) of Theorem 1 has been used to produce upper bounds on the learning curve (via Lemma 1) in three different ways: (i) using the exact values of coefficients IHEli (Fig. 1a), (ii) using the estimate IHEli :::; IHli and the values of IHli and (iii) using the w-uniformity bound (8) with minimal value of Cw,m and as an "apprOximation" with Cw,m = const = 1. Both examples of simple learning tasks considered in the paper allowed us to compare these results with the true learning curves (or their tight bounds) which can serve as benchmarks. Figure 1.a implies that values of parameter w in the w-uniformity bound (approximation) governing a distribution of error patterns between different error shells (c.f, [10)) has a 350 A. KOWALCZYK, J. SZYMANSKI, P. L. BARTLETT, R. C. WILLIAMSON significant impact on learning curve shapes, changing from slow decrease to rapid jumps (''phase transitions',) in generalization. Figure l.b proves that one loses tightness of the bound by using I HI i rather than I HE Ii , and even more is lost if w-unifonnity bounds (with variable C W,17l) are employed. Inspecting Figures l.b and 2.a we find that approximate approaches consisting of replacing IHElr by a simple estimate (w-uniforrnity) can produce learning curves very close to IHlilearning curves suggesting that an application of this formalism to learning systems where neither IHElr nor IHlr can by calculated might be possible. This could lead to a sensible approximate theory capturing at least certain qualitative properties of learning curves for more complex learning tasks. Generally, the results of this paper show that by incorporating the limited knowledge of the statistical distribution of error patterns in the sample space one can dramatically improve bounds on the learning curve with respect to the classical universal estimates of the VCtheory. This is particularly important for "practical" training sample sizes (m ~ 12 x VC-dimension) where the VC-bounds are void. Acknowledgement. The permission of Director, Telstra Research Laboratories, to publish this paper is gratefully acknowledged. A.K. acknowledges the support of the Australian Research Council. References (1) S. Amari, N. Fujita, and S. Shinomoto. Four types of learning curves. Neural Computation, 4(4):605-618, 1992. (2) M. Anthony and N. Biggs. Computational Learning Theory. Cambridge University Press, 1992. (3) A. Blumer, A. Ehrenfeucht, D. Haussler, and M.K. Warmuth. Learnability and the VapnikChervonenkis dimensions. Journal of the ACM, 36:929-965, (Oct. 1989). (4) T.M. Cover. Geometrical and statistical properties of linear inequalities with applications to pattern recognition. IEEE Trans. Elec. Comp., EC-14:326-334, 1965. (5) D. Haussler, M. Keams, H.S. Seung, and N. Tishby. Rigorous learning curve bounds from statistical mechanics. In Proc. 7th Ann. ACM Con[. on Compo Learn. Theory, pages 76-87, 1994. (6) A. Kowalczyk. Estimates of storage capacity of multi-layer perceptron with threshold logic hidden units. Neural Networks, to appear. (7) A. Kowalczyk. VC-formalism with explicit bounds on error shells size distribution. A manuscript, 1994. (8) A. Kowalczykand H. Ferra. Generalisation in feedforward networks. Adv. in NIPS 7, The MIT Press, Cambridge, 1995. (9) A. Kowalczyk, J. Szymanski, and H. Ferra. Combining statistical physics with VC-bounds on generalisation in learning systems. In Proc. ACNN'95, Sydney, 1995. University of Sydney. (10) A. Kowalczyk, J. Szymanski, and R.C. Williamson. Learning curves from a modified vcformalism: a case stUdy. In Proceedings of ICNN'95, Perth (CD'ROM), volume VI, pages 2939-2943, Rundle Mall, South Australia, 1995. IEEE'J'Causal Production. (11) J.G. Mauldon. Random division of an interval. Proc. Cambridge Phil. Soc., 47:331-336,1951. (12) K.R. Muller, M. Finke, N. Murata, and S. Amari. On large scale simulations for learning curves. In Proc. ACNN'95, pages 45-48, Sydney, 1995. University of Sydney. (13) A. Sakurai. n-h-l networks store no less n h + 1 examples but sometimes no more. In Proceedings of the 1992 International Conference on Neural Networks,pagesill-936-ill-941. IEEE, June 1992. (14) H. Sompolinsky, H.S. Seung, and N. Tishby. Statistical mechanics of learning curves. Physical Reviews, A45:6056-6091, 1992. (15) V. Vapnik. Estimation of Dependences Based on Empirical Data. Springer-Verlag, 1982. (16) V. Vapnik, E. Levin, and Y. Le Cun. Measuring the VC-dimension ofa learning machine. Neural Computation, 6 (5):851-876, 1994. (17) C. Wang and S.S. Venkantesh. Temporal dynamics of generalisation in neural networks. Adv. in NIPS 7, The MIT Press, Cambridge, 1995.
|
1995
|
130
|
1,033
|
Using Unlabeled Data for Supervised Learning Geoffrey Towell Siemens Corporate Research 755 College Road East Princeton, N J 08540 Abstract Many classification problems have the property that the only costly part of obtaining examples is the class label. This paper suggests a simple method for using distribution information contained in unlabeled examples to augment labeled examples in a supervised training framework. Empirical tests show that the technique described in this paper can significantly improve the accuracy of a supervised learner when the learner is well below its asymptotic accuracy level. 1 INTRODUCTION Supervised learning problems often have the following property: unlabeled examples have little or no cost while class labels have a high cost. For example, it is trivial to record hours of heartbeats from hundreds of patients. However, it is expensive to hire cardiologists to label each of the recorded beats. One response to the expense of class labels is to squeeze the most information possible out of each labeled example. Regularization and cross-validation both have this goal. A second response is to start with a small set of labeled examples and request labels of only those currently unlabeled examples that are expected to provide a significant improvement in the behavior of the classifier (Lewis & Catlett, 1994; Freund et al., 1993). A third response is to tap into a largely ignored potential source of information; namely, unlabeled examples. This response is supported by the theoretical work of Castelli and Cover (1995) which suggests that unlabeled examples have value in learning classification problems. The algorithm described in this paper, referred to as SULU (Supervised learning Using Labeled and Unlabeled examples), takes this third 648 G. TOWELL path by using distribution information from unlabeled examples during supervised learning. Roughly, SULU uses the centroid of labeled and unlabeled examples in the neighborhood of a labeled example as a new training example. In this way, SULU extracts information about the local variability of the input from unlabeled data. SULU is described in Section 2. In its use of unlabeled examples to alter labeled examples, SULU is reminiscent of techniques for adding noise to networks during training (Hanson, 1990; Matsuoka, 1992). SULU is also reminiscent of instantiations of the EM algorithm that attempt to fill in missing parts of examples (Ghahramani & Jordan, 1994). The similarity of SULU to these, and other, works is explored in Section 3. SULU is intended to work on classification problems for which there is insufficient labeled training data to allow a learner to approach its asymptotic accuracy level. To explore this problem, the experiments described in Section 4 focus on the early parts of the learning curves of six datasets (described in Section 4.1). The results show that SULU consistently, and statistically significantly, improves classification accuracy over systems trained with only the labeled data. Moreover, SULU is consistently more accurate than an implementation of the EM-algorithm that was specialized for the task of filling in missing class labels. From these results, it is reasonable to conclude that SULU is able to use the distribution information in unlabeled examples to improve classification accuracy. 2 THE ALGORITHM SULU uses standard neural-network supervised training techniques except that it occasionally replaces a labeled example with a synthetic example. in addition, the criterion to stop training is slightly modified to require that the network correctly classify almost every labeled example and a majority of the synthetic examples. For instance, the experiments reported in Section 4 generate synthetic examples 50% of the time; the stopping criterion requires that 80% of the examples seen in a single epoch are classified correctly. The main function in Table 1 provides psuedocode for this process. The synthesize function in Table 1 describes the process through which an example is synthesized. Given a labeled example to use as a seed, synthesize collects neighboring examples and returns an example that is the centroid of the collected examples with the label of the starting point. synthesize collects neighboring examples until reaching one of the following three stopping points. First, the maximum number of points is reached; the goal of SULU is to get information about the local variance around known points, this criterion guarantees locality. Second, the next closest example to the seed is a labeled example with a different label; this criterion prevents the inclusion of obviously incorrect information in synthetic examples. Third, the next closest example to the seed is an unlabeled example and the closest labeled example to that unlabeled example has a different label from the seed; this criterion is intended to detect borders between classification areas in example space. The call to synthesize from main effectively samples with replacement from a space defined by a labeled example and its neighbors. As such, there are many ways in which main and synthesize could be written. The principle consideration in this implementation is memory; the space around the labeled examples can be huge. Using Unlabeled Data for Supervised Learning 649 Table 1: Pseudocode for SULU RANDOH(min,max): return a uniformly distributed random integer between min and max, inclusive HAIN(B,H): /* B - in [0 .. 100], controls the rate of example synthesis */ /* H - controls neighborhood size during synthesis */ Let: E /* a set of labeled examples */ U /* a set of unlabeled examples */ N /* an appropriate neural network */ Repeat Permute E Foreach e in E if random(0,100) > B then e (- SYNTHESIZE(e,E,U,random(2,M» TRAIN N using e Until a stopping criterion is reached SYNTHESIZE(e,E,U,m): Let: C /* will hold a collection of examples */ For i from 1 to m c (- ith nearest neighbor of e in E union U if «c is labeled) and (label of c not equal to label of e» then STOP if c is not labeled cc (- nearest neighbor of c in E if label of cc not equal to label of e then STOP add c to C return an example whose input is the centroid of the inputs of the examples in C and has the class label of e. 3 RELATED WORK SULU is similar to two methods of exploring the input space beyond the boundaries of the labeled examples; example generation and noise addition. Example generation commonly uses a model of how a space deforms and an example of the space to generate new examples. For instance, in training a vehicle to turn, Pomerleau (1993) used information about how the scene shifts when a car is turned to gener·ate examples of turns. The major problem with example generation is that deformation models are uncommon. By contrast to example generation, noise addition is a model-free procedure. In general, the idea is to add a small amount of noise to either inputs (Matsuoka, 1992), link weights (Hanson, 1990), or hidden units (Judd & Munro, 1993). For example, Hanson (1990) replaces link weights with a Gaussian. During a forward pass, the Gaussian is sampled to determine the link weight. Training affects both the mean and the variance of the Gaussian. In so doing, Hanson's method uses distribution information in the labeled examples to estimate the global variance of each input dimension. By contrast, SULU uses both labeled and unlabeled examples to make local variance estimates. (Experiments, results not shown, with Hanson's method indicate that it cannot improve classification results as much as SULU.) Finally, there has been some other work on using unclassified examples during training. de Sa (1994) uses the co-occurrence of inputs in multiple sensor modali650 G. TOWELL ties to substitute for missing class information. However, sensor data from multiple modalities is often not available. Another approach is to use the EM algorithm (Ghahramani & Jordan, 1994) which iteratively guesses the value of missing information (both input and output) and builds structures to predict the missing information. Unlike SULU, EM uses global information in this process so it may not perform well on highly disjunctive problems. Also SULU may have an advantage over EM in domains in which only the class label is missing as that is SULU'S specific focus. 4 EXPERIMENTS The experiments reported in this section explore the behavior of SULU on six datasets. Each of the datasets has been used previously so they are only briefly described in the first subsection. The results of the experiments reported in the last part of this section show that SULU significantly and consistently improves classification results. 4.1 DATASETS The first two datasets are from molecular biology. Each take a DNA sequence and encode it using four bits per nucleotide. The first problem, promoter recognition (Opitz & Shavlik, 1994), is: given a sequence of 57 DNA nucleotides, determine if a promoter begins at a particular position in the sequence. Following Opitz and Shavlik, the experiments in this paper use 234 promoters and 702 non promoters. The second molecular biology problem, splice-junction determination (Towell & Shavlik, 1994), is: given a sequence of 60 DNA nucleotides, determine if there is a splice-junction (and the type of the junction) at the middle of the sequence. The data consist of 243 examples of one junction type (acceptors), 228 examples of the other junction type (donors) and 536 examples of non-junctions. For both of these problems, the best randomly initialized neural networks have a small number of hidden units in a single layer (Towell & Shavlik, 1994). The remaining four datasets are word sense disambiguation problems (Le. determine the intended meaning of the word "pen" in the sentence "the box is in the pen"). The problems are to learn to distinguish between six noun senses of "line" or four verb senses of "serve" using either topical or local encodings (Leacock et al., 1993) of a context around the target word. The line dataset contains 349 examples of each sense. Topical encoding, retaining all words that occur more than twice, requires 5700 position vectors. Local encoding, using three words on either side of line, requires 4500 position vectors. The serve dataset contains 350 examples of each sense. Under the same conditions as line, topical encoding requires 4400 position vectors while local encoding requires 4500 position vectors. The best neural networks for these problems have no hidden units (Leacock et al., 1993). 4.2 METHODOLOGY The following methodology was used to test SULU on each dataset. First, the data was split into three sets, 25 percent was set aside to be used for assessing generalization, 50 percent had the class labels stripped off, and the remaining 25 percent was to be used for training. To create learning curves, the training set was Using Unlabeled Data for Supervised Learning 651 Table 2: Endpoints of the learnings curves for standard neural networks and the best result for each of the six datasets. Training Splice Serve Line Set size Promoter Junction Local Topical Local Topical smallest 74.7 66.4 53.9 41.8 38.7 40.6 largest 90.3 85.4 71.7 63.0 58.8 63.3 asymptotic 95.8 94.4 83.1 75.5 70.1 79.2 further subdivided into sets containing 5, 10, 15, 20 and 25 percent of the data such that smaller sets were always subsets of larger sets. Then, a single neural network was created and copied 25 times. At each training set size, a new copy of the network was trained under each of the following conditions: 1) using SULU, 2) using SULU but supplying only the labeled training examples to synthesize, 3) standard network training, 4) using a variant of the EM algorithm that has been specialized to the task of filling in missing class labels, and 5) using standard network training but with the 50% unlabeled prior to stripping the labels. This procedure was repeated eleven times to average out the effects of example selection and network initialization. When SULU was used, synthetic examples replaced labeled examples 50 percent of the time. Networks using the full SULU (case 1) were trained until 80 percent of the examples in a single epoch were correctly classified. All other networks were trained until at least 99.5% of the examples were correctly classified. Stopping criteria intended to prevent overfitting were investigated, but not used because they never improved generalization. 4.3 RESULTS & DISCUSSION Figure 1 and Table 2 summarize the results of these experiments. The graphs in Figure 1 show the efficacy of each algorithm. Except for the largest training set on the splice junction problem, SULU always results in a statistically significant improvement over the standard neural network with at least 97.5 percent confidence (according to a one-tailed paired-sample t-test). Interestingly, SULU'S improvement is consistently between :t and ~ of that achieved by labeling the unlabeled examples. This result contrasts Castelli and Cover's (1995) analysis which suggests that labeled examples are exponentially more valuable than unlabeled examples. In addition, SUL U is consistently and significantly superior to the instantiation of the EM-algorithm when there are very few labeled samples. As the number of labeled samples increases the advantage of SULU decreases. At the largest training set sizes tested, the two systems are roughly equally effective. A possible criticism of SULU is that it does not actually need the unlabeled examples; the procedure may be as effective using only the labeled training data. This hypothesis is incorrect, As shown in Figure 1, SULU when given no unlabeled examples is consistently and significantly inferior ti SULU when given a large number of unlabeled examples. In addition, SULU with no unlabeled examples is consistently, although not always significantly, inferior to a standard neural network. The failure of SULU with only labeled examples points to a significant weakness 652 ~r-----------~~_~_~.~ ~7_~~~~ ~--' --SI..l.U Wl1h _un~ --EM""'_~ SUlU .... O~ + SIIl&IIIcII't'-..penorto5U.U o SIIIblIlcllly _ new .. SULU ~r---~c---~,~---,~--~~--~ Size oIlraining sel ~~~ ____ ~~~~~_Ud_ ~~U~~_Dh~~~~~~-' '-'-,== :-~-:oo~= ' ........ - - - --SlJ..UWllhOurNbe'-d ......... ......... + Stabs\ll::aly' M.lpenor to SlLU ......... ~o $tabstlcaly' ...... m SlJ..U ---..... _---. -SLl..U wrIh 1046 un1 .... ' .......... EM"lrI l04&~ ........... ~----su. Uwrttl ()lA'l~ ......... '+---- __ + SlatlSt.c.Ity alpena, m SUlU __ ....... -.!! _stat\s\lealyn1erD'tDSlJ..U - --+-----. G. TOWELL ~r---~------~~ _~_ ~~---+~~-~----~ " --8ll.u .... 50'2un~ ~ g "\. ~~~ :.~== ..fi~ ................ + Strolllk:lllJ~tDSLl.U _ ....... ~~..: lIIIaaly .. nor » SULU '8~ """-+-_ j:r---4~~ "-" -"~"'~" ~"~"-" -"~"-" --~-" -"~~~':~--~ 0-------G----_---~ ~r---~--~ l ~ O --~ l ~~~--~~~{· Size oIlralnlng sel ~ __ -..e--~o 0.. .... C'" __ __ -Ouu - - _u __ -e-. __ ~ ~ ~o~----~~----~-----&------~ <>. Figure 1: The effect of five training procedures on each of six learning problems. In each of the above graphs, the effect of standard neural learning has been subtracted from all results to suppress the increase in accuracy that results simply from an increase in the number of labeled training examples. Observations marked by a '0' or a '+' respectively indicate that the point is statistically significantly inferior or superior to a network trained using SULU. in its current implementation. Specifically, SULU finds the nearest neighbors of an example using a simple mismatch counting procedure. Tests of this procedure as an independent classification technique (results not shown) indicate that it is consistently much worse than any of the methods plotted in in Figure 1. Hence, its use imparts a downward bias to the generalizatio~ results. A second indication of room for improvement in SULU is the difference in generalization between SULU and a network trained using data in which the unlabeled examples provided to SULU have labels (case 5 above). On every dataset, the gain from labeling the examples is statistically significant. The accuracy of a network trained with all labeled examples is an upper bound for SULU, and one that is likely not reachable. However, the distance between the upper bound and SULU'S current performance indicate that there is room for improvement. Using Unlabeled Data for Supervised Learning 653 5 CONCLUSIONS This paper has presented the SULU algorithm that combines aspects of nearest neighbor classification with neural networks to learn using both labeled and unlabeled examples. The algorithm uses the labeled and unlabeled examples to construct synthetic examples that capture information about the local characteristics of the example space. In so doing, the range of examples seen by the neural network during its supervised learning is greatly expanded which results in improved generalization. Results of experiments on six real-work datasets indicate that SULU can significantly improve generalization when when there is little labeled data. Moreover, the results indicate that SULU is consistently more effective at using unlabeled examples than the EM-algorithm when there is very little labeled data. The results suggest that SULU will be effective given the following conditions: 1) there is little labeled training data, 2) unlabeled training data is essentially free, 3) the accuracy of the classifier when trained with all of the available data is below the level which is expected to be achievable. On problems with all of these properties SULU may significantly improve the generalization accuracy of inductive classifiers. References Castelli, V. & Cover, T. (1995). The relative value of labeled and unlabeled samples in pattern recognition with an unknown mixing parameter. (Technical Report 86), Department of Statistics: Stanford University. de Sa, V. (1994). Learning classification with unlabeled data. Advances in Neural Information Processing Systems, 6. Freund, Y., Seung, H. S., Shamit, E., & Tishby, N. (1993). Information, prediction and query by committee. Advances in Neural Information Processing Systems, 5. Ghahramani, Z. & Jordan, M. I. (1994). Supervised learning from incomplete data via an EM approach. Advances in Neural Information Processing Systems, 6. Hanson, S. J. (1990). A stochastic version of the delta rule. Physica D, 42, 265-272. Judd, J. S. & Munro, P. W. (1993). Nets with unreliable hidden units learn errOrcorrecting codes. Advances in Neural Information Processing Systems, 5. Leacock, C., Towell, G., & Voorhees, E. M. (1993). Towards building contextual representations of word senses using statistical models. Proceedings of SIGLEX Workshop: Acquisition of Lexical Knowledge from Text. Association for Computational Linguistics. Lewis, D. D. & Catlett, J. (1994). Heterogeneous uncertainty sampling for supervised learning. Eleventh International Machine Learning Conference. Matsuoka, K. (1992). Noise injection into inputs in back-propagation learning. IEEE Transactions on Systems, Man and Cybernetics, 22, 436-440. Opitz, D. W. & Shavlik, J. W. (1994). Using genetic search to refine knowledge-based neural networks. Eleventh International Machine Learning Conference. Pomerleau, D. A. (1993). Neural Network Perception for Mobile Robot Guidance. Boston: Kluwer. Towell, G. G. & Shavlik, J. W. (1994). Knowledge-based artificial neural networks. Artificial Intelligence, 70, 119-165.
|
1995
|
131
|
1,034
|
Implementation Issues in the Fourier Transform Algorithm Yishay Mansour" Sigal Sahar t Computer Science Dept. Tel-Aviv University Tel-Aviv, ISRAEL Abstract The Fourier transform of boolean functions has come to play an important role in proving many important learnability results. We aim to demonstrate that the Fourier transform techniques are also a useful and practical algorithm in addition to being a powerful theoretical tool. We describe the more prominent changes we have introduced to the algorithm, ones that were crucial and without which the performance of the algorithm would severely deteriorate. One of the benefits we present is the confidence level for each prediction which measures the likelihood the prediction is correct. 1 INTRODUCTION Over the last few years the Fourier Transform (FT) representation of boolean functions has been an instrumental tool in the computational learning theory community. It has been used mainly to demonstrate the learnability of various classes of functions with respect to the uniform distribution. The first connection between the Fourier representation and learnability of boolean functions was established in [6] where the class ACo was learned (using its FT representation) in O(nPoly-log(n)) time. The work of [5] developed a very powerful algorithmic procedure: given a function and a threshold parameter it finds in polynomial time all the Fourier coefficients of the function larger than the threshold. Originally the procedure was used to learn decision trees [5], and in [8, 2, 4] it was used to learn polynomial size DNF. The FT technique applies naturally to the uniform distribution, though some of the learnability results were extended to product distribution [1, 3] . .. e-mail: manSQur@cs.tau.ac.il t e-mail: gales@cs.tau.ac.il Implementation Issues in the Fourier Transform Algorithm 261 A great advantage of the FT algorithm is that it does not make any assumptions on the function it is learning. We can apply it to any function and hope to obtain "large" Fourier coefficients. The prediction function simply computes the sum of the coefficients with the corresponding basis functions and compares the sum to some threshold. The procedure is also immune to some noise and will be able to operate even if a fraction of the examples are maliciously misclassified. Its drawback is that it requires to query the target function on randomly selected inputs. We aim to demonstrate that the FT technique is not only a powerful theoretical tool, but also a practical one. In the process of implementing the Fourier algorithm we enhanced it in order to improve the accuracy of the hypothesis we generate while maintaining a desirable run time. We have added such feartures as the detection of inaccurate approximations "on the fly" and immediate correction of the errors incurred at a minimal cost. The methods we devised to choose the "right" parameters proved to be essential in order to achieve our goals. Furthermore, when making predictions, it is extremely beneficial to have the prediction algorithm supply an indicator that provides the confidence level we have in the prediction we made. Our algorithm provides us naturally with such an indicator as detailed in Section 4.1. The paper is organized as follows: section 2 briefly defines the FT and describes the algorithm. In Section 3 we describe the experiments and their outcome and in Section 4 the enhancements made. We end with our conclusions in Section 5. 2 FOURIER TRANSFORM (FT) THEORY In this section we briefly introduce the FT theory and algorithm. its connection to learning and the algorithm that finds the large coefficients. A comprehensive survey of the theoretical results and proofs can be found in [7]. We consider boolean functions of n variables: f : {O, l}n -t {-I, I}. We define the inner product: < g, f >= 2-n L::XE{O,l}R f(x)g(x) = E[g . f], where E is the expected value with respect to the uniform distribution. The basis is defined as follows: for each z E {O,l}n, we define the basis function :\:z(Xl,···,Xn) = (_1)L::~=lx;z •. Any function of n boolean inputs can be uniquely expressed as a linear combination of the basis functions. For a function f, the zth Fourier coefficient of f is denoted by j(z), i.e., f(x) = L::zE{O,l}R j(z)XAx). The Fourier coefficients are computed by j(z) =< f, Xz > and we call z the coefficient-name of j(z). We define at-sparse function to be a function that has at most t non-zero Fourier coefficients. 2.1 PREDICTION Our aim is to approximate the target function f by a t-sparse function h. In many cases h will simply include the "large" coefficients of f. That is, if A = {Zl' ... , zm} is the set of z's for which j(Zi) is "large", we set hex) = L::z;EA aiXz;(x), where at is our approximation of j(Zi). The hypothesis we generate using this process, hex), does not have a boolean output. In order to obtain a boolean prediction we use Sign(h(x)), i.e., output +1 if hex) 2 0 and -1 if hex) < o. We want to bound the error we get from approximating f by h using the expected error squared, E[(J - h )2]. It can be shown that bounding it bounds the boolean prediction error probability, i.e., Pr[f(x) f. sign(h(x))] ~ E[(J - h)2]. For a given t, the t-sparse 262 Y. MANSOUR, S. SAHAR hypothesis h that minimizes E[(J - h)2] simply includes the t largest coefficients of f. Note that the more coefficients we include in our approximation and the better we approximate their values, the smaller E[(J - h )2] is going to be. This provides us with the motivation to find the "large" coefficients. 2.2 FINDING THE LARGE COEFFICIENTS The algorithm that finds the "large" coefficients receives as inputs a function 1 (a black-box it can query) and an interest threshold parameter (J > 0. It outputs a list of coefficient-names that (1) includes all the coefficients-names whose corresponding coefficients are "large", i.e., at least (J , and (2) does not include "too many" coefficient-names. The algorithm runs in polynomial time in both 1/() and n. SUBROUTINE search( a) IF TEST[J, a, II] THEN IF lal = n THEN OUTPUT a ELSE search(aO); search(al); Figure 1: Subroutine search The basic idea of the algorithm is to perform a search in the space of the coefficientnames of I. Throughout the search algorithm (see Figure (1)) we maintain a prefix of a coefficient-name and try to estimate whether any of its extensions can be a coefficient-name whose value is "large". The algorithm commences by calling search(A) where A is the empty string. On each invocation it computes the predicate TEST[/, a, (J]. If the predicate is true, it recursively calls search(aO) and search(al). Note that if TEST is very permissive we may reach all the coefficients, in which case our running time will not be polynomial; its implementation is therefore of utmost interest. Formally, T EST[J, a, (J] computes whether Exe{O,l}n-"E;e{O,lP.[J(YX)Xa(Y)] 2: (J2, where k = Iiali . (1) Define la(x) = L:,ae{O,l}n-" j(aj3)x.,a(x). It can be shown that the expected value in (1) is exactly the sum of the squares of the coefficients whose prefix is a , i.e., Exe{o,l}n-"E;e{o,l}d/(yx)x.a(Y)] = Ex[/~(x)] = L:,ae{o,l}n-" p(aj3), implying that if there exists a coefficient Ii( a,8)1 2: (), then E[/;] 2: (J2 . This condition guarantees the correctness of our algorithm, namely that we reach all the "large" coefficients. We would like also to bound the number of recursive calls that search performs. We can show that for at most 1/(J2 of the prefixes of size k, TEST[!, a , (J] is true. This bounds the number of recursive calls in our procedure by O(n/(J2). In TEST we would like to compute the expected value, but in order to do so efficiently we settle for an approximation of its value. This can be done as follows: (1) choose ml random Xi E {a, l}n-k, (2) choose m2 random Yi,j E {a, l}k , (3) query 1 on Yi,jXi (which is why we need the query model-to query f on many points with the same prefix Xi) and receive I(Yi,j xd, and (4) compute the estimate as, Ba = ';1 L:~\ (~~ L:~l I(Yi,iXdXa(Yi,j)f . Again, for more details see [7]. 3 EXPERIMENTS We implemented the FT algorithm (Section 2.2) and went forth to run a series of experiments. The parameters of each experiment include the target function, (J , ml Implementation Issues in the Fourier Transform Algorithm 263 and m2. We briefly introduce the parameters here and defer the detailed discussion. The parameter () determines the threshold between "small" and "large" coefficients, thus controlling the number of coefficients we will output. The parameters wI and w2 determine how accurately we approximate the TEST predicate. Failure to approximate it accurately may yield faulty, even random, results (e.g., for a ludicrous choice of m1 = 1 and m2 = 1) that may cause the algorithm to fail (as detailed in Section 4.3). An intelligent choice of m1 and m2 is therefore indispensable. This issue is discussed in greater detail in Sections 4.3 and 4.4. Figure 2: Typical frequency plots and typical errors. Errors occur in two cases: (1) the algorithm predicts a +1 response when the actual response is -1 (the lightly shaded area), and (2) the algorithm predicts a -1 response, while the true response is +1 (the darker shaded area) . Figures (3)-(5) present representative results of our experiments in the form of graphs that evaluate the output hypothesis of the algorithm on randomly chosen test points. The target function, I, returns a boolean response, ±1, while the FT hypothesis returns a real response. We therefore present, for each experiment, a graph constituting of two curves: the frequency of the values of the hypothesis, h( x), when I( x) = +1, and the second curve for I( x) = -1. If the two curves intersect, their intersection represents the inherent error the algorithm makes. Figure 3: Decision trees of depth 5 and 3 with 41 variables. The 5-deep (3-deep) decision tree returns -1 about 50% (62.5%) of the time. The results shown above are for values (J = 0.03, ml = 100 and m2 = 5600 «(J = 0.06, ml = 100 and m2 = 1300). Both graphs are disjoint, signifying 0% error. 4 RESULTS AND ALGORITHM ENHANCEMENTS 4.1 CONFIDENCE LEVELS One of our most consistent and interesting empirical findings was the distribution of the error versus the value of the algorithm's hypothesis: its shape is always that of a bell shaped curve. Knowing the error distribution permits us to determine with a high (often 100%) confidence level the result for most of the instances, yielding the much sought after confidence level indicator. Though this simple logic thus far has not been supported by any theoretical result, our experimental results provide overwhelming evidence that this is indeed the case. Let us demonstrate the strength of this technique: consider the results of the 16-term DNF portrayed in Figure (4). If the algorithm's hypothesis outputs 0.3 (translated 264 Y. MANSOUR, S. SAHAR Figure 4: 16 terlD DNF. This (randomly generated) DNF of 40 variables returns -1 about 61 % of the time. The results shown above are for the values of 9 = 0.02 , m2 = 12500 and ml = 100. The hypothesis uses 186 non-zero coefficients. A total of 9.628% error was detected. into 1 in boolean terms by the Sign function), we know with an 83% confidence level that the prediction is correct. If the algorithm outputs -0.9 as its prediction, we can virtually guarantee that the response is correct. Thus, although the total error level is over 9% we can supply a confidence level for each prediction. This is an indispensable tool for practical usage of the hypothesis. 4.2 DETERMINING THE THRESHOLD Once the list of large coefficients is built and we compute the hypothesis h( x), we still need to determine the threshold, a, to which we compare hex) (i.e., predict +1 iff hex) > a). In the theoretical work it is assumed that a = 0, since a priori one cannot make a better guess. We observed that fixing a's value according to our hypothesis, improves the hypothesis. a is chosen to minimize the error with respect to a number of random examples. Figure 5: 8 terlD DNF. This (randomly generated) DNF of 40 variables returns -1 about 43% of the time. The results shown above are for the values of 9 = 0.03, m2 = 5600 and ml = 100. The hypothesis consists of 112 non-zero coefficients. For example, when trying to learn an 8-term DNF with the zero threshold we will receive a total of 1.22% overall error as depicted in Figure (5). However, if we choose the threshold to be 0.32, we will get a diminished error of 0.068%. 4.3 ERROR DETECTION ON THE FLY - RETRY During our experimentations we have noticed that at times the estimate Ba for E[J~] may be inaccurate. A faulty approximation may result in the abortion of the traversal of "interesting" subtreees, thus decreasing the hypothesis' accuracy, or in traversal of "uninteresting" subtrees, thereby needlessly increasing the algorithm's runtime. Since the properties of the FT guarantee that E[J~] = E[f~o] + E[J~d, we expect Ba :::::: Bao + Bal . Whenever this is not true, we conclude that at least one of our approximations is somewhat lacking. We can remedy the situation by Implementation Issues in the Fourier Transform Algorithm 265 running the search procedure again on the children, i.e., retry node a. This solution increases the probability of finding all the "large" coefficients. A brute force implementation may cost us an inordinate amount of time since we may retraverse subtrees that we have previously visited. However, since any discrepancies between the parent and its children are discovered-and corrected-as soon as they appear, we can circumvent any retraversal. Thus, we correct the errors without any superfluous additions to the run time. --J: ,-" i\ o " ....... Figure 6: Majority function of 41 variables. The result portrayed are for values m1 = 100, m2 = 800 and (J = 0.08. Note the majority-function characteristic distribution of the results1 . We demonstrate the usefulness of this approach with an example of learning the majority function of 41 boolean variables. Without the retry mechanism, 8 (of a total of 42) large coefficients were missed, giving rise to 13.724% error represented by the shaded area in Figure (6). With the retries all the correct coefficients were found, yielding perfect (flawless) results represented in the dotted curve in Figure (6). 4.4 DETERMINING THE PARAMETERS One of our aims was to determine the values of the different parameters, m1, m2 and (}. Recall that in our algorithm we calculate Ba , the approximation of Ex[f~(x)] where m1 is the number of times we sample x in order to make this approximation. We sample Y randomly m2 times to approximate fa(Xi) = Ey[f(YXih:a(Y)), for each Xi · This approximation of fa(Xi) has a standard deviation of approximately A . Assume that the true value is 13i, i.e. f3i = fa(Xi), then we expect the contribution of the ith element to Ba to be (13i ± )n;? = 131 ± J&; + rr!~. The algorithm tests Ba = rr!1 L 131 ? (}2, therefore, to ensure a low error, based on the above argument, we choose m2 = (J52 • Choosing the right value for m2 is of great importance. We have noticed on more than one occasion that increasing the value of m2 actually decreases the overall run time. This is not obvious at first: seemingly, any increase in the number of times we loop in the algorithm only increases the run time. However, a more accurate value for m2 means a more accurate approximation of the TEST predicate, and therefore less chance of redundant recursive calls (the run time is linear in the number of recursive calls). We can see this exemplified in Figure (7) where the number of recursive calls increase drastically as m2 decreases. In order to present Figure (7), 1The "peaked" distribution of the results is not coincidental. The FT of the majority function has 42 large equal coefficients, labeled cmaj' one for each singleton (a vector of the form 0 .. 010 .. 0) and one for parity (the all-ones vector). The zeros of an input vector with z zeros we will contribute ±1(2z - 41). cmajl to the result and the parity will contribute ±cma) (depending on whether z is odd or even), so that the total contribution is an even factor of cma)' Since cma) = (~g);tcr - 0 .12, we have peaks around factors of 0.24. The distribution around the peaks is due to the f~ct we only approximate each coefficient and get a value close to cma)' 266 Y. MANSOUR, S. SAHAR we learned the same 3 term DNF always using e = 0.05 and mr * m2 The trials differ in the specific values chosen in each trial for m2. 100000. Figure 7: Deter01ining 012' Note that the number of recursive calls grows dramatically as m2 's value decreases. For example, for m2 = 400, the number of recursive calls is 14,433 compared with only 1,329 recursive calls for m2 = 500. SPECIAL CASES: When k = 110'11 is either very small or very large, the values we choose for ml and m2 can be self-defeating: when k ,..... n we still loop ml (~ 2n - k ) times, though often without gaining additional information. The same holds for very small values of k, and the corresponding m2 (~ 2k) values. We therefore add the following feature: for small and large values of k we calculate exactly the expected value thereby decreasing the run time and increasing accuracy. 5 CONCLUSIONS In this work we implemented the FT algorithm and showed it to be a useful practical tool as well as a powerful theoretical technique. We reviewed major enhancements the algorithm underwent during the process. The algorithm successfully recovers functions in a reasonable amount of time. Furthermore, we have shown that the algorithm naturally derives a confidence parameter. This parameter enables the user in many cases to conclude that the prediction received is accurate with extremely high probability, even if the overall error probability is not negligible. Acknowledgements This research was supported in part by The Israel Science Foundation administered by The Israel Academy of Science and Humanities and by a grant of the Israeli Ministry of Science and Technology. References [1) Mihir Bellare. A technique for upper bounding the spectral norm with applications to learning. In 5th Annual Work&hop on Computational Learning Theory, pages 62-70, July 1992. (2) Avrim Blum, Merrick Furst, Jeffrey Jackson, Michael Kearns, Yishay Mansour, and Steven Rudich. Weakly learning DNF and characterizing statistical query learning using fourier analysis. In The 26th Annual AC M Sympo&ium on Theory of Computing, pages 253 - 262, 1994. (3) Merrick L. Furst , Jeffrey C. Jackson, and Sean W. Smith. Improved learning of ACO functions. In 4th Annual Work&hop on Computational Learning Theory, pages 317-325, August 1991. (4) J. Jackson. An efficient membership-query algorithm for learning DNF with respect to the uniform distribution. In Annual Sympo&ium on Switching and Automata Theory, pages 42 - 53, 1994. (5) E. Kushilevitz and Y. Mansour. Learning decision trees using the fourier spectrum. SIAM Journal on Computing 22(6): 1331-1348, 1993. (6) N. Linial, Y. Mansour, and N . Nisan. Constant depth circuits, fourier transform and learnability. JACM 40(3):607-620, 1993. (7) Y. Mansour. Learning Boolean Functions via the Fourier Transform. Advance& in Neural Computation, edited by V.P. Roychodhury and K-Y. Siu and A. Orlitsky, Kluwer Academic Pub. 1994. Can be accessed via Up:/ /ftp.math.tau.ac.iJ/pub/mansour/PAPERS/LEARNING/fourier-survey.ps.Z. (8) Yishay Mansour. An o(nlog log n) learning algorihm for DNF under the uniform distribution. J. of Computer and Sy&tem Science, 50(3):543-550, 1995.
|
1995
|
132
|
1,035
|
A Bound on the Error of Cross Validation Using the Approximation and Estimation Rates, with Consequences for the Training-Test Split 1 INTRODUCTION Michael Kearns AT&T Research We analyze the performance of cross validation 1 in the context of model selection and complexity regularization. We work in a setting in which we must choose the right number of parameters for a hypothesis function in response to a finite training sample, with the goal of minimizing the resulting generalization error. There is a large and interesting literature on cross validation methods, which often emphasizes asymptotic statistical properties, or the exact calculation of the generalization error for simple models. Our approach here is somewhat different, and is pri mari I y inspired by two sources. The first is the work of Barron and Cover [2], who introduced the idea of bounding the error of a model selection method (in their case, the Minimum Description Length Principle) in terms of a quantity known as the index of resolvability. The second is the work of Vapnik [5], who provided extremely powerful and general tools for uniformly bounding the deviations between training and generalization errors. We combine these methods to give a new and general analysis of cross validation performance. In the first and more formal part of the paper, we give a rigorous bound on the error of cross validation in terms of two parameters of the underlying model selection problem: the approximation rate and the estimation rate. In the second and more experimental part of the paper, we investigate the implications of our bound for choosing 'Y, the fraction of data withheld for testing in cross validation. The most interesting aspect of this analysis is the identification of several qualitative properties of the optimal 'Y that appear to be invariant over a wide class of model selection problems: • When the target function complexity is small compared to the sample size, the performance of cross validation is relatively insensitive to the choice of 'Y. • The importance of choosing 'Y optimally increases, and the optimal value for 'Y decreases, as the target function becomes more complex relative to the sample size. • There is nevertheless a single fixed value for'Y that works nearly optimally for a wide range of target function complexity. 2 THE FORMALISM We consider model selection as a two-part problem: choosing the appropriate number of parameters for the hypothesis function, and tuning these parameters. The training sample is used in both steps of this process. In many settings, the tuning of the parameters is determined by a fixed learning algorithm such as backpropagation, and then model selection reduces to the problem of choosing the architecture. Here we adopt an idealized version of this division of labor. We assume a nested sequence of function classes Hl C ... C H d ••• , called the structure [5], where Hd is a class of boolean functions of d parameters, each IPerhaps in conflict with accepted usage in statistics, here we use the term "cross validation" to mean the simple method of saving out an independent test set to perform model selection. Precise definitions will be stated shortly. 184 M.KEARNS function being a mapping from some input space X into {O, I}. For simplicity, in this paper we assume that the Vapnik-Chervonenkis (VC) dimension [6, 5] of the class Hd is O( d). To remove this assumption, one simply replaces all occurrences of d in our bounds by the VC dimension of H d • We assume that we have in our possession a learning algorithm L that on input any training sample 8 and any value d will output a hypothesis function hd E H d that minimizes the training error over H d that is, £ t ( hd) = minhE H" { £ t (h)}, where EtCh) is the fraction of the examples in 8 on which h disagrees with the given label. In many situations, training error minimization is known to be computationally intractable, leading researchers to investigate heuristics such as backpropagation. The extent to which the theory presented here applies to such heuristics will depend in part on the extent to which they approximate training error minimization for the problem under consideration. Model selection is thus the problem of choosing the best value of d. More precisely, we assume an arbitrary target function I (which mayor may not reside in one of the function classes in the structure H1 C ... C H d ••• ), and an input distribution P; I and P together define the generalization error function £g(h) = PrzEP[h(x) =f I(x)]. We are given a training sample 8 of I, consisting of m random examples drawn according to P and labeled by I (with the labels possibly corrupted by a noise process that randomly complements each label independently with probability TJ < 1/2). The goal is to minimize the generalization error of the hypothesis selected. In this paper, we will make the rather mild but very useful assumption that the structure has the property that for any sample size m, there is a value dm.u:(m) such that £t(hdm.u:(m)) = o for any labeled sample 8 of m examples. We call the function dmaz(m) the fitting number of the structure. The fitting number formalizes the simple notion that with enough parameters, we can always fit the training data perfectly, a property held by most sufficiently powerful function classes (including multilayer neural networks). We typically expect the fitting number to be a linear function of m, or at worst a polynomial in m. The significance of the fitting number for us is that no reasonable model selection method should choose hd for d ~ dmaz(m), since doing so simply adds complexity without reducing the training error. In this paper we concentrate on the simplest version of cross validation. We choose a parameter "( E [0, 1], which determines the split between training and test data. Given the input sample 8 of m examples, let 8' be the subsample consisting of the first (1 - "()m examples in 8, and 8" the subsample consisting of the last "(mexamples. In cross validation, rather than giving the entire sample 8 to L, we give only the smaller sample 8', resulting in the sequence h1' ... , hdmaz ((1-"I)m) of increasingly complex hypotheses. Each hypothesis is now obtained by training on only (I - "()m examples, which implies that we will only consider values of d smaller than the corresponding fitting number dmaz((1 - "()m); let us introduce the shorthand d"!naz for dmaz((1 - "()m). Cross validation chooses the hd satisfying hd = mini E {1, ... ,d~az} { £~' (~)} where £~' (~) is the error of hi on the subsample 8". Notice that we are not considering multifold cross validation, or other variants that make more efficient use of the sample, because our analyses will require the independence of the test set. However, we believe that many of the themes that emerge here may apply to these more sophisticated variants as well. We use £ ClI ( m) to denote the generalization error £ g( hd) ofthe hypothesis hd chosen by cross validation when given as input a sample 8 of m random examples of the target function. Obviously, £clI(m) depends on 8, the structure, I, P, and the noise rate. When bounding £cv (m), we will use the expression "with high probability" to mean with probability 1 ~ over the sample 8, for some small .fixed constant ~ > O. All of our results can also be stated with ~ as a parameter at the cost of a loge 1 /~) factor in the bounds, or in terms of the expected value of £clI(m). 3 THE APPROXIMATION RATE It is apparent that any nontrivial bound on £ cv (m) must take account of some measure of the "complexity" of the unknown target function I. The correct measure of this complexity is less obvious. Following the example of Barron and Cover's analysis of MDL performance A Bound on the Error of Cross Validation 185 in the context of density estimation [2], we propose the approximation rate as a natural measure of the complexity of I and P in relation to the chosen structure HI C ... C H d •••• Thus we define the approximation rate function Eg(d) to be Eg(d) = minhEH .. {Eg(h)}. The function E 9 (d) tells us the best generalization error that can be achieved in the class H d, and it is a nonincreasing function of d. If Eg(S) = 0 for some sufficiently large s, this means that the target function I, at least with respect to the input distribution, is realizable in the class H., and thus S is a coarse measure of how complex I is. More generally, even if Eg(d) > 0 for all d, the rate of decay of Eg(d) still gives a nice indication of how much representational power we gain with respect to I and P by increasing the complexity of our models. StilI missing, of course, is some means of determining the extent to which this representational power can be realized by training on a finite sample of a given size, but this will be added shortly. First we give examples of the approximation rate that we will examine following the general bound on E ClI ( m). The Intervals Problem. In this problem, the input space X is the real interval [0,1], and the class Hd of the structure consists of all boolean step functions over [0,1] of at most d steps; thus, each function partitions the interval [0, 1] into at most d disjoint segments (not necessarily of equal width), and assigns alternating positive and negative labels to these segments. The input space is one-dimensional, but the structure contains arbitrarily complex functions over [0, 1]. It is easily verified that our assumption that the VC dimension of Hd is Oed) holds here, and that the fitting number obeys dmllZ(m) S m. Now suppose that the input density P is uniform, and suppose that the target function I is the function of S alternating segments of equal width 1/ s, for some s (thUS, I lies in the class H.). We will refer to these settings as the intervals problem. Then the approximation rate is Eg(d) = (1/2)(1 - dis) for 1 S d < sand Eg(d) = 0 for d ~ s (see Figure 1). The Perceptron Problem. In this problem, the input space X is RN for some large natural number N. The class Hd consists of all perceptrons over the N inputs in which at most d weights are nonzero. If the input density is spherically symmetric (for instance, the uniform density on the unit ball in RN ), and the target function is the function in H. with all s nonzero weights equal to 1, then it can be shown that the approximation rate is Eg(d) = (1/11") cos-I(..jd/s) for d < s [4], and of course Eg(d) = 0 for d ~ s (see Figure 1). Power Law Decay. In addition to the specific examples just given, we would also like to study reasonably natural parametric forms of Eg( d), to determine the sensitivity of our theory to a plausible range of behaviors for the approximation rate. This is important, because in practice we do not expect to have precise knowledge of Eg(d), since it depends on the target function and input distribution. Following the work of Barron [1], who shows a c/dbound on Eg(d) for the case of neural networks with one hidden layer under a squared error generalization measure (where c is a measure of target function complexity in terms of a Fourier transform integrability condition) 2, we can consider approximation rates of the form Eg(d) = (c/d)a + Emin, where Emin ~ 0 is a parameter representing the "degree of unreal izability" of I with respect to the structure, and c, a > 0 are parameters capturing the rate of decay to Emin (see Figure 1). 4 THE ESTIMATION RATE For a fixed I, P and HI C .. . C Hd· .. , we say that a function p( d, m) is an estimation rate boundifforall dand m, with high probability over the sampleSwehave IEt(hd)-Eg(hct)1 S p(d, m), where as usual hd is the result of training error minimization on S within Hd. Thus p( d, m) simply bounds the deviation between the training error and the generalization error of hd • Note that the best such bound may depend in a complicated way on all of the elements of the problem: I, P and the structure. Indeed, much of the recent work on the statistical physics theory of learning curves has documented the wide variety of behaviors that such deviations may assume [4, 3]. However, for many natural problems 2Since the bounds we will give have straightforward generalizations to real-valued function learning under squared error, examining behavior for Eg( d) in this setting seems reasonable. 186 M. KEARNS it is both convenient and accurate to rely on a universal estimation rate bound provided by the powerful theory of unifonn convergence: Namely, for any I, P and any structure, the function p(d, m) = ..j(d/m) log(m/d) is an estimation rate bound [5]. Depending upon the details of the problem, it is sometimes appropriate to omit the loge m/ d) factor, and often appropriate to refine the J dim behavior to a function that interpolates smoothly between dim behavior for small Et to Jd/m for large Et. Although such refinements are both interesting and important, many of the qualitative claims and predictions we will make are invariant to them as long as the deviation kt(hd) - Eg(hd)1 is well-approximated by a power law (d/m)a (0 > 0); it will be more important to recognize and model the cases in which power law behavior is grossly violated. Note that this universal estimation rate bound holds only under the assumption that the training sample is noise-free, but straightforward generalizations exist. For instance, if the training data is corrupted by random label noise at rate 0 ~ TJ < 1/2, then p( d, m) ..j(d/(1 - 2TJ)2m)log(m/d) is again a universal estimation rate bound. 5 THE BOUND Theorem 1 Let HI C ... C Hd · .. be any structure, where the VC dimension 0/ Hd is Oed). Let I and P be any target function and input distribution, let Eg(d) be the approximation rate/unction/or the structure with respect to I and P, and let p(d, m) be an estimation rate bound/or the structure with respect to I and P. Then/or any m, with high probability Ecv(m) ~ min {Eg(d) + p(d, (1 - ,)m)} + 0 ( I~d~di... (1) where, is the/raction o/the training sample used/or testing, and lfYmax is thefitting number dmax( (1 -,)m). Using the universal estimation bound rate and the rather weak assumption that dmax(m) is polynomial in m, we obtain that with high probability 10g«I-,)m)) . ,m (2) Straightforward generalizations 0/ these bounds/or the case where the data is corrupted by classification noise can be obtained, using the modified estimation rate bound given in Section 4 3. We delay the proof of this theorem to the full paper due to space considerations. However, the central idea is to appeal twice to uniform convergence arguments: once within each class Hd to bound the generalization error of the resulting training error minimizer hd E Hd, and a second time to bound the generalization error of the hd minimizing the error on the test set of ,m examples. In the bounds given by (1) and (2), themin{· } expression is analogous to Barron and Cover's index of resolvability [2]; the final tenn in the bounds represents the error introduced by the testing phase of cross validation. These bounds exhibit tradeoff behavior with respect to the parameter,: as we let, approach 0, we are devoting more of the sample to training the hd, and the estimation rate bound tenn p(d, (1 - ,)m) is decreasing. However, the test error tenn O( Jlog(~,u:)/(Tm)) is increasing, since we have less data to accurately estimate the Eg(hd). The reverse phenomenon occurs as we let, approach 1. While we believe Theorem 1 to be enlightening and potentially useful in its own right, we would now like to take its interpretation a step further. More precisely, suppose we ~e main effect of classification noise at rate '1 is the replacement of occurrences in the bound of the sample size m by the smaller "effective" sample size (1 - '1)2m. A Bound on the Error of Cross Validation 187 assume that the bound is an approximation to the actual behavior of EClI(m). Then in principle we can optimize the bound to obtain the best value for "Y. Of course, in addition to the assumptions involved (the main one being that p(d, m) is a good approximation to the training-generalization error deviations of the hd), this analysis can only be carried out given information that we should not expect to have in practice (at least in exact form)in particular, the approximation rate function Eg(d), which depends on f and P. However. we argue in the coming sections that several interesting qualitative phenomena regarding the choice of"Y are largely invariant to a wide range of natural behaviors for Eg (d). 6 A CASE STUDY: THE INTERVALS PROBLEM We begin by performing the suggested optimization of"Y for the intervals problem. Recall that the approximation rate here is Eg(d) = (1/2)(1 - d/8) for d < 8 and Ey(d) = 0 for d ~ 8, where 8 is the complexity of the target function. Here we analyze the behavior obtained by assuming that the estimation rate p(d, m) actually behaves as p(d, m) = Jd/(l - "Y)m (so we are omitting the log factor from the universal bound), and to simplify the formal analysis a bit (but without changing the qualitative behavior) we replace the term Jlog«1 - "Y)m)/bm) by the weaker Jlog(m)/m. Thus, if we define the function F(d, m, "Y) = Ey(d) + Jd/(1 - "Y)m + Jlog(m)/bm) then following Equation (1), we are approximating EclI(m) by EclI (m) ~ min1<d<d" {F(d, m, "Yn 4. __ maa: The first step of the analysis is to fix a value for"Y and differentiate F( d, m, "Y) with respect to d to discover the minimizing value of d; the second step is to differentiate with respect to "Y. It can be shown (details omitted) that the optimal choice of"Y under the assumptions is "Yopt = (log (m)/ 8)1/3/(1 + (Iog(m)/ 8 )1/3). It is importantto remember at this point that despite the fact that we have derived a precise expression for "Yopt. due to the assumptions and approximations we have made in the various constants, any quantitative interpretation of this expression is meaningless. However, we can reasonably expect that this expression captures the qualitative way in which the optimal "Y changes as the amount of data m changes in relation to the target function complexity 8. On this score the situation initially appears rather bleak, as the function (log( m)/ 8)1/3 /(1 + (log(m)/ 8 )1/3) is quite sensitive to the ratio log(m)/8, which is something we do not expect to have the luxury of knowing in practice. However, it is both fortunate and interesting that "Yopt does not tell the entire story. In Figure 2, we plot the function F ( 8, m, "Y) as a function of"Y for m = 10000 and for several different values of 8 (note that for consistency with the later experimental plots, the z axis of the plot is actually the training fraction 1 - "Y). Here we can observe four important qualitative phenomena, which we list in order of increasing subtlety: (A) When 8 is small compared to m, the predicted error is relatively insensitive to the choice of "Y: as a function of "Y, F( 8, m, "Y) has a wide, flat bowl, indicating a wide range of "Y yielding essentially the same near-optimal error. (B) As s becomes larger in comparison to the fixed sample size m, the relative superiority of "Yopt over other values for"Y becomes more pronounced. In particular, large values for"Y become progressively worse as s increases. For example, the plots indicate that for s = 10 (again, m = 10000), even though "Yopt = 0.524 ... the choice "Y = 0.75 will result in error quite near that achieved using "Yopt. However, for s = 500, "Y = 0.75 is predicted to yield greatly suboptimal error. Note that for very large s, the bound predicts vacuously large error for all values of "Y, so that the choice of "Y again becomes irrelevant. (C) Because of the insensitivity to "Y for s small compared to m, there is a fixed value of "Y which seems to yield reasonably good performance for a wide range of values for s. This value is essentially the value of "Yopt for the case where 8 is large but nontrivial generalization is still possible, since choosing the best value for "Y is more important there than for the small 8 case. (D) The value of "Yopt is decreasing as 8 increases. This is slightly difficult to confirm from the plot, but can be seen clearly from the precise expression for "Yopt. 4 Although there are hidden constants in the 0(.) notation of the bounds. it is the relative weights of the estimation and test error terms that is important. and choosing both constants equal to 1 is a reasonable choice (since both terms have the same Chernoff bound origins). 188 M.KEARNS In Figure 3, we plot the results of experiments in which labeled random samples of size m = 5000 were generated for a target function of s equal width intervals, for s = 10,100 and 500. The samples were corrupted by random label noise at rate TJ = 0.3. For each value of 'Y and each value of d, (1 - 'Y)m of the sample was given to a program performing training error minimization within Hd.; the remaining 'Ym examples were used to select the best hd. according to cross validation. The plots show the true generalization error of the hd. selected by cross validation as a function of'Y (the generalization error can be computed exactly for this problem). Each point in the plots represents an average over 10 trials. While there are obvious and significant quantitative differences between these experimental plots and the theoretical predictions of Figure 2, the properties (A), (B) and (C) are rather clearly borne out by the data: (A) In Figure 3, when s is small compared to m, there is a wide range of acceptable 'Y; it appears that any choice of'Y between 0.10 and 0.50 yields nearly optimal generalization error. (B) By the time s = 100, the sensitivity to'Y is considerably more pronounced. For example, the choice 'Y = 0.50 now results in clearly suboptimal performance, and it is more important to have 'Y close to 0.10. (C) Despite these complexities, there does indeed appear to be single value of'Y approximately 0.10that performs nearly optimally for the entire range of s examined. The property (D) namely, that the optimal 'Y decreases as the target function complexity is increased relative to a fixed m is certainly not refuted by the experimental results, but any such effect is simply too small to be verified. It would be interesting to verify this prediction experimentally, perhaps on a different problem where the predicted effect is more pronounced. 7 CONCLUSIONS For the cases where the approximation rate Eg(d) obeys either power law decay or is that derived for the perceptron problem discussed in Section 3, the behavior of EClI(m) as a function of 'Y predicted by our theory is largely the same (for example, see Figure 4). In the full paper, we describe some more realistic experiments in which cross validation is used to determine the number of backpropogation training epochs. Figures similar to Figures 2 through 4 are obtained, again in rough accordance with the theory. In summary, our theory predicts that although significant quantitative differences in the behavior of cross validation may arise for different model selection problems, the properties (A), (B), (C) and (D) should be present in a wide range of problems. At the very least, the behavior of our bounds exhibits these properties for a wide range of problems. It would be interesting to try to identify natural problems for which one or more of these properties is strongly violated; a potential source for such problems may be those for which the underlying learning curve deviates from classical power law behavior [4, 3]. Acknowledgements: I give warm thanks to Yishay Mansour, Andrew Ng and Dana Ron for many enlightening conversations on cross validation and model selection. Additional thanks to Andrew Ng for his help in conducting the experiments. References [1] A. Barron. Universal approximation bounds for superpositions of a sigmoidal function. IEEE Transaclions on Information Theory. 19:930-944. 1991. [2] A. R. Barron and T. M. Cover. Minimum complexity density estimation. IEEE Transaclions on Information Theory, 37:1034-1054, 1991. [3] D. Haussler, M. Kearns. H.S. Seung, and N. Tishby. Rigourous learning curve bounds from statistical mechanics. In Proceedings of the Seventh Annual ACM Confernce on Compulalional Learning Theory. pages 76-87. 1991l. [4] H. S. Seung, H. Sompolinsky. and N. Tishby. Statistical mechanics of learning from examples. Physical Review, A45:6056-6091, 1992. [5] V. N. Vapnik:. Estimalion of Dependences Based on Empirical Dala. Springer-Verlag, New York, 1982. [6] V. N. Vapnik: and A. Y. Chervonenkis. On the uniform convergence of relative frequencies of events to their probabilities. Theory of Probability and ils Applicalions. 16(2):264-280, 1971. A Bound on the Error of Cross Validation rgg pprox mat on a as v grror .. .., rr vs ra n sat s za, sno SQ, m.. . ~ .~ .. ..,J .... ¢i. CV Dun. (c ) or c rom . to ., m' .1 ... 189 Figure 1: Plots of three approximation rates: for the intervals problem with target complexity II = 250 intervals (linear plot intersecting d-axis at 250), for the perceptron problem with target complexity II = 150 nonzero weights (nonlinear plot intersecting d-axis at 150), and for power law decay asymptoting at E",," = 0.05. Figure 2: Plot of the predicted generalization error of cross validation for the intervals model selection problem, as a function of the fraction 1 "'( of data used for training. (In the plot, the fraction of training data is 0 on the left (-y = 1) and 1 on the right ("'( = 0». The fixed sample size m = 10,000 was used, and the 6 plots show the error predicted by the theory for target function complexity values II = 10 (bottom plot), 50, 100, 250, 500, and 1000 (top plot) . Figure 3: Experimental plots of cross validation generalization error in the intervals problem as a function of training set size (1-"'() m. Experiments with the three target complexity values II = 10,100 and 500 (bottom plot to top plot) are shown. Each point represents performance averaged over 10 trials . Figure 4: Plot of the predicted generalization error of cross validation for the power law case E,( d) = (c/d), as a function of the fraction 1-",(ofdata used for training. The fixed sample size m = 25,000 was used, and the 6 plots show the error predicted by the theory for target function complexity values c = 1 (bottom plot), 25,50,75, 100. and 150 (top plot).
|
1995
|
133
|
1,036
|
Classifying Facial Action Marian Stewart Bartlett, Paul A. Viola, Terrence J. Sejnowski, Beatrice A. Golomb Howard Hughes Medical Institute The Salk Institute, La Jolla, CA 92037 marni, viola, terry, beatrice @salk.edu Jan Larsen The Niels Bohr Institute 2100 Copenhagen Denmark jlarsen@fys.ku.dk Paul Ekman Joseph C. Hager Network Information Research Corp Salt Lake City, Utah jchager@ibm.net University of California San Francisco San Francisco, CA 94143 ekmansf@itsa.ucsf.edu Abstract The Facial Action Coding System, (FACS), devised by Ekman and Friesen (1978), provides an objective meanS for measuring the facial muscle contractions involved in a facial expression. In this paper, we approach automated facial expression analysis by detecting and classifying facial actions. We generated a database of over 1100 image sequences of 24 subjects performing over 150 distinct facial actions or action combinations. We compare three different approaches to classifying the facial actions in these images: Holistic spatial analysis based on principal components of graylevel images; explicit measurement of local image features such as wrinkles; and template matching with motion flow fields. On a dataset containing six individual actions and 20 subjects, these methods had 89%, 57%, and 85% performances respectively for generalization to novel subjects. When combined, performance improved to 92%. 1 INTRODUCTION Measurement of facial expressions is important for research and assessment psychiatry, neurology, and experimental psychology (Ekman, Huang, Sejnowski, & Hager, 1992), and has technological applications in consumer-friendly user interfaces, interactive video and entertainment rating. The Facial Action Coding System (FACS) is a method for measuring facial expressions in terms of activity in the underlying facial muscles (Ekman & Friesen, 1978). We are exploring ways to automate FACS. 824 BARTLETI, VIOLA, SEJNOWSKI, GOLOMB, LARSEN, HAGER, EKMAN Rather than classifying images into emotion categories such as happy, sad, or surprised, the goal of this work is instead to detect the muscular actions that comprise a facial expression. FACS was developed in order to allow researchers to measure the activity of facial muscles from video images of faces. Ekman and Friesen defined 46 distinct action units, each of which correspond to activity in a distinct muscle or muscle group, and produce characteristic facial distortions which can be identified in the images. Although there are static cues to the facial actions, dynamic information is a critical aspect of facial action coding. FACS is currently used as a research tool in several branches of behavioral science, but a major limitation to this system is the time required to both train human experts and to manually score the video tape. Automating the Facial Action Coding System would make it more widely accessible as a research tool, and it would provide a good foundation for human-computer interactions tools. Why Detect Facial Actions? Most approaches to facial expression recognition by computer have focused on classifying images into a small set of emotion categories such as happy, sad, or surprised (Mase, 1991; Yacoob & Davis, 1994; Essa & Pentland, 1995). Real facial signals, however, consist ofthousands of distinct expressions, that differ often in only subtle ways. These differences can signify not only which emotion is occurring, but whether two or more emotions have blended together, the intensity of the emotion(s), and if an attempt is being made to control the expression of emotion (Hager & Ekman, 1995). An alternative to training a system explicitly on a large number of expression categories is to detect the facial actions that comprise the expressions. Thousands of facial expressions can be defined in terms of this smaller set of structural components. We can verify the signal value of these expressions by reference to a large body of behavioral data relating facial actions to emotional states which have already been scored with FACS. FACS also provides a meanS for obtaining reliable training data. Other approaches to automating facial measurement have mistakenly relied upon voluntary expressions, which tend to contain exaggerated and redundant cues, while omitting some muscular actions altogether (Hager & Ekman, 1995). 2 IMAGE DATABASE We have collected a database of image sequences of subjects performing specified facial actions. The full database contains over 1100 sequences containing over 150 distinct actions, or action combinations, and 24 different subjects. The sequences contain 6 images, beginning with a neutral expression and ending with a high intensity muscle contraction (Figure 1). For our initial investigation we used data from 20 subjects and attempted to classify the six individual upper face actions illustrated in Figure 2. The information that is available in the images for detecting and discriminating these actions include distortions in the shapes and relative positions of the eyes and eyebrows, the appearance of wrinkles, bulges, and furrows, in specific regions of the face, and motion of the brows and eyelids. Prior to classifying the images, we manually located the eyes, and we used this information to crop a region around the upper face and scale the images to 360 x 240. The images were rotated so that the eyes were horizontal, and the luminance was normalized. Accurate image registration is critical for principal components based approaches. For the holistic analysis and flow fields, the images were further scaled Classifying Facial Action 825 to 22 x 32 and 66 x 96, respectively. Since the muscle contractions are frequently asymmetric about the face, we doubled the size of our data set by reflecting each image about the vertical axis, giving a total of 800 images. Figure 1: Example action sequences from the database. AU 1 AU2 AU4 AU5 AU6 AU7 Figure 2: Examples of the six actions used in this study. AU 1: Inner brow raiser. 2: Outer brow raiser. 4: Brow lower. 5: Upper lid raiser (widening the eyes). 6: Cheek raiser. 7: Lid tightener (partial squint). 3 HOLISTIC SPATIAL ANALYSIS The Eigenface (Thrk & Pentland, 1991) and Holon (Cottrell & Metcalfe, 1991) representations are holistic representations based on principal components, which can be extracted by feed forward networks trained by back propagation. Previous work in our lab and others has demonstrated that feed forward networks taking such holistic representations as input can successfully classify gender from facial images (Cottrell & Metcalfe, 1991; Golomb, Lawrence, & Sejnowski, 1991). We evaluated the ability of a back propagation network to classify facial actions given principal components of graylevel images as input. The primary difference between the present approach and the work referenced above is that we take the principal components of a set of difference images, which we obtained by subtracting the first image in the sequence from the subsequent images (see Figure 3). The variability in our data set is therefore due to the facial distortions and individual differences in facial distortion, and we have removed variability due to surface-level differences in appearance. We projected the difference images onto the first N principal components of the dataset, and these projections comprised the input to a 3 layer neural network with 10 hidden units, and six output units, one per action (Figure 3.) The network is feed forward and fully connected with a hyperbolic tangent transfer function, and was trained with conjugate gradient descent. The output of the network was determined using winner take all, and generalization to novel subjects was determined by using the leave-one-out, or jackknife, procedure in which we trained the network on 19 subjects and reserved all of the images from one subject for testing. This process was repeated for each of the subjects to obtain a mean generalization performance across 20 test cases. 826 BARTLETI, VIOLA, SEJNOWSKI, GOLOMB, LARSEN, HAGER, EKMAN We obtained the best performance with 50 component projections, which gave 88.6% correct across subjects. The benefit obtained by using principal components over the 704-dimensional difference images themselves is not large. Feeding the difference images directly into the network gave a performance of 84% correct. 6 OUtputs I WT A Figure 3: Left: Example difference image. Input values of -1 are mapped to black and 1 to white. Right: Architecture of the feed forward network. 4 FEATURE MEASUREMENT We turned next to explicit measurement of local image features associated with these actions. The presence of wrinkles in specific regions of the face is a salient cue to the contraction of specific facial muscles. We measured wrinkling at the four facial positions marked in Figure 4a, which are located in the image automatically from the eye position information. Figure 4b shows pixel intensities along the line segment labeled A, and two major wrinkles are evident. We defined a wrinkle measure P as the sum of the squared derivative of the intensity values along the segment (Figure 4c.) Figure 4d shows P values along line segment A, for a subject performing each of the six actions. Only AU 1 produces wrinkles in the center of the forehead. The P values remain at zero except for AU 1, for which it increases with increases in action intensity. We also defined an eye opening measure as the area of the visible sclera lateral to the iris. Since we were interested in changes in these measures from baseline, we subtract the measures obtained from the neutral image. a c p b 0'---____ --' Pixel d 3....----------, ....-----, 2 c( c.. 1 ~ _~ .... ----to ~.~~ -.... --']t;. ... ~.t(=-: .... -.... ="!1t:-:-:-..:-::!:: 2 o 1 234 5 Image in Seqence AU1 __ AU2 -+-. AU4 -E!-. AU5 -K-· AU6-4-· AU7-ll--Figure 4: a) Wrinkling was measured at four image locations, A-D. b) Smoothed pixel intensities along the line labeled A. c) Wrinkle measure. d) P measured at image location A for one subject performing each of the six actions. We classified the actions from these five feature measures using a 3-layer neural net with 15 hidden units. This method performs well for some subjects but not for Classifying Facial Action 827 Figure 5: Example flow field for a subject performing AU 7, partial closure of the eyelids. Each flow vector is plotted as an arrow that points in the direction of motion. Axes give image location. others, depending on age and physiognomy. It achieves an overall generalization performance of 57% correct. 5 OPTIC FLOW The motion that results from facial action provides another important source of information. The third classifier attempts to classify facial actions based only on the pattern of facial motion. Motion is extracted from image pairs consisting of a neutral image and an image that displays the action to be classified. An approximation to flow is extracted by implementing the brightness constraint equation (2) where the velocity (vx,Vy) at each image point is estimated from the spatial and temporal gradients of the image I. The velocities can only be reliably extracted at points of large gradient, and we therefore retain only the velocities from those locations. One of the advantages of this simple local estimate of flow is speed. It takes 0.13 seconds on a 120 MHz Pentium to compute one flow field. A resulting flow image is illustrated in Figure 5. 8I(x, y, t) 8I(x, y, t) 8I(x, y, t) _ 0 Vx 8x + Vy 8y + 8t (2) We obtained weighted templates for each of the actions by taking mean flow fields from 10 subjects. We compared novel flow patterns, r to the template ft by the similarity measure S (3). S is the normalized dot product of the novel flow field with the template flow field. This template matching procedure gave 84.8% accuracy for novel subjects. Performance was the same for the ten subjects used in the training set as for the ten in the test set. (3) 6 COMBINED SYSTEM Figure 6 compares performance for the three individual methods described in the previous sections. Error bars give the standard deviation for the estimate of generalization to novel subjects. We obtained the best performance when we combined all three sources of information into a single neural network. The classifier is a 828 BAR1LETI, VIOLA, SEJNOWSKI, GOLOMB, LARSEN, HAGER, EKMAN I 6 Output I WTA 11Ol:i Classifier Figure 6: Left: Combined system architecture. Right: Performance comparisons. Holistic v. Flow r :0.52 • i g ~ ~ 50 Feature v. Row r :0.26 • • • • 60 70 80 90 • 100 Feature v. Holistic i r:O.OO Ii! ~~--...;......,..".~ ~:-----50 60 70 80 90 100 Figure 7: Performance correlations among the three individual classifiers. Each data point is performance for one of the 20 subjects. feed forward network taking 50 component projections, 5 feature measures, and 6 template matches as input (see Figure 6.) The combined system gives a generalization performance of 92%, which is an improvement over the best individual method at 88.6%. The increase in performance level is statistically significant by a paired t-test. While the improvement is small, it constitutes about 30% of the difference between the best individual classifier and perfect performance. Figure 6 also shows performance of human subjects on this same dataset. Human non-experts can correctly classify these images with about 74% accuracy. This is a difficult classification problem that requires considerable training for people to be able to perform well. We can examine how the combined system benefits from multiple input sources by looking at the cprrelations in performance of the three individual classifiers. Combining estimators is most beneficial when the individual estimators make very different patterns of errors.1 The performance of the individual classifiers are compared in Figure 7. The holistic and the flow field classifiers are correlated with a coefficient of 0.52. The feature based system, however, has a more independent pattern of errors from the two template-based methods. Although the stand-alone performance of the featurebased system is low, it contributes to the combined system because it provides estimates that are independent from the two template-based systems. Without the feature measures, we lose 40% of the improvement. Since we have only a small number of features, this data does not address questions about whether templates are better than features, but it does suggest that local features plus templates may be superior to either one alone, since they may have independent patterns of errors. iTom Dietterich, Connectionists mailing list, July 24, 1993. Classifying Facial Action 829 7 DISCUSSION We have evaluated the performance of three approaches to image analysis on a difficult classification problem. We obtained the best performance when information from holistic spatial analysis, feature measurements, and optic flow fields were combined in a single system. The combined system classifies a face in less than a second on a 120 MHz Pentium. Our initial results are promising since the upper facial actions included in this study represent subtle distinctions in facial appearance that require lengthy training for humans to make reliably. Our results compare favorably with facial expression recognition systems developed by Mase (1991), Yacoob and Davis (1994), and Padgett and Cottrell (1995), who obtained 80%, 88%, and 88% accuracy respectively for classifying up to six full face expressions. The work presented here differs from these systems in that we attempt to detect individual muscular actions rather than emotion categories, we use a dataset of labeled facial actions, and our dataset includes low and medium intensity muscular actions as well as high intensity ones. Essa and Pentland (1995) attempt to relate facial expressions to the underlying musculature through a complex physical model of the face. Since our methods are image-based, they are more adaptable to variations in facial structure and skin elasticity in the subject population. We intend to apply these techniques to the lower facial actions and to action combinations as well. A completely automated method for scoring facial actions from images would have both commercial and research applications and would reduce the time and expense currently required for manual scoring by trained observers. Acknow ledgments This research was supported by Lawrence Livermore National Laboratories, IntraUniversity Agreement B291436, NSF Grant No. BS-9120868, and Howard Hughes Medical Institute. We thank Claudia Hilburn for image collection. References Cottrell, G.,& Metcalfe, J. (1991): Face, gender and emotion recognition using holons. In Advances in Neural Information Processing Systems 9, D. Touretzky, (Ed.) San Mateo: Morgan & Kaufman. 564 - 571. Ekman, P., & Friesen, W. (1978): Facial Action Coding System: A Technique for the Measurement of Facial Movement. Palo Alto, CA: Consulting Psychologists Press. Ekman, P., Huang, T., Sejnowski, T., & Hager, J. (1992): Final Report to NSF of the Planning Workshop on Facial Expression Understanding. Available from HIL-0984, UCSF, San Francisco, CA 94143. Essa, I., & Pentland, A. (1995). Facial expression recognition using visually extracted facial action parameters. Proceedings of the International Workshop on Automatic Face- and Gesture-Recognition. University of Zurich, Multimedia Laboratory. Golomb, B., Lawrence, D., & Sejnowski, T. (1991). SEXnet: A neural network identifies sex from human faces. In Advances in Neural Information Processing Systems 9, D. Touretzky, (Ed.) San Mateo: Morgan & Kaufman: 572 - 577. Hager, J., & Ekman, P., (1995). The essential behavioral science of the face and gesture that computer scientists need to know. Proceedings of the International Workshop on Automatic Face- and Gesture-Recognition. University of Zurich, Multimedia Laboratory. Mase, K. (1991): Recognition of facial expression from optical flow. IEICE Transactions E 74(10): 3474-3483. Padgett, C., Cottrell, G., (1995). Emotion in static face images. Proceedings of the Institute for Neural Computation Annual Research Symposium, Vol 5. La Jolla, CA. Turk, M., & Pentland, A. (1991): Eigenfaces for Recognition. Journal of Cognitive Neuroscience 3(1): 71 - 86. Yacoob, Y., & Davis, L. (1994): Recognizin~ human facial expression. University of Maryland Center for Automation Research Technical Report No. 706.
|
1995
|
134
|
1,037
|