index
int64 0
20.3k
| text
stringlengths 0
1.3M
| year
stringdate 1987-01-01 00:00:00
2024-01-01 00:00:00
| No
stringlengths 1
4
|
|---|---|---|---|
1,100
|
Family Discovery Stephen M. Omohundro NEC Research Institute 4 Independence Way, Princeton, N J 08540 om@research.nj.nec.com Abstract "Family discovery" is the task of learning the dimension and structure of a parameterized family of stochastic models. It is especially appropriate when the training examples are partitioned into "episodes" of samples drawn from a single parameter value. We present three family discovery algorithms based on surface learning and show that they significantly improve performance over two alternatives on a parameterized classification task. 1 INTRODUCTION Human listeners improve their ability to recognize speech by identifying the accent of the speaker. "Might" in an American accent is similar to "mate" in an Australian accent. By first identifying the accent, discrimination between these two words is improved. We can imagine locating a speaker in a "space of accents" parameterized by features like pitch, vowel formants, "r" -strength, etc. This paper considers the task of learning such parameterized models from data. Most speech recognition systems train hidden Markov models on labelled speech data. Speaker-dependent systems train on speech from a single speaker. Speakerindependent systems are usually similar, but are trained on speech from many different speakers in the hope that they will then recognize them all. This kind of training ignores speaker identity and is likely to result in confusion between pairs of words which are given the same pronunciation by speakers with different accents. Speaker-independent recognition systems could more closely mimic the human approach by using a learning paradigm we call "family discovery". The system would be trained on speech data partitioned into "episodes" for each speaker. From this data, the system would construct a parameterized family of models representing difFamily Discovery 403 Affine Family Affine Patch Family Coupled Map Family Figure 1: The structure of the three family discovery algorithms. ferent accents. The learning algorithms presented in this paper could determine the dimension and structure of the parameterization. Given a sample of new speech, the best-fitting accent model would be used for recognition. The same paradigm applies to many other recognition tasks. For example, an OCR system could learn a parameterized family of font models (Revow, et. al., 1994). Given new text, the system would identify the document's font parameters and use the corresponding character recognizer. In general, we use "family discovery" to refer to the task of learning the dimension and structure of a parameterized family of stochastic models. The methods we present are equally applicable to parameterized density estimation, classification, regression, manifold learning, reinforcement learning, clustering, stochastic grammar learning, and other stochastic settings. Here we only discuss classification and primarily consider training examples which are explicitly partitioned into episodes. This approach fits naturally into the neural network literature on "meta-learning" (Schmidhuber, 1995) and "network transfer" (Pratt, 1994). It may also be considered as a particular case of the "bias learning" framework proposed by Baxter at this conference (Baxter, 1996). There are two primary alternatives to family discovery: 1) try to fit a single model to the data from all episodes or 2) use separate models for each episode. The first approach ignores the information that the different training sets came from distinct models. The second approach eliminates the possibility of inductive generalization from one set to another. In Section 2, we present three algorithms for family discovery based on techniques for "surface learning" (Bregler and Omohundro, 1994 and 1995). As shown in Figure 1, the three alternative representations of the family are: 1) a single affine subspace of the parameter space, 2) a set of local affine patches smoothly blended together, and 3) a pair of coupled maps from the parameter space into the model space and back. In Section 3, we compare these three approaches to the two alternatives on a parameterized classification task. 404 S. M. OMOHUNDRO 2 THE FIVE ALGORITHMS Let the space of all classifiers under consideration be parameterized by 0 and assume that different values of 0 correspond to different classifiers (ie. it is identifiable). For example, 0 might represent the means, covariances, and class priors of a classifier with normal class-conditional densities. O-space will typically have a much higher dimension than the parameterized family we are seeking. We write P9(X) for the total probability that the classifier 0 assigns to a labelled or unlabelled example x. The true models are drawn from a d-dimensional family parameterized by , . Let the training set be partitioned into N episodes where episode i consists of Ni training examples tij, 1 :S j :S Ni drawn from a single underlying model with parameter 0:. A family discovery learning algorithm uses this training data to estimate the underlying parameterized family. From a parameterized family, we may define the projection operator P from O-space to itself which takes each 0 to the closest member of the family. Using this projection operator, we may define a "family prior" on O-space which dies off exponentially with the square distance of a model from the family mp(O) ex e-(9-P(9))2. Each of the family discovery algorithms chooses a family so as to maximize the posterior probability of the training data with respect to this prior. If the data is very sparse, this MAP approximation to a full Bayesian solution can be supplemented by "Occam" terms (MacKay, 1995) or by using a Monte Carlo approximation. The outer loop of each of the algorithms performs the optimization of the fit of the data by re-estimation in a manner similar to the Expectation Maximization (EM) approach (Jordan and Jacobs, 1994). First, the training data in each episode i is independently fit by a model Oi. Then the dimension of the family is determined as described later and the family projection operator P is chosen to maximize the probability that the episode models Oi came from that family ni mp(Oi). The episode models Oi are then re-estimated including the new prior probability mp. These newly re-estimated models are influenced by the other episodes through mp and so exhibit training set "transfer". The re-estimation loop is repeated until nothing changes. The learned family can then be used to classify a set of Ntest unlabelled test examples Xk, 1 :S k :S Ntest drawn from a model O;est in the family. First, the parameter Otest is estimated by selecting the member of the family with the highest likelihood on the test samples. This model is then used to perform the classification. A good approximation to the best-fit family member is often to take the image of the best-fit model in the entire O-space under the projection operator P. In the next five sections, we describe the two alternative approaches and the three family discovery algorithms. They differ only in their choice of family representation as encoded in the projection operator P. 2.1 The Single Model Approach The first alternative approach is to train a single model on all of the training data. It selects 0 to maximize the total likelihood L( 0) = n~l n~l P9 (tij ). New test data is classified by this single selected model. Family Discovery 405 2.2 The Separate Models Approach The second alternative approach fits separate models for each training }£isode. It chooses Bi for 1::; i::; N to maximize the episode likelihood Li(Bi) = TIj~IPIJ(tij). Given new test data, it determines which of the individual models Bi fit best and classifies the data with it. 2.3 The Affine Algorithm The affine model represents the underlying model family as an affine subspace of the model parameter space. The projection operator Pal line projects a parameter vector B orthogonally onto the affine subspace. The subspace is determined by selecting the top principal vectors in a principal components analysis of the bestfit episode model parameters. As described in (Bregler & Omohundro, 1994) the dimension is chosen by looking for a gap in the principal values. 2.4 The Affine Patch Algorithm The second family discovery algorithm is based on the "surface learning" procedure described in (Bregler and Omohundro, 1994). The family is represented by a collection of local affine patches which are blended together using Gaussian influence functions. The projection mapping Ppatch is a smooth convex combination of projections onto the affine patches Ppatch(B) = 2::=1 10: (B)Ao: (B) where Ao: is the projection operator for an affine patch and Io:(B) = E:"J:)(IJ) is a normalized Gaussian blending function. The patches are initialized using k-means clustering on the episode models to choose k patch centers. A local principal components analysis is performed on the episode models which are closest to each center. The family dimension is determined by examining how the principal values scale as successive nearest neighbors are considered. Each patch may be thought of as a "pancake" lying in the surface. Dimensions which belong to the surface grow quickly as more neighbors are considered while dimensions across the surface grow only because of the curvature of the surface. The Gaussian influence functions and the affine patches are then updated by the EM algorithm (Jordan and Jacobs, 1994). With the affine patches held fixed, the Gaussians Go: are refit to the errors each patch makes in approximating the episode models. Then with the Gaussians held fixed, the affine patches Ao: are refit to the epsiode models weighted by the the corresponding Gaussian Go:. Similar patches may be merged together to form a more parsimonious model. 2.5 The Coupled Map Algorithm The affine patch approach has the virtue that it can represent topologically complex families (eg. families representing physical objects might naturally be parameterized by the rotation group which is topologically a projective plane). It cannot, however, provide an explicit parameterization of the family which is useful in some applications (eg. optimization searches). The third family discovery algorithm therefore attempts to directly learn a parameterization of the model family. Recall that the model parameters define B-space, while the family parameters de406 S. M. OMOHUNDRO fine 'Y-space. We represent a family by a mapping G from B-space to 'Y-space together with a mapping F from 'Y-space back to B-space. The projection operation is Pmap(B) = F(G(B)). The map G(O) defines the family parameter l' on the full O-space. This representation is similar to an "auto-associator" network in which we attempt to "encode" the best-fit episode parameters Oi in the lower dimensional 'Y-space by the mapping G in such a way that they can be correctly reconstructed by the function F. Unfortunately, if we try to train F and G using back-propagation on the identity error function, we get no training data away from the family. There is no reason for G to project points away from the family to the closest family member. We can rectify this by training F and G iteratively. First an arbitrary G is chosen and F is trained to send the images 'Yi = G(Oi) back to 0i' G is trained, however, on images under F corrupted by additive spherical Gaussian noise! This provides samples away from the family and on average the training signal sends each point in B space to the closest family member. To avoid iterative training, our experiments used a simpler approach. G was taken to be the affine projection operator defined by a global principal components analysis of the best-fit episode model parameters. Once G is defined, F is chosen to minimize the difference between F(G(Oi)) and Oi for each best-fit episode parameter Oi. Any form of trainable nonlinear mapping could be used for F (eg. backprop neural networks or radial basis function networks). We represent F as a mixture of experts (Jordan and Jacobs, 1994) where each expert is an affine mapping and the mixture coefficients are Gaussians. The mapping is trained by the EM algorithm. 3 ALGORITHM COMPARISON To compare these five algorithms, we consider a two-class classification task with unit-variance normal class-conditional distributions on a 5-dimensional feature space. The means of the class distributions are parameterized by a nonlinear twoparameter family: ml = (1'1 + ~cos¢»e~1 + ('Y2 + ~sin¢»e~2 m2 = ('Yl ~ cos ¢> ) e~1 + ('Y2 ~ sin ¢> ) l2 . where 0 ~ 1'1, 1'2 ~ 10 and ¢> = ('Yl + 1'2)/3. The class means are kept at a unit distance apart, ensuring significant class overlap over the whole family. The angle ¢> varies with the parameters so that the correct classification boundary changes orientation over the family. This choice of parameters introduces sufficient nonlinearity in the task to distinguish the non-linear algorithms from the linear one. Figure 1 shows the comparative performance of the 5 algorithms. The x-axis is the total number of training examples. Each set of examples consisted of approximately N = ..;x episodes of approximately Ni = ..;x examples each. The classifier parameters for an episode were drawn uniformly from the classifier family. The episode training examples were then sampled from the chosen classifier according to the classifier's distribution. Each of the 5 algorithms was then trained on these examples. The number of patches in the surface patch algorithm and the number of affine components in the surface map algorithm were both taken to be the square-root of Family Discovery 407 0.52 r---.---.---""T""----r----,-----r---r---~-__, 0.5 0.48 0.46 I!? 0.44 g w '0 0.42 c: 0 :u I!! 0.4 u. 0.38 0.36 0.34 400 600 800 1000 1200 1400 Number of Examples Single model -+Separate models -+-_. Affine family -EJ -Affine Patch family ··x···· Map Mixture family -A-.1600 1800 2000 Figure 2: A comparison of the 5 family discovery algorithms on the classification task. the number of training episodes. The y-axis shows the percentage correct for each algorithm on an independent test set. Each test set consisted of 50 episodes of 50 examples each. The algorithms were presented with unlabelled data and their classification predictions were then compared with the correct classification label. The results show significant improvement through the use of family discovery for this classification task. The single model approach performed significantly worse than any of the other approaches, especially for larger numbers of episodes (where the family discovery becomes possible). The separate model approach improves with the number of episodes, but is nearly always bested by the approaches which take explicit account of the underlying parameterized family. Because of the nonlinearity in this task, the simple affine model performs more poorly than the two nonlinear methods. It is simple to implement, however, and may well be the method of choice when the parameters aren't so nonlinear. From this data, there is not a clear winner between the surface patch and surface map approaches. 4 TRAINING SET DISCOVERY Throughout this paper, we have assumed that the training set was partitioned into episodes by the teacher. Agents interacting with the world may not be given this explicit information. For example, a speech recognition system may not be told when it is conversing with a new speaker. Similarly, a character recognition system 408 s. M. OMOHUNDRO would probably not be given explicit information about font changes. Learners can sometimes use the data itself to detect these changes, however. In many situations there is a strong prior that successive events are likely to have come from a single model with only occasional model changes. The EM algorithm is often used for segmenting unlabelled speech. It may be used in a similar manner to find the training set episode boundaries. First, a clustering algorithm is used to partition the training examples into episodes. A parameterized family is then fit to these episodes. The data is then repartitioned according to the similarity of the induced family parameters and the process is repeated until it converges. A similar approach may be applied when the model parameters vary slowly with time rather than occasionally jumping discontinously. Acknowledgements I'd like to thank Chris Bregler for work on the affine patch approach to surface learning, Alexander Linden for suggesting coupled maps for surface learning, and Peter Blicher for discussions. References Baxter, J. (1995) Learning model bias. This volume. Bregler, C. & Omohundro, S. (1994) Surface learning with applications to lipreading. In J. Cowan, G. Tesauro and J. Alspector (eds.), Advances in Neural Information Processing Systems 6, pp. 43-50. San Francisco, CA: Morgan Kaufmann Publishers. Bregler, C. & Omohundro, S. (1995) Nonlinear image interpolation using manifold learning. In G. Tesauro, D. Touretzky and T. Leen (eds.), Advances in Neural Information Processing Systems 7. Cambridge, MA: MIT Press. Bregler, C. & Omohundro, S. (1995) Nonlinear manifold learning for visual speech recognition. In W. Grimson (ed.), Proceedings of the Fifth International Conference on Computer Vision. Jordan, M. & Jacobs, R. (1994) Hierarchical mixtures of experts and the EM algorithm. Neural Computation, 6:181-214. MacKay, D. (1995) Probable networks and plausible predictions - a review of practical Bayesian methods for supervised neural networks. Network, to appear. Pratt, L. (1994) Experiments on the transfer of knowledge between neural networks. In S. Hanson, G. Drastal, and R. Rivest (eds.), Computational Learning Theory and Natural Learning Systems, Constraints and Prospects, pp. 523-560. Cambridge, MA: MIT Press. Revow, M., Williams, C. and Hinton, G. (1994) Using generative models for handwritten digit recognition. Technical report, University of Toronto. Schmidhuber, J. (1995) On learning how to learn learning strategies. Technical Report FKI-198-94, Fakultat fur Informatik, Technische Universitat Munchen.
|
1995
|
55
|
1,101
|
Modeling Interactions of the Rat's Place and Head Direction Systems A. David Redish and David S. Touretzky Computer Science Department & Center for the Neural Basis of Cognition Carnegie Mellon University, Pittsburgh PA 15213-3891 Internet: {dredi sh, ds t}@es . emu. edu Abstract We have developed a computational theory of rodent navigation that includes analogs of the place cell system, the head direction system, and path integration. In this paper we present simulation results showing how interactions between the place and head direction systems can account for recent observations about hippocampal place cell responses to doubling and/or rotation of cue cards in a cylindrical arena (Sharp et at., 1990). Rodents have multiple internal representations of their relationship to their environment. They have, for example, a representation of their location (place cells in the hippocampal formation, see Muller et at., 1991), and a location-independent representation of their heading (head direction cells in the postsubiculum and the anterior thalamic nuclei, see Taube et at., 1990; Taube, 1995). If these representations are to be used for navigation, they must be aligned consistently whenever the animal reenters a familiar environment. This process was examined in a set of experiments by Sharp et at. (1990). 1 The Sharp et al., 1990 experiment Rats spent multiple sessions finding food scattered randomly on the floor of a black cylindrical arena with a white cue card along the wall subtending 90° of arc. The animals were not disoriented before entering the arena, and they always entered at the same location: the northwest corner. See Figure 3a. Hippocampal place fields were mapped by single-cell recording. A variety of probe trials were then introduced. When an identical second cue 62 Head r-----~ Direction ~ Path r-----'--~ Integral_.I.o ........ .J (xp,Y,,> Goal Memory Local View (T~ 'I' .;) Place COde A (It) A. D. REDISH, D. S. TOURETZKY Figure 1: Organization of the rodent navigation model. card was added opposite the first (Figure 3c), most place fields did not double. J Instead, the cells continued to fire at their original locations. However, if the rat was introduced into the double-card environment at the southeast corner (Figure 3d), the place fields rotated by 1800 • But rotation did not occur in single-card probe trials with a southeast entry point (Figure 3b). When tested with cue cards rotated by ±30°, Sharp et al. observed that place field locations were controlled by an interaction of the choice of entry point with the cue card positions (Figure 3f.) 2 The CRAWL model In earlier work (Wan et al., 1994a; Wan et al., 1994b; Redish and Touretzky, 1996) we described a model of rodent navigation that includes analogs of both place cells and the head direction system. This model also includes a local view module representing egocentric spatial information about landmarks, and a separate metric representation of location which serves as a substrate for path integration. The existence of a path integration faculty in rodents is strongly supported by behavioral data; see Maurer and Seguinot (1995) for a discussion. Hypotheses about the underyling neural mechanismss are presently being explored by several researchers, including us. The structure of our model is shown in Figure 1. Visual inputs are represented as triples of form (Ti, 'i, (Ji), each denoting the type, distance, and egocentric bearing ofa landmark. The experiments reported here used two point-type landmarks representing the left and right edges of the cue card, and one surface-type landmark representing the arena wall. For the latter, 'i and (Ji define the normal vector between the rat and the surface. In the local view module, egocentric bearings (Ji are converted to allocentric form <Pi by adding the current value represented in the head direction system, denoted as tPh . The visual angle CYij between pairs of landmarks is also part of the local view, and can be used to help localize the animal when its head direction is unknown. See Figure 2. I Five of the 18 cells recorded by Sharp et al. changed their place fields over the various recording sessions. Our model does not reproduce these effects, since it does not address changes in place cell tuning. Such changes could occur due to variations in the animal's mental state from one trial to the next, or as a result of learning across trials. Modeling Interactions of the Rat's Place and Head Direction Systems (T., r., 4>.) } 1 1 63 Figure 2: Spatial variables used in tuning a place cell to two landmarks i and j when the animal is at path integrator coordinates (xl" Yl') . Our simulated place units are radial basis functions tuned to combinations of individual landmark bearings and distances, visual angles between landmark pairs, and path integrator coordinates. Place units can be driven by visual input alone when the animal is trying to localize itself upon initial entry at a random spot in the environment, or by the path integrator alone when navigating in the dark. But normally they are driven by both sources simultaneously. A key role of the place system is to maintain associations between the two representations, so that either can be reconstructed from the other. The place system also maintains a record of allocentric bearings of landmarks when viewed from the current position; this enables the local view module to compare perceived with remembered landmark bearings, so that drift in the head direction system can be detected and corrected. In computer simulations using a single parameter set, the model reproduces a variety of behavioral and neurophysiological results including control of place fields by visual landmarks, persistence of place fields in the dark, and place fields drifting in synchrony with drift in the head direction system. Its predictions for open-field landmark-based navigation behavior match many of the experimental results of Collett et al. (1986) for gerbils. 2.1 Entering a familiar environment Upon entering a familiar environment, the model's four spatial representations (local view, head direction, place code, and path integrator coordinates) must be aligned with the current sensory input and with each other. Note that local view information is completely determined given the visual input and head direction, and place cell activity is completely determined given the local view and path integrator representations. Thus, the alignment process manipulates just two variables: head direction and path integrator coordinates. When the animal enters the environment with initial estimates for them, the alignment process can produce four possible outcomes: (1) Retain the initial values of both variables, (2) Reset the head direction, (3) Reset the path integrator, or (4) Reset both head direction and the path integrator. 2.2 Prioritizing the outcomes When the animal was placed at the northwest entry point and there were two cue cards (Figure 3c), we note that the orientation of the wall segment adjacent to the place field is identical with that in the training case. This suggests that the animal's head direction 64 A. D. REDISH, D. S. TOURETZKY did not change. The spatial relationship between the entry point and place field was also unchanged: notice that the distance from the entry point to the center of the field is the same as in Figure 3a. Therefore, we conclude that the initially estimated path integrator coordinates were retained. Alternatively, the animal could have changed both its head direction (by 180°) and its path integrator coordinates (to those of the southeast comer) and produced consistent results, but to the experimenter the place field would appear to have flipped to the other card. Because no flip was observed, the first outcome must have priority over the fourth. In panel d, where the place field has flipped to the northwest comer, the orientation of the segment of wall adjacent to the field has changed, but the spatial relationship between the entry point and field center has not. Resetting the path integrator and not the head direction would also give a solution consistent with this local view, but with the place field unflipped (as in panel b). We conclude that the second outcome (reset head direction) must have priority over the third (reset the path integrator). The third and fourth outcomes are demonstrated in Figures 3b and 3f. In panel b, the orientation of the wall adjacent to the place field is unchanged from panel a, but the spatial relationship between the entry point and the place field center is different, as evidenced by the fact that the distance between them is much reduced. This is outcome 3. In panel f, both variables have changed (outcome 4). Finally, the fact that place fields are stable over an entire session, even when there are multiple cue cards (and therefore multiple consistent pairings of head directions and path integrator coordinates) implies that animals do not reset their head direction or path integrator in visually ambiguous environments as long as the current values are reasonably consistent with the local view. We therefore assume that outcome 1 is preferred over the others. This analysis establishes a partial ordering over the four outcomes: 1 is preferred over 4 by Figure 3c, and over the others by the stability of place fields, and outcome 2 is preferred over 3 by Figure 3d. This leaves open the question of whether outcome 3 or 4 has priority over the other. In this experiment, after resetting the path integrator it's always safe for the animal to attempt to reset its head direction. If the head direction does not change by more than a few degrees, as in panel b, we observe outcome 3; if it does change substantially, as in panel f, we observe outcome 4. 2.3 Consistency The viability of an outcome is a function of the consistency between the local view and path integrator representations. The place system maintains the association between the two representations and mediates the comparison between them. The activity A(u) of a place unit is the product of a local view term LV(u) and a path integrator term C(u). LV(u) is in turn a product of five Gaussians: two tuned to bearings and two to distances (for the same' pair of landmarks), and one tuned to the retinal angle between a pair of landmarks. C(u) is a Gaussian tuned to the path integrator coordinates of the center of the place field. If the two representations agree, then the place units activated by path integrator input will be the same as those activated by the local view module, so the product A(u) computed by those units will be significantly greater than zero. The consistency K, of the association Modeling Interactions of the Rat's Place and Head Direction Systems 65 between path integrator and local view representations is given by: K, = Lu A(u)/ Lu C(u). Because A(u) < C(u) for all place units, K, ranges between 0 and 1. When the current local view is compatible with that predicted by the current path integrator coordinates, K, will be high; when the two are not compatible, K, will be low. Earlier we showed that the navigation system should choose the highest priority viable outcome. If the consistency of an outcome is more than K, * better than all higher-priority outcomes, that outcome is a viable choice and higher-priority ones are not. K,* is an empirically derived constant that we have set equal to 0.04. 3 Discussion Our results match all of the cases already discussed. (See Figure 3, panels a through d as well as f and h.) Sharp et al. (1990) did not actually test the rotated cue cards with a northwest entry point, so our result in panel e is a prediction. When the animals entered from the northwest, but only one cue card was available at 1800 , Sharp et al. report that the place field did not rotate. In our model the place field does rotate, as a result of outcome 4. This discrepancy can be explained by the fact that this particular manipulation was the last one in the sequence done by Sharp et at. McNaughton et al. (1994) and Knierim et al. (1995) have shown that if rats experience the cue card moving over a number of sessions, they eventually come to ignore it and it loses control over place fields. When we tested our model without a cue card (equivalent to a card being present but ignored), the resulting place field was more diffuse than normal but showed no rotation; see Figure 3g. We thus predict that if this experiment had been done before the other manipulations rather than after, the place field would have foIlowed the cue card. In the Sharp et al. experiment, the animals were always placed in the environment at the same location during training. Therefore, they could reliably estimate their initial path integrator coordinates. They also had a reliable head direction estimate because they were not disoriented. We predict that were the rats trained with a variety of entry points instead of just one, using an environment with a single cue card at 00 (the training environment used by Sharp et al.), and then tested with two cue cards at 00 and 1800 , the place field would not rotate no matter what entry point was used. This is because when trained with a variable entry point, the animal would not learn to anticipate its path integrator coordinates upon entry; a path integratorreset would have to be done every time in order to establish the animal's coordinates. The reset mechanism uses allocentric bearing information derived from the head direction estimate, and in this task the resulting path integrator coordinates will be consistent with the initial head direction estimate. Hence, outcome 3 will always prevail. If the animal is disoriented, however, then both the path integrator and the head direction system must be reset upon entry (because consistency will be low with a faulty head direction), and the animal must choose one cue card or the other to match against its memory. So with disorientation and a variable entry point, the place field will be controlled by one or the other cue card with a 50/50 probability. This was found to be true in a related behavioral experiment by Cheng (1986). Our model shows how interactions between the place and head direction systems handle the various combinations of entry point, number of cue cards, and amount of cue card rotation. It predicts that head direction reset will be observed in certain tasks and not in others. In 66 A. D. REDISH, D. S. TOURETZKY experiments such as the single cue card task with an entry in the southeast, it predicts the place code will shift from an initial value corresponding to the northwest entry point to the value for the southeast entry point, but the head direction will not change. This could be tested by recording simultaneously from place cells and head direction cells. References Cheng, K. (1986). A purely geometric module in the rat's spatial representation. Cognition, 23: 149-178. Collett, T., Cartwright, B. A., and Smith, B. A. (1986). Landmark learning and visuospatial memories in gerbils. Journal of Comparative Physiology A, 158:835-851. Knierim, J. J., Kudrimoti, H. 5., and McNaughton, B. L. (1995). Place cells, head direction cells, and the learning of landmark stability. Journal of Neuroscience, 15: 164859. Maurer, R. and Seguinot, V. (1995). What is modelling for? A critical review of the models of path integration. Journal of Theoretical Biology, 175:457-475. McNaughton, B. L., Mizumori, S. J. Y., Barnes, C. A., Leonard, B. 1., Marquis, M., and Green, E. J. (1994). Cortical rpresentation of motion during unrestrained spatial navigation in the rat. Cerebral Cortex, 4(1):27-39. Muller, R. U., Kubie, 1. L., Bostock, E. M., Taube, J. 5., and Quirk, G. 1. (1991). Spatial firing correlates of neurons in the hippocampal formation of freely moving rats. In Paillard, J., editor, Brain and Space, chapter 17, pages 296-333. Oxford University Press, New York. Redish, A. D. and Touretzky, D. s. (1996). Navigating with landmarks: Computing goal locations from place codes. In Ikeuchi, K. and Veloso, M., editors, Symbolic Visual Learning. Oxford University Press. In press. Sharp, P. E., Kubie, J. L., and Muller, R. U. (1990). Firing properties of hippocampal neurons in a visually symmetrical environment: Contributions of multiple sensory cues and mnemonic processes. Journal of Neuroscience, 10(9):3093-3105. Taube, 1. s. (1995). Head direction cells recorded in the anterior thalamic nuclei of freely moving rats. Journal of Neuroscience, 15(1): 1953-1971. Taube, J. 5., Muller, R. I., and Ranck, Jr., J. B. (1990). Head direction cells recorded from the postsubiculum in freely moving rats. I. Description and quantitative analysis. Journal of Neuroscience, 10:420-435. Wan, H. 5., Touretzky, D. 5., and Redish, A. D. (1994a). Computing goal locations from place codes. In Proceedings of the 16th annual conference of the Cognitive Science society, pages 922-927. Lawrence Earlbaum Associates, Hillsdale N1. Wan, H. 5., Touretzky, D. 5., and Redish, A. D. (1994b). Towards a computational theory of rat navigation. In Mozer, M., Smolen sky, P., Touretzky, D., Elman, J., and Weigend, A., editors, Proceedings of the 1993 Connectionist Models Summer School, pages 11-19. Lawrence Earlbaum Associates, Hillsdale NJ. Modeling Interactions of the Rat's Place and Head Direction Systems (a) 1 cue card at 0° (East) entry in Northwest comer angle of rotation (Sharp et al.) = 2.7° precession of HD system = 0 0 (c) 2 cue cards at 00 (East) & 1800 (West) entry in Northwest comer angle of rotation (Sharp et al.) = -2.3° precession of HD system = 0 0 (e) 2 cue cards at 330 0 & 150 0 entry in Northwest comer not done by Sharp et al. precession of HD system = 331 0 (g) I cue card at 1800 (West) entry in Northwest comer angle of rotation (Sharp et al.) ::: -5.5 0 precession of HD system = 00 (b) 1 cue card at 00 entry in Southeast comer angle of rotation (Sharp et al.) = -6.0 0 precession of HD system = 2° (d) 2 cue cards at 00 & 180 0 entry in Southeast comer angle of rotation (Sharp et al.) = 182.5 0 precession of HD system::: 178 0 (f) 2 cue cards at 3300 & 150 0 entry in Southeast comer angle of rotation (Sharp et al.) = 158.3° precession of HD system = 151 ° (h) 1 cue card at 180 0 entry in Southeast comer angle of rotation (Sharp et al.) = 182.2° precession of HD system = 179 0 67 Figure 3: Computer simulations of the Sharp et al. (1990) experiment showing that place fields are controlled by both cue cards (thick arcs) and entry point (arrowhead). "Angle of rotation" is the angle at which the correlation between the probe and training case place fields is maximal. Because head direction and place code are tightly coupled in our model, precession of HD is an equivalent measure in our model.
|
1995
|
56
|
1,102
|
Empirical Entropy Manipulation for Real-World Problems Paul Viola: Nicol N. Schraudolph, Terrence J. Sejnowski Computational Neurobiology Laboratory The Salk Institute for Biological Studies 10010 North Torrey Pines Road La Jolla, CA 92037-1099 viola@salk.edu Abstract No finite sample is sufficient to determine the density, and therefore the entropy, of a signal directly. Some assumption about either the functional form of the density or about its smoothness is necessary. Both amount to a prior over the space of possible density functions. By far the most common approach is to assume that the density has a parametric form. By contrast we derive a differential learning rule called EMMA that optimizes entropy by way of kernel density estimation. Entropy and its derivative can then be calculated by sampling from this density estimate. The resulting parameter update rule is surprisingly simple and efficient. We will show how EMMA can be used to detect and correct corruption in magnetic resonance images (MRI). This application is beyond the scope of existing parametric entropy models. 1 Introduction Information theory is playing an increasing role in unsupervised learning and visual processing. For example, Linsker has used the concept of information maximization to produce theories of development in the visual cortex (Linsker, 1988). Becker and Hinton have used information theory to motivate algorithms for visual processing (Becker and Hinton, 1992). Bell and Sejnowski have used information maximization • Author to whom correspondence should be addressed. Current address: M.LT., 545 Technology Square, Cambridge, MA 02139. 852 P. VIOLA, N. N. SCHRAUDOLPH, T. J. SEJNOWSKI to solve the "cocktail party" or signal separation problem (Bell and Sejnowski, 1995). In order to simplify analysis and implementation, each of these techniques makes specific assumptions about the nature of the signals used, typically that the signals are drawn from some parametric density. In practice, such assumptions are very inflexible. In this paper we will derive a procedure that can effectively estimate and manipulate the entropy of a wide variety of signals using non-parametric densities. Our technique is distinguished by is simplicity, flexibility and efficiency. We will begin with a discussion of principal components analysis (PCA) as an example of a simple parametric entropy manipulation technique. After pointing out some of PCA's limitation, we will then derive a more powerful non-parametric entropy manipulation procedure. Finally, we will show that the same entropy estimation procedure can be used to tackle a difficult visual processing problem. 1.1 Parametric Entropy Estimation Typically parametric entropy estimation is a two step process. We are given a parametric model for the density of a signal and a sample. First, from the space of possible density functions the most probable is selected. This often requires a search through parameter space. Second, the entropy of the most likely density function is evaluated. Parametric techniques can work well when the assumed form of the density matches the actual data. Conversely, when the parametric assumption is violated the resulting algorithms are incorrect. The most common assumption, that the data follow the Gaussian density, is especially restrictive. An entropy maximization technique that assumes that data is Gaussian, but operates on data drawn from a non-Gaussian density, may in fact end up minimizing entropy. 1.2 Example: Principal Components Analysis There are a number of signal processing and learning problems that can be formulated as entropy maximization problems. One prominent example is principal component analYllill (PCA). Given a random variable X, a vector v can be used to define a new random variable, Y" = X . v with variance Var(Y,,) = E[(X . v - E[X . v])2]. The principal component v is the unit vector for which Var(Yv) is maximized. In practice neither the density of X nor Y" is known. The projection variance is computed from a finite sample, A, of points from X, Var(Y,,) ~ Var(Y,,) == EA[(X . v - EA[X . v])2] , (1) A where VarA(Y,,) and E A [·] are shorthand for the empirical variance and mean evaluated over A. Oja has derived an elegant on-line rule for learning v when presented with a sample of X (Oja, 1982). Under the assumption that X is Gaussian is is easily proven that Yv has maximum entropy. Moreover, in the absence of noise, Yij, contains maximal information about X. However, when X is not Gaussian Yij is generally not the most informative projection. 2 Estimating Entropy with Parzen Densities We will now derive a general procedure for manipulating and estimating the entropy of a random variable from a sample. Given a sample of a random variable X, we can Empirical Entropy Manipulation for Real-world Problems 853 construct another random variable Y = F(X,l1). The entropy, heY), is a function of v and can be manipulated by changing 11. The following derivation assumes that Y is a vector random variable. The joint entropy of a two random variables, h(Wl' W2), can be evaluated by constructing the vector random variable, Y = [Wl' w2jT and evaluating heY). Rather than assume that the density has a parametric form, whose parameters are selected using maximum likelihood estimation, we will instead use Parzen window density estimation (Duda and Hart, 1973). In the context of entropy estimation, the Parzen density estimate has three significant advantages over maximum likelihood parametric density estimates: (1) it can model the density of any signal provided the density function is smooth; (2) since the Parzen estimate is computed directly from the sample, there is no search for parameters; (3) the derivative of the entropy of the Parzen estimate is simple to compute. The form of the Parzen estimate constructed from a sample A is p.(y, A) = ~A I: R(y - YA) = EA[R(y - YA)] , YAEA (2) where the Parzen estimator is constructed with the window function R(·) which integrates to 1. We will assume that the Parzen window function is a Gaussian density function. This will simplify some analysis, but it is not necessary. Any differentiable function could be used. Another good choice is the Cauchy density. Unfortunately evaluating the entropy integral hey) ~ -E[log p.(~, A)] = -i: log p.(y, A)dy is inordinately difficult. This integral can however be approximated as a sample mean: (3) where EB{ ] is the sample mean taken over the sample B. The sample mean converges toward the true expectation at a rate proportional to 1/ v' N B (N B is the size of B). To reiterate, two samples can be used to estimate the entropy of a density: the first is used to estimate the density, the second is used to estimate the entropyl. We call h· (Y) the EMMA estimate of entropy2. One way to extremize entropy is to use the derivative of entropy with respect to v. This may be expressed as ~h(Y) ~ ~h·(Y) = __ 1_ '" LYAEA f;gt/J(YB - YA) (4) dl1 dv N B L....iB Ly EA gt/J(YB - YA) YBE A 1 d 1 = NB I: I: Wy (YB , YA) dl1 "2 Dt/J(YB - YA), (5) YBEB YAEA _ gt/J(Yl - Y2) where WY(Yl' Y2) = L ( ) , (6) YAEA gt/J Yl - YA Dt/J(Y) == yT.,p-ly, and gt/J(Y) is a multi-dimensional Gaussian with covariance .,p. Wy(Yl' Y2) is an indicator of the degree of match between its arguments, in a "soft" lUsing a procedure akin to leave-one-out cross-validation a single sample can be used for both purposes. 2EMMA is a random but pronounceable subset of the letters in the words "Empirical entropy Manipulation and Analysis". 854 P. VIOLA, N. N. SCHRAUDOLPH, T. J. SEJNOWSKl sense. It will approach one if Yl is significantly closer to Y2 than any element of A. To reduce entropy the parameters v are adjusted such that there is a reduction in the average squared distance between points which Wy indicates are nearby. 2.1 Stochastic Maximization Algorithm Both the calculation of the EMMA entropy estimate and its derivative involve a double summation. As a result the cost of evaluation is quadratic in sample size: O(NANB). While an accurate estimate of empirical entropy could be obtained by using all of the available data (at great cost), a stochastic estimate of the entropy can be obtained by using a random subset of the available data (at quadratically lower cost). This is especially critical in entropy manipulation problems, where the derivative of entropy is evaluated many hundreds or thousands of times. Without the quadratic savings that arise from using smaller samples entropy manipulation would be impossible (see (Viola, 1995) for a discussion of these issues). 2.2 Estimating the Covariance In addition to the learning rate .A, the covariance matrices of the Parzen window functions, g,p, are important parameters of EMMA. These parameters may be chosen so that they are optimal in the maximum likelihood sense. For simplicity, we assume that the covariance matrices are diagonal,.,p = DIAG(O"~,O"~, ... ). Following a derivation almost identical to the one described in Section 2 we can derive an equation analogous to (4), d. 1"" "" ( 1 ) ([y]~ ) -h (Y) = L...J L...J WY(YB' YA) -- - 1 dO"k N B b O"k O"~ YsE YAEa (7) where [Y]k is the kth component of the vector y. The optimal, or most likely, .,p minimizes h· (Y). In practice both v and .,p are adjusted simultaneously; for example, while v is adjusted to maximize h· (YlI ), .,p is adjusted to minimize h· (y,,). 3 Principal Components Analysis and Information As a demonstration, we can derive a parameter estimation rule akin to principal components analysis that truly maximizes information. This new EMMA based component analysis (ECA) manipulates the entropy of the random variable Y" = X·v under the constraint that Ivl = 1. For any given value of v the entropy of Yv can be estimated from two samples of X as: h·(Yv ) = -EB[logEA[g,p(xB·v - XA· v)]], where .,p is the variance of the Parzen window function. Moreover we can estimate the derivative of entropy: d~ h·(YlI ) = ; L L Wy(YB, YA) .,p-l(YB - YA)(XB - XA) , B B A where YA = XA . v and YB = XB . v. The derivative may be decomposed into parts which can be understood more easily. Ignoring the weighting function Wy.,p-l we are left with the derivative of some unknown function f(y"): d 1 dvf(Yv ) = N N L L(YB - YA)(XB - XA) (8) B A B A What then is f(y")? The derivative of the squared difference between samples is: d~ (YB - YA)2 = 2(YB - YA)(XB - XA) . So we can see that f(Y,,) = 2N IN L L(YB - YA)2 B A B A Empirical Entropy Manipulation for Real-world Problems 3 2 o -I -2 •• t -3 -4 -2 • I o : . ECA-MIN ECA-MAX BCM BINGO PCA 2 4 Figure 1: See text for description. 855 is one half the expectation of the squared difference between pairs of trials of Yv • Recall that PCA searches for the projection, Yv , that has the largest sample variance. Interestingly, f(Yv ) is precisely the sample variance. Without the weighting term Wll,p-l, ECA would find exactly the same vector that PCA does: the maximum variance projection vector. However because of Wll , the derivative of ECA does not act on all points of A and B equally. Pairs of points that are far apart are forced no further apart. Another way of interpreting ECA is as a type of robust variance maximization. Points that might best be interpreted as outliers, because they are very far from the body of other points, playa very small role in the minimization. This robust nature stands in contrast to PCA which is very sensitive to outliers. For densities that are Gaussian, the maximum entropy projection is the first principal component. In simulations ECA effectively finds the same projection as PCA, and it does so with speeds that are comparable to Oja's rule. ECA can be used both to find the entropy maximizing (ECA-MAX) and minimizing (ECA-MIN) axes. For more complex densities the PCA axis is very different from the entropy maximizing axis. To provide some intuition regarding the behavior of ECA we have run ECAMAX, ECA-MIN, Oja's rule, and two related procedures, BCM and BINGO, on the same density. BCM is a learning rule that was originally proposed to explain development of receptive fields patterns in visual cortex (Bienenstock, Cooper and Munro, 1982). More recently it has been argued that the rule finds projections that are far from Gaussian (Intrator and Cooper, 1992). Under a limited set of conditions this is equivalent to finding the minimum entropy projection. BINGO was proposed to find axes along which there is a bimodal distribution (Schraudolph and Sejnowski, 1993). Figure 1 displays a 400 point sample and the projection axes discussed above. The density is a mixture of two clusters. Each cluster has high kurtosis in the horizontal direction. The oblique axis projects the data so that it is most uniform and hence has the highest entropy; ECA-MAX finds this axis. Along the vertical axis the data is clustered and has low entropy; ECA-MIN finds this axis. The vertical axis also has the highest variance. Contrary to published accounts, the first principal component can in fact correspond to the minimum entropy projection. BCM, while it may find minimum entropy projections for some densities, is attracted to the kurtosis along the horizontal axis. For this distribution BCM neither minimizes nor maximizes entropy. Finally, BINGO successfully discovers that the vertical axis is very bimodal. 856 P. VIOLA, N. N. SCHRAUOOLPH, T. J. SEJNOWSKI 1200 1000 800 600 400 200 ~ .1 0 0.1 0.2 0.3 0.4 \ Corrupted:. Corrected .• . ' .. . : '. 0.7 0.8 0.9 Figure 2: At left: A slice from an MRI scan of a head. Center: The scan after correction. Right: The density of pixel values in the MRI scan before and after correction. 4 Applications EMMA has proven useful in a number of applications. In object recognition EMMA has been used align 3D shape models with video images (Viola and Wells III, 1995). In the area of medical imaging EMMA has been used to register data that arises from differing medical modalities such as magnetic resonance images, computed tomography images, and positron emission tomography (Wells, Viola and Kikinis, 1995). 4.1 MRI Processing In addition, EMMA can be used to process magnetic resonance images (MRI). An MRI is a 2 or 3 dimensional image that records the density of tissues inside the body. In the head, as in other parts of the body, there are a number of distinct tissue classes including: bone, water, white matter, grey matter, and fat. ~n principle the density of pixel values in an MRI should be clustered, with one cluster for each tissue class. In reality MRI signals are corrupted by a bias field, a multiplicative offset that varies slowly in space. The bias field results from unavoidable variations in magnetic field (see (Wells III et al., 1994) for an overview of this problem). Because the densities of each tissue type cluster together tightly, an uncorrupted MRI should have relatively low entropy. Corruption from the bias field perturbs the MRI image, increasing the values of some pixels and decreasing others. The bias field acts like noise, adding entropy to the pixel density. We use EMMA to find a low-frequency correction field that when applied to the image, makes the pixel density have a lower entropy. The resulting corrected image will have a tighter clustering than the original density. Call the uncorrupted scan s(z); it is a function of a spatial random variable z. The corrupted scan, c( x) = s( z) + b( z) is a sum of the true scan and the bias field. There are physical reasons to believe b( x) is a low order polynomial in the components of z. EMMA is used to minimize the entropy of the corrected signal, h( c( x) - b( z, v», where b( z, v), a third order polynomial with coefficients v, is an estimate for the bias corruption. Figure 2 shows an MRI scan and a histogram of pixel intensity before and after correction. The difference between the two scans is quite subtle: the uncorrected scan is brighter at top right and dimmer at bottom left. This non-homogeneity Empirical Entropy Manipulation for Real-world Problems 857 makes constructing automatic tissue classifiers difficult. In the histogram of the original scan white and grey matter tissue classes are confounded into a single peak ranging from about 0.4 to 0.6. The histogram of the corrected scan shows much better separation between these two classes. For images like this the correction field takes between 20 and 200 seconds to compute on a Sparc 10. 5 Conclusion We have demonstrated a novel entropy manipulation technique working on problems of significant complexity and practical importance. Because it is based on nonparametric density estimation it is quite flexible, requiring no strong assumptions about the nature of signals. The technique is widely applicable to problems in signal processing, vision and unsupervised learning. The resulting algorithms are computationally efficient. Acknowledgements This research was support by the Howard Hughes Medical Institute. References Becker, S. and Hinton, G. E. (1992). A self-organizing neural network that discovers surfaces in random-dot stereograms. Nature, 355:161-163. Bell, A. J. and Sejnowski, T. J. (1995). An information-maximisation approach to blind separation. In Tesauro, G., Touretzky, D. S., and Leen, T. K., editors, Advance8 in Neural Information Proce88ing, volume 7, Denver 1994. MIT Press, Cambridge. Bienenstock, E., Cooper, L., and Munro, P. (1982). Theory for the development of neuron selectivity: Orientation specificity and binocular interaction in visual cortex. Journal of Neur08cience, 2. Duda, R. and Hart, P. (1973). Pattern Cla88ification and Scene AnalY8i8. Wiley, New York. Intrator, N. and Cooper, L. N. (1992). Objective function formulation of the bcm theory of visual cortical plasticity: Statistical connections, stability conditions. Neural Network., 5:3-17. Linsker, R. (1988). Self-organization in a perceptual network. IEEE Computer, pages 105-117. Oja, E. (1982). A simplified neuron model as a principal component analyzer. Journal of Mathematical Biology, 15:267-273. Schraudolph, N. N. and Sejnowski, T. J. (1993). Unsupervised discrimination of clustered data via optimization of binary information gain. In Hanson, S. J., Cowan, J. D., and Giles, C. L., editors, Advance. in Neural Information Proce88ing, volume 5, pages 499-506, Denver 1992. Morgan Kaufmann, San Mateo. Viola, P. A. (1995). Alignment by Ma:cimization of Mutual Information. PhD thesis, Massachusetts Institute of Technology. MIT AI Laboratory TR 1548. Viola, P. A. and Wells III, W. M. (1995). Alignment by maximization of mutual information. In Fifth Inti. Conf. on Computer Vi8ion, pages 16-23, Cambridge, MA. IEEE. Wells, W., Viola, P., and Kikinis, R. (1995). Multi-modal volume registration by maximization of mutual information. In Proceeding. of the Second International Sympo8ium on Medical Robotic. and Computer A88i8ted Surgery, pages 55 - 62. Wiley. Wells III, W., Grimson, W., Kikinis, R., and Jolesz, F. (1994). Statistical Gain Correction and Segmentation of MRI Data. In Proceeding. of the Computer Society Conference on Computer Vi.ion and Pattern Recognition, Seattle, Wash. IEEE, Submitted.
|
1995
|
57
|
1,103
|
A Neural Network Model of 3-D Lightness Perception Luiz Pessoa Federal Univ. of Rio de Janeiro Rio de Janeiro, RJ, Brazil pessoa@cos.ufrj.br Abstract William D. Ross Boston University Boston, MA 02215 bill@cns.bu.edu A neural network model of 3-D lightness perception is presented which builds upon the FACADE Theory Boundary Contour System/Feature Contour System of Grossberg and colleagues. Early ratio encoding by retinal ganglion neurons as well as psychophysical results on constancy across different backgrounds (background constancy) are used to provide functional constraints to the theory and suggest a contrast negation hypothesis which states that ratio measures between coplanar regions are given more weight in the determination of lightness of the respective regions. Simulations of the model address data on lightness perception, including the coplanar ratio hypothesis, the Benary cross, and White's illusion. 1 INTRODUCTION Our everyday visual experience includes surface color constancy. That is, despite 1) variations in scene lighting and 2) movement or displacement across visual contexts, the color of an object appears to a large extent to be the same. Color constancy refers, then, to the fact that surface color remains largely constant despite changes in the intensity and composition of the light reflected to the eyes from both the object itself and from surrounding objects. This paper discusses a neural network model of 3D lightness perception i.e., only the achromatic or black to white dimension of surface color perception is addressed. More specifically, the problem of background constancy (see 2 above) is addressed and mechanisms to accomplish it in a system exhibiting illumination constancy (see 1 above) are proposed. A landmark result in the study of lightness was an experiment reported by Wallach (1948) who showed that for a disk-annulus pattern, lightness is given by the ratio of disk and annulus luminances (i.e., independent of overall illumination); the A Neural Network Model of 3-D Lightness Perception 845 so-called ratio principle. In another study, Whittle and Challands (1969) had subjects perform brightness matches in a haploscopic display paradigm. A striking result was that subjects always matched decrements to decrements, or increments to increments, but never increments to decrements. Whittle and Challands' (1969) results provide psychophysical support to the notion that the early visual system codes luminance ratios and not absolute luminance. These psychophysical results are in line with results from neurophysiology indicating that cells at early stages of the visual system encode local luminance contrast (Shapley and Enroth-Cugell, 1984). Note that lateral inhibition mechanisms are sensitive to local ratios and can be used as part of the explanation of illumination constancy. Despite the explanatory power of the ratio principle, and the fact that the early stages of the visual system likely code contrast, several experiments have shown that, in general, ratios are insufficient to account for surface color perception. Studies of background constancy (Whittle and Challands, 1969; Land and McCann, 1971; Arend and Spehar, 1993), of the role of 3-D spatial layout and illumination arrangement on lightness perception (e.g. , Gilchrist, 1977) as well as many other effects, argue against the sufficiency of local contrast measures (e.g., Benary cross, White's, 1979 illusion). The neural network model presented here addresses these data using several fields of neurally plausible mechanisms of lateral inhibition and excitation. 2 FROM LUMINANCE RATIOS TO LIGHTNESS The coplanar ratio hypothesis (Gilchrist, 1977) states that the lightness of a given region is determined predominantly in relation to other coplanar surfaces, and not by equally weighted relations to all retinally adjacent regions. We propose that in the determination of lightness, contrast measures between non-coplanar adjacent surfaces are partially negated in order to preserve background constancy. Consider the Benary Cross pattern (input stimulus in Fig. 2). If the gray patch on the cross is considered to be at the same depth as the cross, while the other gray patch is taken to be at the same depth as the background (which is below the cross), the gray patch on the cross should look lighter (since its lightness is determined in relation to the black cross), and the other patch darker (since its lightness is determined in relation to the white background) . White's (1979) illusion can be discussed in similar terms (see the input stimulus in Fig. 3). The mechanisms presented below implement a process of partial contrast negation in which the initial retinal contrast code is modulated by depth information such that the retinal contrast consistent with the depth interpretation is maintained while the retinal contrast not supported by depth is negated or attenuated. 3 A FILLING-IN MODEL OF 3-D LIGHTNESS Contrast/Filling-in models propose that initial measures of boundary contrast followed by spreading of neural activity within filling-in compartments produce a response profile isomorphic with the percept (Gerrits & Vendrik, 1970; Cohen & Grossberg, 1984; Grossberg & Todorovic, 1988; Pessoa, Mingolla, & Neumann, 1995). In this paper we develop a neural network model of lightness perception in the tradition of contrast/filling-in theories. The neural network developed here is an extension of the Boundary Contour System/Feature Contour System (BCS/FCS) proposed by Cohen and Grossberg (1984) and Grossberg and Mingolla (1985) to explain 3-D lightness data. 846 L. PESSOA. W. D. ROSS A fundamental idea of the BCS/FCS theory is that lateral inhibition achieves illumination constancy but requires the recovery of lightness by the filling-in, or diffusion, of featural quality ("lightness" in our case). The final diffused activities correspond to lightness, which is the outcome of interactions between boundaries and featural quality, whereby boundaries control the process of filling-in by forming gates of variable resistance to diffusion. H ow can the visual system construct 3-D lightness percepts from contrast measures obtained by retinotopic lateral inhibition? A mechanism that is easily instantiated in a neural model and provides a straightforward modification to the contrast/fillingin proposal of Grossberg and Todorovic (1988) is the use of depth-gated filling-in. This can be accomplished through a pathway that modulates boundary strength for boundaries between surfaces or objects across depth. The use of permeable or "leaky" boundaries was also used by Grossberg and Todorovic (1988) for 2-D stimuli. In the current usage, permeability is actively increased at depth boundaries to partially negate the contrast effect since filling-in proceeds more freely and thus preserve lightness constancy across backgrounds. Figure 1 describes the four computational stages of the system. I BOUNDARIES ,...---------, ~ ~ ON/OFF ~FILTERING j ~ I RLLlNG-IN I Figure 1: Model components. I '" I DEPTH I MAP Stage 1: Contrast Measurement. At this stage both ON and OFF neural fields with lateral inhibitory connectivity measure the strength of contrast at image regions in uniform regions a contrast measurement of zero results. Formally, the ON field is given by dyi; _ + + ) + (+ ) + dt - -aYij + ((3 Yij Cij Yij + 'Y Eij (1) where a , (3 and 'Yare constants; ct is the total excitatory input to yi; and Et; is the total inhibitory input to yi;. These terms denote discrete convolutions of the input Iij with Gaussian weighting functions, or kernels. An analogous equation specifies Yi; for the OFF field. Figure 2 shows the ON-contrast minus the OFF-contrast. Stage 2: 2-D Boundary Detection. At Stage 2, oriented odd-symmetric boundary detection cells are excited by the oriented sampling of the ON and OFF Stage 1 cells. Responses are maximal when ON activation is strong on one side of a cell's receptive field and OFF activation is strong on the opposite side. In other words, the cells are tuned to ON/OFF contrast co-occurrence, or juxtaposition (see Pessoa et aI., 1995). The output at this stage is the sum of the activations of such cells at each location for all orientations. The output responses are sharpened and localized through lateral inhibition across space; an equation similar to Equation 1 is used. The final output of Stage 2 is given by the signals Zij (see Fig. 2, Boundaries). Stage 3: Depth Map. In the current implementation a simple scheme was employed for the determination of the depth configuration. Initially, four types of A Neural Network Model of 3-D Lightness Perception 847 T-junction cells detect such configurations in the image. For example, Iij = Zi-d,j x Zi+d,j x Zi ,j+d, (2) where d is a constant, detects T-junctions, where left, right, and top positions of the boundary stage are active; similar cells detect T-junctions of different orientations. The activities of the T-junction cells are then used in conjunction with boundary signals to define complete boundaries. Filling-in within these depth boundaries results in a depth map (see Fig. 2, Depth Map). Stage 4: Depth-modulated Filling-in. In Stage 4, the ON and OFF contrast measures are allowed to diffuse across space within respective filling-in regions. Diffusion is blocked by boundary activations from Stage 2 (see Grossberg & Todorovic, 1988, for details). The diffusion process is further modulated by depth information. The depth map provides this information; different activities code different depths. In a full blown implementation of the model, depth information would be obtained by the depth segmentation of the image supported by both binocular disparity and monocular depth cues. Depth-modulated filling-in is such that boundaries across depths are reduced in strength. This allows a small percentage of the contrast on either side ofthe boundary to leak across it resulting in partial contrast negation, or reduction, at these boundaries. ON and OFF filling-in domains are used which receive the corresponding ON and OFF contrast activities from Stage 1 as inputs (see Fig. 2, Filled-in). 4 SIMULATIONS The present model can account for several important phenomena, including 2 - D effects of lightness constancy and contrast (see Grossberg and Todorovic, 1988). The simulations that follow address 3 -D lightness effects. 4.1 Benary Cross Figure 2 shows the simulation for the Benary Cross. The plotted gray level values for filling-in reflect the activities of the ON filling-in domain minus the OFF domain. The model correctly predicts that the patch on the cross appears lighter than the patch on the background. This result is a direct consequence of contrast negation. The depth relationships are such that the patch on the cross is at the same depth as the cross and the patch on the background is at the same depth as the background (see Fig. 2, Depth Map) . Therefore, the ratio of the background to the patch on the cross (across a depth boundary) and the ratio of the cross to the patch on the background (also across a depth boundary), are given a smaller weight in the lightness computation. Thus, the background will have a stronger effect on the appearance of the patch on the background, which will appear darker. At the same time, the cross will have a greater effect on the appearance of the patch on the cross, which will appear lighter. 4.2 White's lllusion White's (1979) illusion (Fig. 3) is such that the gray patches on the black stripes appear lighter than the gray patches on the white stripes. This effect is considered a puzzling violation of simultaneous contrast since the contour length of the gray patches is larger for the stripes they do not lie on. Simultaneous contrast would predict that the gray patches on the black stripes appear lighter than the ones on white. 848 Stimulus I I I L~ - ---I Boundaries ON-OFF Contrast L. PESSOA, W. D. ROSS Depth Map Filled-in Figure 2: Benary Cross. The filled-in values of the gray patch on the cross are higher than the ones for the gray patch on the background. Gray levels code intensity; darker grays code lower values, lighter grays code higher values. Figure 3 shows the result of the model for White's effect. The T-junction information in the stimulus determines that the gray patches are coplanar with the patches they lie on. Therefore, their appearance will be determined in relation to the contrast of their respective backgrounds. This is obtained, again, through contrast modulation, where the contrast of, say, the gray patch on a black stripe is preserved, while the contrast of the same patch with the white is partially negated (due to the depth arrangement). 4.3 Coplanar Hypothesis Gilchrist (1977) showed that the perception of lightness is not determined by retinal adjacency, and that depth configuration and spatial layout help specify lightness. More specifically, it was proposed that the ratio of coplanar surfaces, not necessarily retinally adjacent, determines lightness, the so-called coplanar ratio hypothesis. Gilchrist was able to convincingly demonstrate this by comparing the perception of lightness in two equivalent displays (in terms of luminance values), aside from the perceived depth relationships in the displays. Figure 4 shows computer simulations of the coplanar ratio effect. The same stimulus is given as input in two simulations with different depth specifications. In one (Depth Map 1), the depth map specifies that the rightmost patch is at a different depth than the two leftmost patches which are coplanar. In the other (Depth Map 2), the two rightmost patches are coplanar and at a different depth than the leftmost patch. In all, the depth organization alters the lightness of the central region, which should appear darker in the configuration of Depth Map 1 than the one for Depth Map 2. For Depth Map 1, since the middle patch is coplanar with a white patch, this patch is darkened by simultaneous contrast. For Depth Map 2, the middle patch will be lightened by contrast since it is coplanar with a black patch. It should be noted that the depth maps for the simulations shown in Fig. 4 were given as input. A Neural Network Model of 3-D Lightness Perception 849 -, - --- 1 1 Boundaries Stimulus ON-OFF Contrast Filled-in Figure 3: White's effect. The filled-in values of the gray patches on the black stripes are higher than the ones for the gray patches on white stripes. The current implementation cannot recover depth trough binocular disparity and only employs monocular cues as in the previous simulations. 5 CONCLUSIONS In this paper, data from experiments on lightness perception were used to extend the BCSjFCS theory of Grossberg and colleagues to account for several challenging phenomena. The model is an initial step towards providing an account that can take into consideration the complex factors involved in 3-D vision see Grossberg (1994) for a comprehensive account of 3-D vision. Acknowledgements The authors would like to than Alan Gilchrist and Fred Bonato for their suggestions concerning this work. L. P. was supported in part by Air Force Office of Scientific Research (AFOSR F49620-92-J-0334) and Office of Naval Research (ONR N0001491-J-4100); W. R. was supported in part by HNC SC-94-001. Reference Arend, L., & Spehar, B. (1993) Lightness, brightness, and brightness contrast: 2. Reflectance variation. Perception {3 Psychophysics 54:4576-468. Cohen, M., & Grossberg, S. (1984) Neural dynamics of brightness perception: Features, boundaries, diffusion, and resonance. Perception {3 Psychophysics 36:428-456. Gerrits, H. & Vendrik, A. (1970) Simultaneous contrast, filling-in process and information processing in man's visual system. Experimental Brain Research 11:411-430. 850 L. PESSOA, W. D. ROSS Filled-in 2 Stimulus Depth Map 1 Filled-in 1 Figure 4: Gilchrist's coplanarity. The Filled-in values for the middle patch on top are higher than on bottom. Gilchrist, A. (1977) Perceived lightness depends on perceived spatial arrangement. Science 195:185-187. Grossberg, S. (1994) 3-D vision and figure-ground separation by visual cortex. Perception & Psychophysics 55:48-120. Grossberg, S., & Mingolla, E. (1985) Neural dynamics of form perception: Boundary completion, illusory figures, and neon color spreading. Psychological Review 92:173-211. Grossberg, S., & Todorovic. D. (1988). Neural dynamics of 1-D and 2-D brightness perception: A unified model of classical and recent phenomena. Perception & Psychophysics 43:241-277. Land, E., & McCann, J. (1971). Lightness and retinex theory. Journal of the Optical Society of America 61:1-11. Pessoa, L., Mingolla, E., & Neumann, H. (1995) A contrast- and luminance-driven multiscale network model of brightness perception. Vision Research 35:22012223. Shapley, R., & Enroth-Cugell, C. (1984) Visual adaptation and retinal gain controls. In N. Osborne and G. Chader (eds.), Progress in Retinal Research, pp. 263346. Oxford: Pergamon Press. Wallach, H. (1948) Brightness constancy and the nature of achromatic colors. Journal of Experimental Psychology 38: 310-324. White, M. (1979) A new effect of pattern on perceived lightness. Perception 8:413416. Whittle, P., & Challands, P. (1969) The effect of background luminance on the brightness of flashes. Vision Research 9:1095-1110.
|
1995
|
58
|
1,104
|
Adaptive Retina with Center-Surround Receptive Field Shih-Chii Lin and Kwabena Boahen Computation and Neural Systems 139-74 California Institute of Technology Pasadena, CA 91125 shih@pcmp.caltech.edu, buster@pcmp.caltech.edu Abstract Both vertebrate and invertebrate retinas are highly efficient in extracting contrast independent of the background intensity over five or more decades. This efficiency has been rendered possible by the adaptation of the DC operating point to the background intensity while maintaining high gain transient responses. The centersurround properties of the retina allows the system to extract information at the edges in the image. This silicon retina models the adaptation properties of the receptors and the antagonistic centersurround properties of the laminar cells of the invertebrate retina and the outer-plexiform layer of the vertebrate retina. We also illustrate the spatio-temporal responses of the silicon retina on moving bars. The chip has 59x64 pixels on a 6.9x6.8mm2 die and it is fabricated in 2 J-tm n-well technology. 1 Introduction It has been observed previously that the initial layers of the vertebrate and invertebrate retina systems perform very similar processing functions on the incoming input signal[1]. The response versus log intensity curves of the receptors in invertebrate and vertebrate retinas look similar. The curves show that the receptors have a larger gain for changes in illumination than to steady illumination, i.e, the receptors adapt. This adaptation property allows the receptor to respond over a large input range without saturating. Anatomically, the eyes of invertebrates differ greatly from that of vertebrates. VerAdaptive Retina with Center-Surround Receptive Field 679 tebrates normally have two simple eyes while insects have compound eyes. Each compound eye in the fly consists of 3000-4000 ommatidia and each ommatidium consists of 8 photoreceptors. Six of these receptors (which are also called RI-R6) are in a single spectral class. The other two receptors, R7 and R8 provide channels for wavelength discrimination and polarization. The vertebrate eye is divided into the outer-plexiform layer and the inner-plexiform layer. The outer-plexiform layer consists of the rods and cones, horizontal cells and bipolar cells. Invertebrate receptors depolarise in response to an increase in light, in contrast to vertebrate receptors, which hyperpolarise to an increase in light intensity. Both vertebrate and invertebrate receptors show light adaptation over at least five decades of background illumination. This adaptation property allows the retina to maintain a high transient gain to contrast over a wide range of background intensities. The invertebrate receptors project to the next layer which is called the lamina layer. This layer consists primarily of monopolar cells which show a similar response versus log intensity curve to that of vertebrate bipolar cells in the outer-plexiform layer. Both cells respond with graded potentials to changes in illumination. These cells also show a high transient gain to changes in illumination while ignoring the background intensity and they possess center-surround receptive fields. In vertebrates, the cones which are excited by the incoming light, activate the horizontal cells which in tum inhibit the cones. The horizontal cells thus mediate the lateral inhibition which produces the center-surround properties. In insects, a possible process of this lateral inhibition is done by current flow from the photoreceptors through the epithelial glial cells surrounding an ommatidium or the modulation of the local field potential in the lamina to influence the transmembrane potential of the photoreceptor[2]. The center-surround receptive fields allow contrasts to be accentuated since the surround computes a local mean and subtracts that from the center signal. Mahowald[3] previously described a silicon retina with adaptive photoreceptors and Boahen et al.[4] recently described a compact current-mode analog model of the outer-plexiform layer of the vertebrate retina and analysed the spatio-temporal processing properties of this retina[5]. A recent array of photoreceptors from Delbriick[6] uses an adaptive photoreceptor circuit that adapts its operating point to the background intensity so that the pixel shows a high transient gain over 5 decades of background illumination. However this retina does not have spatial coupling between pixels. The pixels in the silicon retina described here has a compact circuit that incorporates both spatial and temporal filtering with light adaptation over 5 decades of background intensity. The network exhibits center-surround behavior. Boahen et al.[4] in their current-mode diffusor retina, draw an analogy between parts of the diffusor circuit and the different cells in the outer-plexiform layer. While the same analogy cannot be drawn from this silicon retina to the invertebrate retina since the function of the cells are not completely understood, the output responses of the retina circuit are similar to the output responses of the photoreceptor and monopolar cells in invertebrates. The circuit details are described in Section 2 and the spatio-temporal processing performed by the retina on stimulus moving at different speeds is shown in Section 680 S.-C. LIU, K. BOAHEN 3. 2 Circuit -----VI Vb VI VI 1 p1 1 M4 Vh VI·I Vh VI+I .1. .1. .bel --------Vr MI (a) im.l iia iI ... I rrr rrr rrr 'II 'II (b) Figure 1: (a) One-dimensional version of the retina. (b) Small-signal equivalent of circuit in (a). A one-dimensional version of the retina is shown in Figure l(a). The retina consists of an adaptive photoreceptor circuit at each pixel coupled together with diffusors, controlled by voltages, Vg and Vh. The output of this network can either be obtained at the voltage output, V, or at the current output, 10 but the outputs have different properties. Phototransduction is obtained by using a reverse-biased photodiode which produces current that is proportional to the incident light. The logarithmic properties are obtained by operating the feedback transistor shown in Figure l(a) in the subthreshold region. The voltage change at the output photoreceptor, Vr , is proportional to a small contrast since UT UTdI UT i Vr = -d(logl) = -- = -K, K, 1 K, h g where UT is the thermal voltage, K, = CO:rCd ' Coz is the oxide capacitance and Cd is the depletion capacitance of a transistor. The circuit works as follows: If the photocurrent through the photodiode increases, Vr will be pulled low and the output voltage at V, increases by VI = AVr where A is the amplifier gain of the output stage. This output change in V, is coupled into Vel through a capacitor Adaptive Retina with Center-Surround Receptive Field 681 divider ratio, Cl~2C2. The feedback transistor, M4, operates in the subthreshold region and supplies the current necessary to offset the photocurrent. The increase in Vel (i.e. the gate voltage of M4) causes the current supplied by M3 to increase which pulls the node voltage, Vr , back to the voltage level needed by Ml to sink the bias current from transistor, M2. 3.5 3.45 3.4 ... -= 0 ~ • 3.35 • -2 c 0 Q. • • a: 3.3 -1 3.25 0 3.2 0 5 10 15 20 25 Time (Sec) Figure 2: This figure shows the output response of the receptor to a variation of about 40% p-p in the intensity of a flickering LED light incident on the chip. The response shows that the high sensitivity of the receptor to the LED is maintained over 5 decades of differing background intensities. The numbers on the section of the curve indicate the log intensity of the mean value. 0 log is the absolute intensity from the LED. The adaptive element, M3, has an I-V curve which looks like a hyperbolic sine. The small slope of the I-V curve in the middle means that for small changes of voltages across M3, the element looks like an open-circuit. With large changes of voltage across M3, the current through M3 becomes exponential and Vel is charged or discharged almost instantaneously. Figure 2 shows the output response of the photoreceptor to a square-wave variation of about 40% p-p in the intensity of a red LED (635 nm). The results show that the circuit is able to discern the small contrast over five decades of background intensity while the steady-state voltage of the photoreceptor output varies only about 15mV. Further details of the photoreceptor circuit and its adaptation properties are described in Delbriick[6]. 3 Spatio-Temporal Response The spatio-temporal response of the network to different moving stimuli is explored in this section. The circuit shown in Figure l(a) can be transferred to an equivalent network of resistors and capacitors as shown in Figure l(b) to obtain the transfer function of the circuit. The capacitors at each node are necessary to model the 682 8.5 i ~ 7.5 :; ... :; o i I ~ r. ... (a) 0.4 0.6 0.8 Time (Sec) 1.2 1.4 3.8 ~_---:-":--_--::'= __ ':"':-_--::':-_--::,'::-_--.J 0.3 0.4 0.5 0.6 0.7 0.8 (b) Time (Sec) S.-C. LIU, K. BOAHEN 1 lJ; Figure 3: (a) Response of a pixel to a grey strip 2 pixels wide of gray-level "0.4" on a dark background of level "0" moving past the pixel at different speeds. (b) Response of a pixel to a dark strip of gray-level "0.6" on a white background of level "1" moving past the pixel at different speeds. The voltage shown on these curves is not the direct measurement of the voltage at V, but rather V, drives a current-sensing transistor and this current is then sensed by an offchip current sense-amplifier. Adaptive Retina with Center-Surround Receptive Field 683 temporal responses of the circuit. The chip results from the experiments below illustrate the center-surround properties of the network and the difference in time-constants between the surround and center. 3.1 Chip Results Data from the 2D chip is shown in the next few figures. In these experiments, we are only looking at one pixel of the 2D array. A rotating circular fly-wheel stimulus with strips of alternating contrasts is mounted above the chip. The stimulus was created using Mathematica. Figure 3a shows the spati~temporal impulse response of one pixel measured at V, with a small strip at level "0.4" on a dark background of level "0" moving past the pixels on the row. At slow speeds, the impulse response shows a center-surround behavior where the pixel first receives inhibition from the preceding pixels which are excited by the stimulus. When the stimulus moves by the pixel of interest, it is excited and then it is inhibited by the subsequent pixels seeing the stimulus. Tim. (Sec) I o f I i Figure 4: Response of a pixel to a strip of varying contrasts on a dark background moving past the pixel at a constant speed. At faster speeds, the initial inhibition in the response grows smaller until at some even faster speed, the initial inhibition is no longer observed. This response comes about because the inhibition from the surround has a longer-time constant than the center. When the stimulus moves past the pixel of interest, the inhibition from the preceding pixels excited by the stimulus does not have time to inhibit the pixel of interest. Hence the excitation is seen first and then the inhibition comes into place when the stimulus passes by. Note that in these figures (Figures 3-4), the curves have been displaced to show the pixel response at different speeds of the moving stimulus. The voltage shown on these curves is not the direct measurement of the voltage at V, but rather V, drives a current-sensing transistor and this current is then sensed by an off-chip current sense-amplifier. Figure 3b shows the spati~temporal impulse response of one pixel with a similar 684 s.-c. LlU, K. BOAHEN size strip of level "0.6" on a light background of level "1" moving past the row of pixels. The same inhibition behavior is seen for increasing stimulus speeds. Figure 4 shows the output response at V, for the same stimulus of gray-levels varying from "0.2" to "0.8" on a dark background of level "0" moving at one speed. The peak excitation response is plotted against the contrast in Figure 5. A level of "0.2" corresponds to a irradiance of 15mW/m2 while a level of "0.8" corresponds to a irradiance of 37.4mW/m2. These measurements are done with a photometer mounted about 1.5in above a piece of paper with the contrast which is being measured. The irradiance varies exponentially with increasing level. 4 Conclusion In this paper, we described an adaptive retina with a center-surround receptive field. The system properties of this retina allows it to model functionally either the responses of the laminar cells in the invertebrate retina or the outer-plexiform layer of vertebrate retina. We show that the circuit shows adaptation to changes over 5 decades of background intensities. The center-surround property of the network can be seen from its spatio-temporal response to different stimulus speeds. This property serves to remove redundancy in space and time of the input signal. Acknowledgements We thank Carver Mead for his support and encouragement. SC Liu is supported by an NIMH fellowship and K Boahen is supported by a Sloan fellowship. We thank Tobias Delbriick for the inspiration and help in testing the design. We also thank Rahul Sarpeshkar and Bradley Minch for comments. Fabrication was provided by MOSIS. References [1] S. B. Laughlin, "Coding efficiency and design in retinal processing", In: Facets of Vision (D. G. Stavenga and R. C. Hardie, eds) pp. 213-234. Springer, Berlin, 1989. [2] S. R. Shaw, "Retinal resistance barriers and electrica1lateral inhibition", Nature, Lond.255,: 480-483, 1975. [3] M. A. Mahowald, "Silicon Retina with Adaptive Photoreceptors" in SPIE/SPSE Symposium on Electronic Science and Technology: From Neurons to Chips. Orlando, FL, April 1991. [4] K. A. Boahen and A. G. Andreou, "A Contrast Sensitive Silicon Retina with Reciprocal Synapses", In D. S. Touretzky (ed.), Advances in Neural Information Processing Systems 4, 764-772. San Mateo, CA: Morgan Kaufmann, 1992. [5] K. A. Boahen, "Spatiotemporal sensitivity of the retina: A physical model", CNS Memo CNS-TR-91-06, California Institute of Technology, Pasadena, CA 91125, June 1991. [6] T. Delbriick, "Analog VLSI Phototransduction by continous-time, adaptive, logarithmic photoreceptor circuits", CNS Memo No.30, California Institute of Technology, Pasadena, CA 91125, 1994.
|
1995
|
59
|
1,105
|
Laterally Interconnected Self-Organizing Maps in Hand-Written Digit Recognition Yoonsuck Choe, Joseph Sirosh, and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 yschoe,sirosh,risto@cs. u texas .ed u Abstract An application of laterally interconnected self-organizing maps (LISSOM) to handwritten digit recognition is presented. The lateral connections learn the correlations of activity between units on the map. The resulting excitatory connections focus the activity into local patches and the inhibitory connections decorrelate redundant activity on the map. The map thus forms internal representations that are easy to recognize with e.g. a perceptron network. The recognition rate on a subset of NIST database 3 is 4.0% higher with LISSOM than with a regular Self-Organizing Map (SOM) as the front end, and 15.8% higher than recognition of raw input bitmaps directly. These results form a promising starting point for building pattern recognition systems with a LISSOM map as a front end. 1 Introduction Hand-written digit recognition has become one of the touchstone problems in neural networks recently. Large databases of training examples such as the NIST (National Institute of Standards and Technology) Special Database 3 have become available, and real-world applications with clear practical value, such as recognizing zip codes in letters, have emerged. Diverse architectures with varying learning rules have been proposed, including feed-forward networks (Denker et al. 1989; Ie Cun et al. 1990; Martin and Pittman 1990), self-organizing maps (Allinson et al. 1994), and dedicated approaches such as the neocognitron (Fukushima and Wake 1990). The problem is difficult because handwriting varies a lot, some digits are easily confusable, and recognition must be based on small but crucial differences. For example, the digits 3 and 8, 4 and 9, and 1 and 7 have several overlapping segments, and the differences are often lost in the noise. Thus, hand-written digit recognition can be seen as a process of identifying the distinct features and producing an internal representation where the significant differences are magnified, making the recognition easier. Laterally Interconnected Self-organizing Maps in Handwritten Digit Recognition 737 In this paper, the Laterally Interconnected Synergetically Self-Organizing Map architecture (LISSOM; Sirosh and Miikkulainen 1994, 1995, 1996) was employed to form such a separable representation. The lateral inhibitory connections of the LISSOM map decorrelate features in the input, retaining only those differences that are the most significant. Using LISSOM as a front end, the actual recognition can be performed by any standard neural network architecture, such as the perceptron. The experiments showed that while direct recognition of the digit bitmaps with a simple percept ron network is successful 72.3% of the time, and recognizing them using a standard self-organizing map (SOM) as the front end 84.1% of the time, the recognition rate is 88.1 % based on the LISSOM network. These results suggest that LISSOM can serve as an effective front end for real-world handwritten character recognition systems. 2 The Recognition System 2.1 Overall architecture The system consists of two networks: a 20 x 20 LISSOM map performs the feature analysis and decorrelation of the input, and a single layer of 10 perceptrons the final recognition (Figure 1 (a)) . The input digit is represented as a bitmap on the 32 x 32 input layer. Each LISSOM unit is fully connected to the input layer through the afferent connections, and to the other units in the map through lateral excitatory and inhibitory connections (Figure 1 (b)). The excitatory connections are short range, connecting only to the closest neighbors of the unit, but the inhibitory connections cover the whole map. The percept ron layer consists of 10 units, corresponding to digits 0 to 9. The perceptrons are fully connected to the LISSOM map, receiving the full activation pattern on the map as their input. The perceptron weights are learned through the delta rule, and the LISSOM afferent and lateral weights through Hebbian learning. 2.2 LISSOM Activity Generation and Weight Adaptation The afferent and lateral weights in LISSOM are learned through Hebbian adaptation. A bitmap image is presented to the input layer, and the initial activity of the map is calculated as the weighted sum of the input. For unit (i, j), the initial response TJij IS TJij = (7 ('2: eabllij,ab) , a,b (1) where eab is the activation of input unit (a, b), Ilij ,ab is the afferent weight connecting input unit ( a, b) to map unit (i, j), and (7 is a piecewise linear approximation of the sigmoid activation function. The activity is then settled through the lateral connections. Each new activity TJij (t) at step t depends on the afferent activation and the lateral excitation and inhibition: TJiAt) = (7 ('2: eabllij,ab + Ie '2: Eij,kITJkl(t - 1) - Ii '2: Iij,kITJkl(t - 1)), (2) a,b k,l k,l where Eij,kl and Iij,kl are the excitatory and inhibitory connection weights from map unit (k, l) to (i, j) and TJkl(t - 1) is the activation of unit (k , I) during the previous time step. The constants I e and Ii control the relative strength of the lateral excitation and inhibition. After the activity has settled, the afferent and lateral weights are modified according to the Hebb rule. Afferent weights are normalized so that the length of the weight 738 Y. CHOE, J. SIROSH, R. MIIKKULAINEN Output Layer (10) .Lq?'Li7.L:17.La7'LV.87..,.Li7.LWLp' : ...... LISSOM Map LaY~/~~~X20) L::7.L7.L7""'-.L::7LI7 ~ ..... ..,. .... ..c:7\ . ..c:7L:7.&l§7'..,. ..... L7 \. ; .L7.L7.£7.L7LSJ7L7 'L7.AlFL7.L7..c:7L7 .L7.A11P".AIIP"L7.L7.o " :mput L~yer (32x32) ". L7L7~~~~~~~L7L7 L7.L7L:7.L7.L7..c:7L7 L7L7~~~~~~~L7L7 . .L7.L7 .......... ..,..L7..c:7 L7L7L7L7L7L7L7L7L7L7L7. ' .L7..,..L7L::7.L7.L7..c:7 L7L7L7L7L7L7L7L7L7L7L70 ' ..c:7..,..L7 ..... ~..c:7..c:7 0 20 . L7..,...,..L7.L?..,.L7 .... ..c:7..,..L7L7.L7..,..L7 .... L7.L:7..,...,...,.L/.L:7 :L:7.L7..c:7.L7.L7..c:7L7 (a) • Unit OJ) tII'd Units with excitatory lateral connections to (iJ) • Units with inhibitory lateral connections to (iJ) (b) Figure 1: The system architecture. (a) The input layer is activated according to the bitmap image of digit 6. The activation propagates through the afferent connections to the LISSOM map, and settles through its lateral connections into a stable pattern. This pattern is the internal representation of the input that is then recognized by the perceptron layer. Through ,the connections from LISSOM to the perceptrons, the unit representing 6 is strongly activated, with weak activations on other units such as 3 and 8. (b) The lateral connections to unit (i, j), indicated by the dark square, are shown. The neighborhood of excitatory connections (lightly shaded) is elevated from the map for a clearer view. The units in the excitatory region also have inhibitory lateral connections (indicated by medium shading) to the center unit. The excitatory radius is 1 and the inhibitory radius 3 in this case. vector remains the same; lateral weights are normalized to keep the sum of weights constant (Sirosh and Miikkulainen 1994): .. (t + 1) Ilij,mn(t) + crinp1]ij~mn IllJ,mn VLmn[llij,mn(t) + crinp1]ij~mnF' (3) .. (t + 1) _ Wij,kl(t) + cr1]ij1]kl W1J,kl "'" [ ( ) ] , wkl Wij ,kl t + cr1]ij1]kl (4) where Ilij,mn is the afferent weight from input unit (m, n) to map unit (i, j), and crinp is the input learning rate; Wij ,kl is the lateral weight (either excitatory Eij ,kl or inhibitory Iij ,kl) from map unit (k, I) to (i, j), and cr is the lateral learning rate (either crexc or crinh). 2.3 Percept ron Output Generation and Weight Adaptation The perceptrons at the output of the system receive the activation pattern on the LISSOM map as their input. The perceptrons are trained after the LISSOM map has been organized. The activation for the perceptron unit Om is Om = CL1]ij Vij,m, i,j (5) where C is a scaling constant, 1]ij is the LISSOM map unit (i,j), and Vij,m is the connection weight between LISSOM map unit (i,j) and output layer unit m. The delta rule is used to train the perceptrons: the weight adaptation is proportional to the map activity and the difference between the output and the target: Vij,m(t + 1) = Vij,m(t) + crout1]ij((m Om), (6) where crout is the learning rate of the percept ron weights, 1]ij is the LISSOM map unit activity, (m is the target activation for unit m. ((m = 1 if the correct digit = m, 0 otherwise). Laterally Interconnected Self-organizing Maps in Handwritten Digit Recognition 739 I Representation I Training Test LISSOM 93.0/ 0.76 88.1/ 3.10 SOM 84.5/ 0.68 84.1/ 1.71 Raw Input 99.2/ 0.06 72.3/ 5.06 Table 1: Final Recognition Results. The average recognition percentage and its variance over the 10 different splits are shown for the training and test sets. The differences in each set are statistically significant with p > .9999. 3 Experiments A subset of 2992 patterns from the NIST Database 3 was used as training and testing data. 1 The patterns were normalized to make sure taht each example had an equal effect on the LISSOM map (Sirosh and Miikkulainen 1994). LISSOM was trained with 2000 patterns. Of these, 1700 were used to train the perceptron layer, and the remaining 300 were used as the validation set to determine when to stop training the perceptrons. The final recognition performance of the whole system was measured on the remaining 992 patterns, which neither LISSOM nor the perceptrons had seen during training. The experiment was repeated 10 times with different random splits of the 2992 input patterns into training, validation, and testing sets. The LISSOM map can be organized starting from initially random weights. However, if the input dimensionality is large, as it is in case of the 32 X 32 bitmaps, each unit on the map is activated roughly to the same degree, and it is difficult to bootstrap the self-organizing process (Sirosh and Miikkulainen 1994, 1996). The standard Self-Organizing Map algorithm can be used to preorganize the map in this case. The SOM performs preliminary feature analysis of the input, and forms a coarse topological map of the input space. This map can then be used as the starting point for the LISSOM algorithm, which modifies the topological organization and learns lateral connections that decorrelate and represent a more clear categorization of the input patterns. The initial self-organizing map was formed in 8 epochs over the training set, gradually reducing the neighborhood radius from 20 to 8. The lateral connections were then added to the system, and over another 30 epochs, the afferent and lateral weights of the map were adapted according to equations 3 and 4. In the beginning, the excitation radius was set to 8 and the inhibition radius to 20. The excitation radius was gradually decreased to 1 making the activity patterns more concentrated and causing the units to become more selective to particular types of input patterns. For comparison, the initial self-organized map was also trained for another 30 epochs, gradually decreasing the neighborhood size to 1 as well. The final afferent weights for the SOM and LISSOM maps are shown in figures 2 and 3. After the SOM and LISSOM maps were organized, a complete set of activation patterns on the two maps were collected. These patterns then formed the training input for the perceptron layer. Two separate versions were each trained for 500 epochs, one with SOM and the other with LISSOM patterns. A third perceptron layer was trained directly with the input bitmaps as well. Recognition performance was measured by counting how often the most highly active perceptron unit was the correct one. The results were averaged over the 10 different splits. On average, the final LISSOM+perceptron system correctly recognized 88.1% of the 992 pattern test sets. This is significantly better than the 84.1% 1 Downloadable at ftp:j jsequoyah.ncsl.nist.gov jpubjdatabasesj. 740 Y. CHOE, J. SIROSH, R. MIIKKULAINEN Iliiji '·~1,;i;:!il , '8 ....... ···· Slll .... ". "1111 "Q" .. '11 .111/1 ""' <·1,1111 Figure 2: Final Afferent Weights of the SOM map. The digit-like patterns represent the afferent weights of each map unit projected on the input layer. For example, the lower left corner represents the afferent weights of unit (0,0). High weight values are shown in black and low in white. The pattern of weights shows the input pattern to which this unit is most sensitive (6 in this case). There are local clusters sensitive to each digit category. of the SOM+perceptron system, and the 72.3% achieved by the perceptron layer alone (Table 1). These results suggest that the internal representations generated by the LISSOM map are more distinct and easier to recognize than the raw input patterns and the representations generated by the SOM map. 4 Discussion The architecture was motivated by the hypothesis that the lateral inhibitory connections of the LISSOM map would decorrelate and force the map activity patterns to become more distinct. The recognition could then be performed by even the simplest classification architectures, such as the perceptron. Indeed, the LISSOM representations were easier to recognize than the SOM patterns, which lends evidential support to the hypothesis. In additional experiments, the percept ron output layer was replaced by a two-weight-Iayer backpropagation network and a Hebbian associator net, and trained with the same patterns as the perceptrons. The recognition results were practically the same for the perceptron, backpropagation, and Hebbian output networks, indicating that the internal representations formed by the LISSOM map are the crucially important part of the recognition system. A comparison of the learning curves reveals two interesting effects (figure 4). First, even though the perceptron net trained with the raw input patterns initially performs well on the test set, its generalization decreases dramatically during training. This is because the net only learns to memorize the training examples, which does not help much with new noisy patterns. Good internal representations are therefore crucial for generalization. Second, even though initially the settling process of the LISSOM map forms patterns that are significantly easier to recognize than Laterally Interconnected Self-organizing Maps in Handwritten Digit Recognition 741 Figure 3: Final Afferent Weights of the LISSOM map. The squares identify the above-average inhibitory lateral connections to unit (10,4) (indicated by the thick square). Note that inhibition comes mostly from areas of similar functionality (i.e. areas sensitive to similar input), thereby decorrelating the map activity and forming a sparser representation of the input. the initial, unsettled patterns (formed through the afferent connections only), this difference becomes insignificant later during training. The afferent connections are modified according to the final, settled patterns, and gradually learn to anticipate the decorrelated internal representations that the lateral connections form. 5 Conclusion The experiments reported in this paper show that LISSOM forms internal representations of the input patterns that are easier to categorize than the raw inputs and the patterns on the SOM map, and suggest that LISSOM can form a useful front end for character recognition systems, and perhaps for other pattern recognition systems as well (such as speech) . The main direction of future work is to apply the approach to larger data sets, including the full NIST 3 database, to use a more powerful recognition network instead of the perceptron, and to increase the map size to obtain a richer representation of the input space. Acknowledgements This research was supported in part by National Science Foundation under grant #IRI-9309273. Computer time for the simulations was provided by the Pittsburgh Supercomputing Center under grants IRI930005P and IRI940004P, and by a High Performance Computer Time Grant from the University of Texas at Austin. References Allinson, N. M., Johnson, M. J., and Moon, K. J. (1994). Digital realisation of selforganising maps. In Touretzky, D. S., editor, Advances in Neural Information Processing Systems 6. San Mateo, CA: Morgan Kaufmann. 742 "0 ~ 0 () -oe. 100 95 90 85 80 75 Y. CHOE. J. SIROSH. R. MIIKKULAINEN Comparison:Test 'SettIEi<CLlSSOU' 'Unsettled LISSOM' ----. . 'SOM' .... . :.Rawj~~~t' ... . ~---... ------------_.-----~------ -- . ... j . . .... -.---_ ..... --.-......... . .. . . . . ... ,.. . ", .. ~ . '" ..... - ~. ... .. .................... --... .. 7 0 ~ __ ~ ____ L-__ -L ____ L-__ -L ____ L-__ -L ____ L-__ ~ __ ~ o 50 100 150 200 250 300 350 400 450 500 Epochs Figure 4: Comparison of the learning curves, A perceptron network was trained to recognize four different kinds of internal representations: the settled LISSOM patterns, the LISSOM patterns before settling, the patterns on the final SOM network, and raw input bitmaps. The recognition accuracy on the test set was then measured and averaged over 10 simulations. The generalization of the raw input + perceptron system decreases rapidly as the net learns to memorize the training patterns. The difference of using settled and unsettled LISSOM patterns diminishes as the afferent weights of LISSOM learn to take into account the decorrelation performed by the lateral weights. Denker, J. S., Gardner, W. R., Graf, H. P., Henderson, D., Howard, R. E., Hubbard, W., Jackel, L. D., Baird, H. S., and Guyon, I. (1989). Neural network recognizer for hand-written zip code digits. In Touretzky, D. S., editor, Advances in Neural Information Processing Systems 1. San Mateo, CA: Morgan Kaufmann. Fukushima, K., and Wake, N. (1990). Alphanumeric character recognition by neocognitron. In Advanced Neural Computers, 263- 270. Elsevier Science Publishers B.V. (North-Holland). Ie Cun, Y., Boser, B., Denker, J. S., Henderson, D., Howard, R. E., Hubbard, W., and Jackel, 1. D. (1990) . Handwritten digit recognition with a backpropagation network. In Touretzky, D. S., editor, Advances in Neural Information Processing Systems 2. San Mateo, CA: Morgan Kaufmann. Martin, G. L., and Pittman, J. A. (1990). Recognizing hand-printed letters and digits. In Touretzky, D. S., editor, Advances in Neural Information Processing Systems 2. San Mateo, CA: Morgan Kaufmann. Sirosh, J. , and Miikkulainen, R. (1994). Cooperative self-organization of afferent and lateral connections in cortical maps. Biological Cybernetics, 71:66- 78. Sirosh, J., and Miikkulainen, R. (1995). Ocular dominance and patterned lateral connections in a self-organizing model of the primary visual cortex. In Tesauro, G ., Touretzky, D. S., and Leen, T . K., editors, Advances in Neural Information Processing Systems 7. Cambridge, MA: MIT Press. Sirosh, J., and Miikkulainen, R. (1996). Topographic receptive fields and patterned lateral interaction in a self-organizing model of the primary visual cortex. Neural Computation (in press).
|
1995
|
6
|
1,106
|
Does the Wake-sleep Algorithm Produce Good Density Estimators? Peter Dayan Brendan J. Frey, Geoffrey E. Hinton Department of Computer Science University of Toronto Toronto, ON M5S 1A4, Canada {frey, hinton} @cs.toronto.edu Department of Brain and Cognitive Sciences Massachusetts Institute of Technology Cambridge, MA 02139, USA dayan@ai.mit.edu Abstract The wake-sleep algorithm (Hinton, Dayan, Frey and Neal 1995) is a relatively efficient method of fitting a multilayer stochastic generative model to high-dimensional data. In addition to the top-down connections in the generative model, it makes use of bottom-up connections for approximating the probability distribution over the hidden units given the data, and it trains these bottom-up connections using a simple delta rule. We use a variety of synthetic and real data sets to compare the performance of the wake-sleep algorithm with Monte Carlo and mean field methods for fitting the same generative model and also compare it with other models that are less powerful but easier to fit. 1 INTRODUCTION Neural networks are often used as bottom-up recognition devices that transform input vectors into representations of those vectors in one or more hidden layers. But multilayer networks of stochastic neurons can also be used as top-down generative models that produce patterns with complicated correlational structure in the bottom visible layer. In this paper we consider generative models composed of layers of stochastic binary logistic units. Given a generative model parameterized by top-down weights, there is an obvious way to perform unsupervised learning. The generative weights are adjusted to maximize the probability that the visible vectors generated by the model would match the observed data. Unfortunately, to compute the derivatives of the log probability of a visible vector, d, with respect to the generative weights, e, it is necessary to consider all possible ways in which d could be generated. For each possible binary representation a in the hidden units the derivative needs to be weighted by the posterior probability of a given d and e: P(ald, e) = P(ale)p(dla, e)ILP(~le)p(dl~, e). 13 (1) 662 B. J. FREY. G. E. HINTON, P. DAYAN It is intractable to compute P(ald, 9), so instead of minimizing -logP(dI9), we minimize an easily computed upper bound on this quantity that depends on some additional parameters, <1>: -logP(dI9) ~ F(dI9, <1» = - I, Q(al d, <I»logP(a, d19) + I,Q(ald, <I»logQ(ald, <1». (2) a a F(dI9, <1» is a Helmholtz free energy and is equal to -logP(dI9) when the distribution Q(-Id, <1» is the same as the posterior distribution P(-Id, 9). Otherwise, F(dI9, <1» exceeds -logP(dI9) by the asymmetric divergence: D = I,Q(ald, <I»log (Q(ald, <I»IP(ald, 9» . (3) a We restrict Q( -I d, <1» to be a product distribution within each layer that is conditional on the binary states in the layer below and we can therefore compute it efficiently using a bottom-up recognition network. We call a model that uses bottom-up connections to minimize the bound in equation 2 in this way a Helmholtz machine (Dayan, Hinton. Neal and Zemel 1995). The recognition weights <I> take the binary activities in one layer and stochastically produce binary activities in the layer above using a logistic function. So, for a given visible vector, the recognition weights may produce many different representations in the hidden layers, but we can get an unbiased sample from the distribution Q(-Id, <1» in a single bottom-up pass through the recognition network. The highly restricted form of Q( -I d, <1» means that even if we use the optimal recognition weights, the gap between F(dI9, <1» and -logP(dI9) is large for some generative models. However, when F(dI9, <1» is minimized with respect to the generative weights, these models will generally be avoided. F(dI9, <1» can be viewed as the expected number of bits required to communicate a visible vector to a receiver. First we use the recognition model to get a sample from the distribution Q( -I d, <1». Then, starting at the top layer, we communicate the activities in each layer using the top-down expectations generated from the already communicated activities in the layer above. It can be shown that the number of bits required for communicating the state of each binary unit is sklog(qk1pk) + (l-sk)log[(1-qk)/(1-Pk)], where Pk is the top-down probability that Sk is on and qk is the bottom-up probability that Sk is on. There is a very simple on-line algorithm that minimizes F(dI9, <1» with respect to the generative weights. We simply use the recognition network to generate a sample from the distribution Q(-Id, <1» and then we increment each top-down weight 9kj by ESk(SrPj), where 9kj connects unit k to unit j. It is much more difficult to exactly follow the gradient of F(dI9, <1» with respect to the recognition weights, but there is a simple approximate method (Hinton, Dayan, Frey and Neal 1995). We generate a stochastic sample from the generative model and then we increment each bottom-up weight <l>ij by ESi(Sj- f/j) to increase the log probability that the recognition weights would produce the correct activities in the layer above. This way of fitting a Helmholtz machine is called the "wake-sleep" algorithm and the purpose of this paper is to assess how effective it is at performing highdimensional density estimation on a variety of synthetically constructed data sets and two real-world ones. We compare it with other methods of fitting the same type of generative model and also with simpler models for which there are efficient fitting algorithms. 2 COMPETITORS We compare the wake-sleep algorithm with six other density estimation methods. All data units are binary and can take on values dk = 1 (on) and dk = 0 (off). Gzip. Gzip (Gailly, 1993) is a practical compression method based on Lempel-Ziv coding. This sequential data compression technique encodes future segments of data by transmitDoes the Wake-sleep Algorithm Produce Good Density Estimators? 663 ting codewords that consist of a pointer into a buffer of recent past output together with the length of the segment being coded. Gzip's perfonnance is measured by subtracting the length of the compressed training set from the length of the compressed training set plus a subset of the test set. Taking all disjoint test subsets into account gives an overall test set code cost. Since we are interested in estimating the expected perfonnance on one test case, to get a tight lower bound on gzip's perfonnance, the subset size should be kept as small as possible in order to prevent gzip from using early test data to compress later test data. Base Rate Model. Each visible unit k is assumed to be independent of the others with a probability Pk of being on. The probability of vector d is p(d) = Ilk Pkdk (1 - Pk)l- dk . The arithmetic mean of unit k's activity is used to estimate Pk' except in order to avoid serious overfitting, one extra on and one extra off case are included in the estimate. Binary Mixture Model. This method is a hierarchical extension of the base rate model which uses more than one set of base rates. Each set is called a component. Component j has probability 1tj and awards each visible unit k a probability Pjk of being on. The net probability of dis p(d) = Lj 1tj Ilk Pj/k (1 - Pjk)l-dk . For a given training datum, we consider the component identity to be a missing value which must be filled in before the parameters can be adjusted. To accomplish this, we use the expectation maximization algorithm (Dempster, Laird and Rubin 1977) to maximize the log-likelihood of the training set, using the same method as above to avoid serious overfitting. Gibbs Machine (GM). This machine uses the same generative model as the Helmholtz machine, but employs a Monte Carlo method called Gibbs sampling to find the posterior in equation 1 (Neal, 1992). Unlike the Helmholtz machine it does not require a separate recognition model and with sufficiently prolonged sampling it inverts the generative model perfectly. Each hidden unit is sampled in fixed order from a probability distribution conditional on the states of the other hidden and visible units. To reduce the time required to approach equilibrium, the network is annealed during sampling. Mean Field Method (MF). Instead of using a separate recognition model to approximate the posterior in equation 1, we can assume that the distribution over hidden units is factorial for a given visible vector. Obtaining a good approximation to the posterior is then a matter of minimizing free energy with respect to the mean activities. In our experiments, we use the on-line mean field learning algorithm due to Saul, Jaakkola, and Jordan (1996). Fully Visible Belief Network (FVBN). This method is a special case of the Helmholtz machine where the top-down network is fully connected and there are no hidden units. No recognition model is needed since there is no posterior to be approximated. 3 DATA SETS The perfonnances of these methods were compared on five synthetic data sets and two real ones. The synthetic data sets had matched complexities: the generative models that produced them had 100 visible units and between 1000 and 2500 parameters. A data set with 100,000 examples was generated from each model and then partitioned into 10,000 for training, 10,000 for validation and 80,000 for testing. For tractable cases, each data set entropy was approximated by the negative log-likelihood of the training set under its generative model. These entropies are approximate lower bounds on the perfonnance. The first synthetic data set was generated by a mixture model with 20 components. Each component is a vector of 100 base rates for the 100 visible units. To make the data more realistic, we arranged for there to be many different components whose base rates are all extreme (near 0 or 1) representing well-defined clusters and a few components with most base rates near 0.5 representing much broader clusters. For componentj, we selected base rate Pjk from a beta distribution with mean Ilt and variance 1lt(1-1lt)/40 (we chose this variance to keep the entropy of visible units low for Ilt near 0 or 1, representing well-defined clusters). Then, as often as not we randomly replaced each Pjk with 1-Pjk to 664 B. 1. FREY, G. E. HINTON, P. DAY AN make each component different (without doing this, all components would favor all units off). In order to obtain many well-defined clusters, the component means Il.i were themselves sampled from a beta distribution with mean 0.1 and variance 0.02. The next two synthetic data sets were produced using sigmoidal belief networks (Neal 1992) which are just the generative parts of binary stochastic Helrnhol tz machines. These networks had full connectivity between layers, one with a 20~100 architecture and one with a 5~10~15~2~100 architecture. The biases were set to 0 and the weights were sampled uniformly from [-2,2), a range chosen to keep the networks from being deterministic. The final two synthetic data sets were produced using Markov random fields. These networks had full bidirectional connections between layers. One had a 10<=>20<=>100 architecture, and the other was a concatenation of ten independent 10<=>10 fields. The biases were set to 0 and the weights were sampled from the set {-4, 0, 4} with probabilities {0.4, 0.4, 0.2}. To find data sets with high-order structure, versions of these networks were sampled until data sets were found for which the base rate method performed badly. We also compiled two versions of a data set to which the wake-sleep algorithm has previously been applied (Hinton et al. 1995). These data consist of normalized and quantized 8x8 binary images of handwritten digits made available by the US Postal Service Office of Advanced Technology. The first version consists of a total of 13,000 images partitioned as 6000 for training, 2000 for validation and 5000 for testing. The second version consists of pairs of 8x8 images (ie. 128 visible units) made by concatenating vectors from each of the above data sets with those from a random reordering of the respective data set. 4 TRAINING DETAILS The exact log-likelihoods for the base rate and mixture models can be computed, because these methods have no or few hidden variables. For the other methods, computing the exact log-likelihood is usually intractable. However, these methods provide an approximate upper bound on the negative log-likelihood in the form of a coding cost or Helmholtz free energy, and results are therefore presented as coding costs in bits. Because gzip performed poorly on the synthetic tasks, we did not break up the test and validation sets into subsets. On the digit tasks, we broke the validation and test sets up to make subsets of 100 visible vectors. Since the "-9" gzip option did not improve performance significantly, we used the default configuration. To obtain fair results, we tried to automate the model selection process subject to the constraint of obtaining results in a reasonable amount of time. For the mixture model, the Gibbs machine, the mean field method, and the Helmholtz machine, a single learning run was performed with each of four different architectures using performance on a validation set to avoid wasted effort. Performance on the validation set was computed every five epochs, and if two successive validation performances were not better than the previous one by more than 0.2%, learning was terminated. The network corresponding to the best validation performance was selected for test set analysis. Although it would be desirable to explore a wide range of architectures, it would be computationally ruinous. The architectures used are given in tables 3 and 4 in the appendix. The Gibbs machine was annealed from an initial temperature of 5.0. Between each sweep of the network, during which each hidden unit was sampled once, the temperature was multiplied by 0.9227 so that after 20 sweeps the temperature was 1.0. Then, the generative weights were updated using the delta rule. To bound the datum probability, the network is annealed as above and then 40 sweeps at unity temperature are performed while summing the probability over one-nearest-neighbor configurations, checking for overlap. A learning rate of 0.01 was used for the Gibbs machine, the mean field method, the Helmholtz machine, and the fully visible belief network. For each of these methods, this value was found to be roughly the largest possible learning rate that safely avoided oscillations. Does the Wake-sleep Algorithm Produce Good Density Estimators? 70r-----------------------------------------------------, 60 50 40 20 Gzip Base rate model -Mixture model -e-Gibbs machine Mean field method -4-Fully visible belief network Entropy • l:~m - 10~--------------------------------------------------~ Mixture 2~1()() BN BN 5~IO~ MRF MRF Single 2~I()() 15~20~I()() 1~2~l()() lOx (IO~IO) digits Tasks Digit pairs 665 Figure 1. Compression performance relative to the Helmholtz machine. Lines connecting the data points are for visualization only, since there is no meaningful interpolant. 5 RESULTS The learning times and the validation performances are given in tables 3 and 4 of the appendix. Test set appraisals and total learning times are given in table 1 for the synthetic tasks and in table 2 for the digit tasks. Because there were relatively many training cases in each simulation, the validation procedure serves to provide timing information more than to prevent overfitting. Gzip and the base rate model were very fast, followed by the fully visible belief network, the mixture model, the Helmholtz machine, the mean field method, and finally the Gibbs machine. Test set appraisals are summarized by compression performance relative to the Helmholtz machine in figure 1 above. Greater compression sizes correspond to lower test set likelihoods and imply worse density estimation. When available, the data set entropies indicate how close to optimum each method comes. The Helmholtz machine yields a much lower cost compared to gzip and base rates on all tasks. Compared to the mixture model, it gives a lower cost on both BN tasks and the MRF 10 x (1O~1O) task. The latter case shows that the Helmholtz machine was able to take advantage of the independence of the ten concatenated input segments, whereas the mixture method was not. Simply to represent a problem where there are only two distinct clusters present in each of the ten segments, the mixture model would require 210 components. Results on the two BN tasks indicate the Helmholtz machine is better able to model multiple simultaneous causes than the mixture method, which requires that only one component (cause) is active at a time. On the other hand, compared to the mixture model, the Helmholtz machine performs poorly on the Mixture 20~100 task. It is not able to learn that only one cause should be active at a time. This problem can be avoided by hard-wiring softmax groups into the Helmholtz machine. On the five synthetic tasks, the Helmholtz machine performs about the same as or better than the Gibbs machine, and runs two orders of magnitude faster. (The Gibbs machine was too slow to run on the digit tasks.) While the quality of density estimation produced by the mean field method is indistinguishable from the Helmholtz machine, the latter runs an order of magnitude faster than the mean field algorithm we used. The fully visible belief network performs significantly better than the Helmholtz machine on the two digit tasks and significantly worse on two of the synthetic tasks. It is trained roughly two orders of magnitude faster than the Helmholtz machine. 666 B. J. FREY, G. E. HINTON, P. DAYAN Table 1. Test set cost (bits) and total training time (hrs) for the synthetic tasks. Model used to produce synthetic data Mixture BN BN5~1O~ MRF MRF 20=>100 20=>100 15~20~100 10~20~100 10 x (1O~1O) Entropy 36.5 63.5 unknown 19.2 36.8 gzip 61.4 o 98.0 0 92.1 o 35.6 0 59.9 0 Base rates 96.6 o 80.7 0 69.2 o 42.2 0 68.1 0 Mixture 36.7 o 74.0 0 62.6 1 19.3 1 49.6 1 GM 44.1 131 63.9 240 58.1 251 26.1 195 40.3 145 MF 42.2 68 64.7 80 58.4 68 19.3 75 38.7 89 HM 42.7 8 65.2 3 58.5 4 19.4 2 38.6 4 FVBN 50.9 o 67.8 0 60.6 0 19.8 0 38.2 0 Table 2. Test set cost (bits) and training time (hrs) for the digit tasks. Method Single digits Method Digit pairs gzip 44.3 0 gZlp 89.2 0 Base rates 59.2 o Base rates 118.4 0 Mixture 37.5 o Mixture 92.7 1 MF 39.5 38 MF 80.7 104 HM 39.1 2 HM 80.4 7 FVBN 35.9 o FVBN 72.9 0 6 CONCLUSIONS If we were given a new data set and asked to leave our research biases aside and do efficient density estimation, how would we proceed? Evidently it would not be worth trying gzip and the base rate model. We'd first try the fully visible belief network and the mixture model, since these are fast and sometimes give good estimates. Hoping to extract extra higher-order structure, we would then proceed to use the Helmholtz machine or the mean field method (keeping in mind that our implementation of the Helmholtz machine is considerably faster than Saul et al. 's implementation of the mean field method). Because it is so slow, we would avoid using the Gibbs machine unless the data set was very small. Acknowledgments We greatly appreciate the mean field software provided by Tommi Jaakkola and Lawrence Saul. We thank members of the Neural Network Research Group at the University of Toronto for helpful advice. The financial support from ITRC, IRIS, and NSERC is appreciated. References Dayan, P., Hinton, G. E., Neal, R. M., and Zemel, R. S. 1995. The Helmholtz machine. Neural Computation 7, 889-904. Dempster, A. P., Laird, N. M. and Rubin, D. B. 1977. Maximum likelihood from incomplete data via the EM algorithm. J. Royal Statistical Society, Series B 34, 1-38. Gailly, J. 1993. gzip program for unix. Hinton, G. E., Dayan, P., Frey, B. J., Neal, R. M. 1995. The wake-sleep algorithm for unsupervised neural networks. Science 268, 1158-1161. Neal, R. M. 1992. Connectionist learning of belief networks. Artificial Intelligence 56,71-113. Saul, L. K., Jaakkola, T., and Jordan, M.I. 1996. Mean field theory for sigmoid belief networks. Submitted to Journal of Artificial Intelligence. Does the Wake-sleep Algorithm Produce Good Density Estimators? 667 Appendix The average validation set cost per example and the associated learning time for each simulation are listed in tables 3 and 4. Architectures judged to be optimal according to validation performance are indicated by "*,, and were used to produce the test results given in the body of this paper. Table 3. Validation set cost (bits) and learning time (min) for the synthetic tasks. Model used to produce synthetic data Mixture BN BN 5~lO~ MRF MRF 20~100 20~100 15~20~100 10<=>20<=> 1 00 10 x (1O<=> 10) gzip 61.6 o 98.1 0 92.3 o 35.6 o 60.0 0 Base rates 96.7 o 80.7 0 69.4 o 42.1 o 68.1 0 Mixture 20~100 44.6 3 75.6 3 63.9 4 19.2* 3 54.8 5 Mixture 40~100 36.8* 5 74.8 5 63.2 7 19.2 7 52.4 15 Mixture 60~100 36.8 7 74.4 7 62.9 8 19.2 8 51.0 17 Mixture lOO~lOO 37.0 14 74.0* 12 62.7* 13 19.3 12 49.6* 22 OM 20~lOO 50.6 1187 63.9* 1639 58.1* 2084 26.1* 934 40.3* 1425 OM 50~lOO 68.8 2328 80.4 3481 76.4 5234 49.2 6472 56.5 3472 OM 1O~20~100 44.1* 872 66.4 1771 59.8 3084 28.0 767 42.3 1033 OM 20~50~100 52.7 3476 91.3 7504 88.0 4647 55.3 3529 63.5 2781 MF 20~100 49.5 518 64.6 427 58.4* 497 19.4 862 39.2 471 MF 50~100 49.9 1644 64.8 1945 58.6 1465 20.4 1264 38.7* 2427 MF 1O~20~100 46.0 306 64.6* 658 58.5 543 19.3* 569 38.9 882 MF 20~50~100 42.1* 1623 65.0 1798 58.6 1553 19.3 1778 38.8 1575 HM 20~lOO 50.0 41 65.2 28 58.8 41 19.7 15 38.6* 30 HM 50~lOO 50.7 81 65.5 66 59.4 78 20.2 27 38.9 46 HM lO~20~100 43.4 32 65.1* 38 58.5* 45 19.4* 21 38.9 46 HM 20~50~lOO 42.6* 308 67.2 69 59.2 93 19.5 64 39.4 102 FVBN 51.0 7 67.8 7 60.7 6 19.8 8 38.3 6 Table 4. Validation set cost (bits) and learning time (min) for the digit tasks. Method Single digits Method Digit pairs gzip 44.2 0 gzip 88.8 1 Base rates 59.0 0 Base rates 117.9 0 Mixture 16~64 43.2 1 Mixture 32~128 96.9 6 Mixture 32~64 40.0 4 Mixture 64~128 93.8 8 Mixture 64~64 38.0 5 Mixture 128~128 92.4* 14 Mixture 128~64 37.1* 6 Mixture 256~128 92.8 27 MF 16~24~64 39.9 341 MF 16~24~32~128 82.7 1335 MF24~32~64 39.1* 845 MF 16~32~64~128 81.2 1441 MF 12~16~24~64 39.8 475 MF 12~16~24~32~128 82.8 896 MF 16~24~32~64 39.1 603 MF 12~16~32~64~128 80.1* 2586 HM 16~24~64 39.7 24 HM 16~24~32~128 83.8 76 HM 24~32~64 39.4 34 HM 16~32~64~128 80.1* 138 HM 12~16~24~64 40.4 16 HM 12~16~24~32~128 84.6 74 HM 16~24~32~64 38.9* 52 HM 12~16~32~64~128 80.1 135 FVBN 35.8 1 FVBN 72.5 7 PART V IMPLEMENTATIONS
|
1995
|
60
|
1,107
|
EM Optimization of Latent-Variable Density Models Christopher M Bishop, Markus Svensen and Christopher K I Williams Neural Computing Research Group Aston University, Birmingham, B4 7ET, UK c.m.bishop~aston.ac.uk svensjfm~aston.ac.uk c.k.i.williams~aston.ac.uk Abstract There is currently considerable interest in developing general nonlinear density models based on latent, or hidden, variables. Such models have the ability to discover the presence of a relatively small number of underlying 'causes' which, acting in combination, give rise to the apparent complexity of the observed data set. Unfortunately, to train such models generally requires large computational effort. In this paper we introduce a novel latent variable algorithm which retains the general non-linear capabilities of previous models but which uses a training procedure based on the EM algorithm. We demonstrate the performance of the model on a toy problem and on data from flow diagnostics for a multi-phase oil pipeline. 1 INTRODUCTION Many conventional approaches to density estimation, such as mixture models, rely on linear superpositions of basis functions to represent the data density. Such approaches are unable to discover structure within the data whereby a relatively small number of 'causes' act in combination to account for apparent complexity in the data. There is therefore considerable interest in latent variable models in which the density function is expressed in terms of of hidden variables. These include density networks (MacKay, 1995) and Helmholtz machines (Dayan et al., 1995). Much of this work has been concerned with predicting binary variables. In this paper we focus on continuous data. 466 c. M. BISHOP, M. SVENSEN, C. K. I. WILLIAMS y(x;W) Figure 1: The latent variable density model constructs a distribution function in t-space in terms of a non-linear mapping y(x; W) from a latent variable x-space. 2 THE LATENT VARIABLE MODEL Suppose we wish to model the distribution of data which lives in aD-dimensional space t = (tl, ... , tD). We first introduce a transformation from the hidden variable space x = (Xl, ... , xL) to the data space, governed by a non:-linear function y(x; W) which is parametrized by a matrix of weight parameters W. Typically we are interested in the situation in which the dimensionality L of the latent variable space is less than the dimensionality D of the data space, since we wish to capture the fact that the data itself has an intrinsic dimensionality which is less than D. The transformation y(x; W) then maps the hidden variable space into an L-dimensional non-Euclidean subspace embedded within the data space. This is illustrated schematically for the case of L = 2 and D = 3 in Figure 1. If we define a probability distribution p(x) on the latent variable space, this will induce a corresponding distribution p(y) in the data space. We shall refer to p(x) as the prior distribution of x for reasons which will become clear shortly. Since L < D, the distribution in t-space would be confined to a manifold of dimension L and hence would be singular. Since in reality data will only approximately live on a lower-dimensional space, it is appropriate to include a noise model for the t vector. We therefore define the distribution of t, for given x and W, given by a spherical Gaussian centred on y(x; W) having variance {3-1 so that ( 1) The distribution in t-space, for a given value of the weight matrix W, lS then obtained by integration over the x-distribution p(tIW) = J p(tlx, W)p(x) dx. (2) For a given data set V = (t l , ... , t N ) of N data points, we can determine the weight matrix W using maximum likelihood. For convenience we introduce an error function given by the negative log likelihood: N N E(W) = -In 11 p(tn IW) = - ~ In {J p(tn Ixn, W)p(xn) dxn } . (3) EM Optimization of Latent-Variable Density Models 467 In principle we can now seek the maximum likelihood solution for the weight matrix, once we have specified the prior distribution p(x) and the functional form of the mapping y(x; W), by minimizing E(W). However, the integrals over x occuring in (3), and in the corresponding expression for 'iJ E, will, in general, be analytically intractable. MacKay (1995) uses Monte Carlo techniques to evaluate these integrals and conjugate gradients to find the weights. This is computationally very intensive, however, since a Monte Carlo integration must be performed every time the conjugate gradient algorithm requests a value for E(W) or 'iJ E(W). We now show how, by a suitable choice of model, it is possible to find an EM algorithm for determining the weights. 2.1 EM ALGORITHM There are three key steps to finding a tractable EM algorithm for evaluating the weights. The first is to use a generalized linear network model for the mappmg function y(x; W). Thus we write y(x; W) = W ¢(x) (4) where the elements of ¢(x) consist of M fixed basis functions cPj(x), and W is a D x M matrix with elements Wkj' Generalized linear networks possess the same universal approximation capabilities as multi-layer adaptive networks. The price which has to be paid, however, is that the number of basis functions must typically grow exponentially with the dimensionality L of the input space. In the present context this is not a serious problem since the dimensionality is governed by the latent variable space and will typically be small. In fact we are particularly interested in visualization applications, for which L = 2. The second important step is to use a simple Monte Carlo approximation for the integrals over x. In general, for a function Q(x) we can write J 1 K Q(x)p(x) dx ~ f{ ~ Q(xi ) z=l (5) where xi represents a sample drawn from the distribution p(x). If we apply this to (3) we obtain E(W) = - t,ln{ ~ tp(tnlxni,w)} (6) The third key step to choose the sample of points {xni} to be the same for each term in the summation over n. Thus we can drop the index n on x ni to give N {I K } E(W) = - ~ In f{ ~p(tnlxi, W) (7) We now note that (7) represents the negative log likelihood under a distribution consisting of a mixture of f{ kernel functions. This allows us to apply the EM algorithm to find the maximum likelihood solution for the weights. Furthermore, as a consequence of our choice (4) for the non-linear mapping function, it will turn out that the M-step can be performed explicitly, leading to a solution in terms of a set 468 c. M. BISHOP, M. SYENSEN, C. K. I. WILLIAMS of linear equations. We note that this model corresponds to a constrained Gaussian mixture distribution of the kind discussed in Hinton et al. (1992). We can formulate the EM algorithm for this system as follows. Setting the derivatives of (7) with respect to Wkj to zero we obtain t, t, R;n(W) {t, w"f,(x;) - t~ } f;(x;) = 0 (8) where we have used Bayes' theorem to introduce the posterior probabilities, or responsibilities, for the mixture components given by R- (W) = p(tnlxi, W) m L:~=1 p(tnlxil, W) (9) Similarly, maximizing with respect to (3 we obtain K N ~ = N1D I: I: Rni(W) lIy(xn; W) - t n ll 2 . i=l n=l (10) The EM algorithm is obtained by supposing that, at some point in the algorithm, the current weight matrix is given by wold and the current value of (3 is (30ld. Then we can evaluate the responsibilities using these values for Wand (3 (the E-step), and then solve (8) for the weights to give W new and subsequently solve (10) to give (3new (the M-step). The two steps are repeated until a suitable convergence criterion is reached. In practice the algorithm converges after a relatively small number of iterations. A more formal justification for the EM algorithm can be given by introducing auxiliary variables to label which component is responsible for generating each data point, and then computing the expectation with respect to the distribution of these variables. Application of Jensen's inequality then shows that, at each iteration of the algorithm, the error function will decrease unless it is already at a (local) minimum, as discussed for example in Bishop (1995). If desired, a regularization term can be added to the error function to control the complexity of the model y(x; W). From a Bayesian viewpoint, this corresponds to a prior distribution over weights. For a regularizer which is a quadratic function of the weight parameters, this leads to a straightforward modification to the weight update equations. It is convenient to write the condition (8) in matrix notation as (~TGold~ + AI)(Wnew)T = ~TTold (11) where we have included a regularization term with coefficient A, and I denotes the unit matrix. In (11) ~ is a f{ x M matrix with elements <l>ij = (/Jj(xi ), T is a I< x D matrix, and G is a I< x I< diagonal matrix, with elements N N Tik = I: Rin(W)t~ Gjj = I: ~n(W). (12) n=l n=l We can now solve (11) for w new using standard linear matrix inversion techniques, based on singular value decomposition to allow for possible ill-conditioning. Note that the matrix ~ is constant throughout the algorithm, and so need only be evaluated once at the start. EM Optimization of Latent-Variable Density Models 469 4~----------------------~ 4~----------------------~ 3 3 • , : 1' . • 2 2 o 11 o • -_11~--~----~--~----~--~ -1~--~----~------~~--~ o 2 3 4 -1 0 2 3 4 Figure 2: Results from a toy problem involving data (' x') generated from a 1-dimensional curve embedded in 2 dimensions, together with the projected sample points ('+') and their Gaussian noise distributions (filled circles). The initial configuration, determined by principal component analysis, is shown on the left, and an intermediate configuration, obtained after 4 iterations of EM, is shown on the right. 3 RESULTS We now present results from the application of this algorithm first to a toy problem involving data in three dimensions, and then to a more realistic problem involving 12-dimensional data arising from diagnostic measurements of oil flows along multiphase pipelines. For simplicity we choose the distribution p(x) to be uniform over the unit square. The basis functions ¢j (x) are taken to be spherically symmetric Gaussian functions whose centres are distributed on a uniform grid in x-space, with a common width parameter chosen so that the standard deviation is equal to the separation of neighbouring basis functions. For both problems the weights in the network were initialized by performing principal components analysis on the data and then finding the least-squares solution for the weights which best approximates the linear transformation which maps latent space to target space while generating the correct mean and variance in target space. As a simple demonstration of this algorithm, we consider data generated from a one-dimensional distribution embedded in two dimensions, as shown in Figure 2. 3.1 OIL FLOW DATA Our second example arises in the problem of determining the fraction of oil in a multi-phase pipeline carrying a mixture of oil, water and gas (Bishop and James, 1993). Each data point consists of 12 measurements taken from dual-energy gamma densitometers measuring the attenuation of gamma beams passing through the pipe. Synthetically generated data is used which models accurately the attenuation processes in the pipe, as well as the presence of noise (arising from photon statistics). The three phases in the pipe (oil, water and gas) can belong to one of three different geometrical configurations, corresponding to stratified, homogeneous, and annular flows, and the data set consists of 1000 points distributed equally between the 3 470 c. M. BISHOP, M. SVENSEN, C. K. I. W1LUAMS 2~----~------~-------------, 1.5 ~ ..,..~ • • ..... .,,;' .. .." 0.5 00 ...", 0 " 0 ~ ~iI:" • , ~,J1 :"." +~+ .,. • ~ ........ 0 -0.5 .. ,.,..t:'" .. -, ++ , . ... (~ ~+ -1 • ~O ~ #+ .. 0 6 C -1.5 : .. ~ ~_" ".a:. 0 .'IIIt. .. • aa.; ..... -32 -1 0 2 -2 0 2 4 Figure 3: The left plot shows the posterior-mean projection of the oil data in the latent space of the non-linear model. The plot on the right shows the same data set projected onto the first two principal components. In both plots, crosses, circles and plus-signs represent the stratified, annular and homogeneous configurations respectively. classes. We take the latent variable space to be two-dimensional. This is appropriate for this problem as we know that, locally, the data must have an intrinsic dimensionality of two (neglecting noise on the data) since, for any given geometrical configuration of the three phases, there are two degrees of freedom corresponding to the fractions of oil and water in the pipe (the fraction of gas being redundant since the three fractions must sum to one). It also allows us to use the latent variable model to visualize the data by projection onto x-space. For the purposes of visualization, we note that a data point t n induces a posterior distribution p(xltn, W*) in x-space, where W* denotes the value of the weight matrix for the trained network. This provides considerably more information in the visualization space than many simple techniques (which generally project each data point onto a single point in the visualization space). For example, the posterior distribution may be multi-modal, indicating that there is more than one region of x-space which can claim significant responsibility for generating the data point. However, it is often convenient to project each data point down to a unique point in x-space. This can be done by finding the mean of the posterior distribution, which itself can be evaluated by a simple Monte Carlo integration using quantities already calculated in the evaluation of W* . Figure 3 shows the oil data visualized in the latent-variable space in which, for each data point, we have plotted the posterior mean vector. Again the points have been labelled according to their multi-phase configuration. We have compared these results with those from a number of conventional techniques including factor analysis and principal component analysis. Note that factor analysis is precisely the model which results if a linear mapping is assumed for y(x; W), a Gaussian distribution p(x) is chosen in the latent space, and the noise distribution in data space is taken to be Gaussian with a diagonal covariance matrix. Of these techniques, principal component analysis gave the best class separation (assessed subjectively) and is illustrated in Figure 3. Comparison with the results from the non-linear model clearly shows that the latter gives much better separation of the three classes, as a consequence of the non-linearity permitted by the latent variable mapping. EM Optimization of Latent-Variable Density Models 471 4 DISCUSSION There are interesting relationships between the model discussed here and a number of well-known algorithms for unsupervised learning. We have already commented that factor analysis is a special case of this model, involving a linear mapping from latent space to data space. The Kohonen topographic map algorithm (Kohonen, 1995) can be regarded as an approximation to a latent variable density model of the kind outlined here. Finally, there are interesting similarities to a 'soft' version of the 'principal curves' algorithm (Tibshirani, 1992). The model we have described can readily be extended to deal with the problem of missing data, provided we assume that the missing data is ignorable and missing at random (Little and Rubin, 1987). This involves maximizing the likelihood function in which the missing values have been integrated out. For the model discussed here, the integrations can be performed analytically, leading to a modified form of the EM algorithm. Currently we are extending the model to allow for mixed continuous and categorical variables. We are also exploring Bayesian approaches, based on Markov chain Monte Carlo, to replace the maximum likelihood procedure. Acknowledgements This work was partially supported by EPSRC grant GR/J75425: Novel Developments in Learning Theory. Markus Svensen would like to thank the staff of the SANS group in Stockholm for their hospitality during part of this project. References Bishop, C. M. (1995). Neural Networks for Pattern Recognition. Oxford University Press. Bishop, C. M. and G. D. James (1993). Analysis of multiphase flows using dualenergy gamma densitometry and neural networks. Nuclear Instruments and Methods in Physics Research A327, 580-593. Dayan, P., G. E. Hinton, R. M. Neal, and R. S. Zemel (1995). The HelmQoltz machine. Neural Computation 7 (5), 889- 904. Hinton, G. E., C. K. 1. Williams, and M. D. Revow (1992). Adaptive elastic models for hand-printed character recognition. In J. E. Moody, S. J. Hanson, and R. P. Lippmann (Eds.), Advances in Neural Information Processing Systems 4. Morgan Kauffmann. Kohonen, T. (1995). Self-Organizing Maps. Berlin: Springer-Verlag. Little, R. J. A. and D. B. Rubin (1987). Statistical Analysis with Missing Data. New York: John Wiley. MacKay, D. J. C. (1995). Bayesian neural networks and density networks. Nuclear Instruments and Methods in Physics Research, A 354 (1), 73- 80. Tibshirani, R. (1992). Principal curves revisited. Statistics and Computing 2, 183-190.
|
1995
|
61
|
1,108
|
Modern Analytic Techniques to Solve the Dynamics of Recurrent Neural Networks A.C.C. Coolen Dept. of Mathematics King's College London Strand, London WC2R 2LS, U.K. S.N. Laughton Dept. of Physics - Theoretical Physics University of Oxford 1 Keble Road, Oxford OX1 3NP, U.K. D. Sherrington .. Center for Non-linear Studies Los Alamos National Laboratory Los Alamos, New Mexico 87545 Abstract We describe the use of modern analytical techniques in solving the dynamics of symmetric and nonsymmetric recurrent neural networks near saturation. These explicitly take into account the correlations between the post-synaptic potentials, and thereby allow for a reliable prediction of transients. 1 INTRODUCTION Recurrent neural networks have been rather popular in the physics community, because they lend themselves so naturally to analysis with tools from equilibrium statistical mechanics. This was the main theme of physicists between, say, 1985 and 1990. Less familiar to the neural network community is a subsequent wave of theoretical physical studies, dealing with the dynamics of symmetric and nonsymmetric recurrent networks. The strategy here is to try to describe the processes at a reduced level of an appropriate small set of dynamic macroscopic observables. At first, progress was made in solving the dynamics of extremely diluted models (Derrida et al, 1987) and of fully connected models away from saturation (for a review see (Coolen and Sherrington, 1993)). This paper is concerned with more recent approaches, which take the form of dynamical replica theories, that allow for a reliable prediction of transients, even near saturation. Transients provide the link between initial states and final states (equilibrium calculations only provide ·On leave from Department of Physics - Theoretical Physics, University of Oxford 254 A. C. C. COOLEN, S. N. LAUGHTON, D. SHERRINGTON information on the possible final states). In view of the technical nature of the subject, we will describe only basic ideas and results for simple models (full details and applications to more complicated models can be found elsewhere). 2 RECURRENT NETWORKS NEAR SATURATION Let us consider networks of N binary neurons ai E {-I, I}, where neuron states are updated sequentially and stochastically, driven by the values of post-synaptic potentials hi. The probability to find the system at time t in state 0' = (a1,' .. , aN) is denoted by Pt(O'). For the rates Wi(O') of the transitions ai -t -(7i and for the potentials hi (0') we make the usual choice 1 Wi (0') = - [1-ai tanh [,Bhi (0')]] 2 hi(O') = L Jijaj j:f:i The parameter ,B controls the degree of stochasticity: the ,B = 0 dynamics is completely random, whereas for ,B = 00 we find the deterministic rule ai -t sgn[hi(O')]. The evolution in time of Pt(O') is given by the master equation d N dtPt (0') = l: [Pt (FkO' )Wk (FkO') - Pt (0' )Wk (0')] k=l (1) with Fk<P(O') = <P(a1 , ... ,-(7k, ... ,aN)' For symmetric models, where Jij = Jji for all (ij), the dynamics (1) leads asymptotically to the Boltzmann equilibrium distribution Peq(O') '" exp [-,BE(O')], with the energy E(O') = - Li<j adijaj. For associative memory models with Hebbian-type synapses, required to store a set of P random binary patterns e/.1 = (€i, ... , €~ ), the relevant macroscopic observable is the overlap m between the current microscopic state 0' and the pattern to be retrieved (say, pattern 1): m = -Iv Li €lai. Each post-synaptic potential can now be written as the sum of a simple signal term and an interference-noise term, e.g. 1 p=o:N Jij = N L €f€j /.1=1 hi(O') = m€l + ~ l: €f l: €jaj /.1>1 j:f:i (2) All complications arise from the noise terms. The 'Local Chaos Hypothesis' (LCH) consists of assuming the noise terms to be independently distributed Gaussian variables. The macroscopic description then consists of the overlap m and the width ~ of the noise distribution (Amari and Maginu, 1987). This, however, works only for states near the nominated pattern, see also (Nishimori and Ozeki, 1993). In reality the noise components in the potentials have far more complicated statisticsl . Due to the build up of correlations between the system state and the non-nominated patterns, the noise components can be highly correlated and described by bi-modal distributions. Another approach involves a description in terms of correlation- and response functions (with two timearguments). Here one builds a generating functional, which is a sum over all possible trajectories in state space, averaged over the distribution of the non-nominated patterns. One finds equations which are exact for N -t 00 , but, unfortunately, also rather complicated. For the typical neural network models solutions are known only in equilibrium (Rieger et aI, 1988); information on transients has so far only been obtained through cumbersome approximation schemes (Horner et aI, 1989). We now turn to a theory that takes into account the non-trivial statistics of the post-synaptic potentials, yet involves observables with one time-argument only. lCorrelations are negligible only in extremely diluted (asymmetric) networks (Derrida et aI, 1987), and in networks with independently drawn (asymmetric) random synapses Modem Analytic Techniques to Solve the Dynamics of Recurrent Neural Networks 255 3 DYNAMICAL REPLICA THEORIES The evolution of macroscopic observables n( 0') = (01 (0'), ... , OK (0')) can be described by the so-called Kramers-Moyal expansion for the corresponding probability distribution pt(n) (derived directly from (1)). Under certain conditions on the sensitivity of n to single-neuron transitions (7i -t -1J'i, one finds on finite time-scales and for N -t 00 the macroscopic state n to evolve deterministically according to: ~n = EO' pt(O')8 [n-n(O')] Ei Wi(O') [n(FiO')-n(O')] (3) dt EO' pt(O')8 [n-n(O')] This equation depends explicitly on time through Pt(O'). However, there are two natural ways for (3) to become autonomous: (i) by the term Ei Wi(O') [n(FiO') -n(O')] depending on u only through n(O') (as for attractor networks away from saturation), or (ii) by (1) allowing for solutions of the form Pt(O') = fdn(O')] (as for extremely diluted networks). In both cases Pt(O') drops out of (3). Simulations further indicate that for N -t 00 the macroscopic evolution usually depends only on the statistical properties of the patterns {ell}, not on their microscopic realisation ('self-averaging'). This leads us to the following closure assumptions: 1. Probability equipartitioning in the n subshells of the ensemble: Pt(O') '" 8 [nt-n(O')]. If n indeed obeys closed equations, this assumption is safe. 2. Self-averaging of the n flow with resfect to the microscopic details of the non-nominated patterns: tt n -t (dt n)patt. Our equations (3) are hereby transformed into the closed set: ~n _ (EO' 8 [n-n(O')] Ei Wi(O') [n(FiO') - n(O')]) dt EO' 8[n-n(O')] patt The final observation is that the tool for averaging fractions is replica theory: dd n = lim lim ~ (~Wi(O'l) [n(FiO'1)-n(O'1)] rrn 8[n-n(O'O )])patt (4) t n--tO N --too ~ ~ O'I ···O' n i 0=1 The choice to be made for the observables n(O'), crucial for the closure assumptions to make sense, is constrained by requiring the theory to be exact in specific limits: exactness for a -t 0 : n = (m, ... ) exactness for t -t 00: n = (E, ... ) (for symmetric models only) 4 SIMPLE VERSION OF THE THEORY For the Hopfield model (2) the simplest two-parameter theory which is exact for a -t o and for t -t 00 is consequently obtained by choosing n = (m,E). Equivalently we can choose n = (m,r), where r(O') measures the 'interference energy': m = ~ L~I(7i i The result of working out (4) for n = (m, r) is: !m = J dz Dm,r[z] tanh,B (m+z) - m 1 d 1 J "2 dt r =; dz Dm,r[z]z tanh,B (m+z) + 1 - r 256 A. C. C. COOLEN, S. N. LAUGHTON, D. SHERRINGTON 15 ~----------------------------~ r / I / / / / o L-____________________________ ~ o m 1 Figure 1: Simulations (N = 32000, dots) versus simple RS theory (solid lines), for a = 0.1 and j3 = 00. Upper dashed line: upper boundary of the physical region. Lower dashed line: upper boundary of the RS region (the AT instability). in which Dm,r[z] is the distribution of 'interference-noise' terms in the PSP's, for which the replica calculation gives the outcome (in so-called RS ansatz): Dm,r[z] = e-~2 {l-jDY tanh [>.y [~] t+(~+Z) -~+{tl} 2 27rar apr apr + e-~)2 {1-jDY tanh [>.y [~] t +(~-Z)~-{tl} 2 27rar apr apr with Dy = [27rj-te- h2dy, ~ = apr->.2jp and>' = pyaq[l-p(l-q)]-l, and with the remaining parameters {q, {t, p} to be solved from the coupled equations: j j 1-p(1-q)2 m = Dy tanh[>'y+{tj q = Dy tanh2 [>.y+{t] r = [1-p(1-q)]2 Here we only give (partly new) results of the calculation; details can be found in (Coolen and Sherrington, 1994). The noise distribution is not Gaussian (in agreement with simulations, in contrast to LCH). Our simple two-parameter theory is found to be exact for t '" 0, t -7 00 and for a -7 O. Solving numerically the dynamic equations leads to the results shown in figures 1 and 2. We find a nice agreement with numerical simulations in terms of the flow in the (m, r) plane. However, for trajectories leading away from the recall state m '" 1, the theory fails to reproduce an overall slowing down. These deviations can be quantified by comparing cumulants of the noise distributions (Ozeki and Nishimori, 1994), or by applying the theory to exactly solvable models (Coolen and Franz, 1994). Other recent applications include spin-glass models (Coolen and Sherrington, 1994) and more general classes of attractor neural network models (Laughton and Coolen, 1995). The simple two-parameter theory always predicts adequately the location of the transients in the order parameter plane, but overestimates the relaxation speed. In fact, figure 2 shows a remarkable resemblance to the results obtained for this model in (Horner et al, 1989) with the functional integral formalism; the graphs of m(t) are almost identical, but here they are derived in a much simpler way. Modem Analytic Techniques to Solve the Dynamics of Recurrent Neural Networks 257 1 .8 10 2 ·6 --..., ..., ~ '-' ..... !-..... ..... ..... .4 ..... .... .... .... 5 -.... .... ........ .2 -----0 0 0 2 4 6 B 10 0 2 4 6 B 10 t t Figure 2: Simulations (N = 32000, dots) versus simple RS theory (RS stable: solid lines, RS unstable: dashed lines), now as functions of time, for Q; = 0.1 and f3 = 00. 5 ADVANCED VERSION OF THE THEORY Improving upon the simple theory means expanding the set n beyond n = (m,E). Adding a finite number of observables will only have a minor impact; a qualitative step forward, on the other hand, results from introducing a dynamic order parameter function. Since the microscopic dynamics (1) is formulated entirely in terms of neuron states and post-synaptic potentials we choose for n (u) the joint distribution: 1 D[(, h](u) = N L <5 [( -O"i] <5 [h-hi(U)] i This choice has the advantages that (a) both m and (for symmetric systems) E are integrals over D[(, h], so the advanced theory automatically inherits the exactness at t = 0 and t = 00 of the simple one, (b) it applies equally well to symmetric and nonsymmetric models and (c) as with the simple version, generalisation to models with continuous neural variables is straightforward. Here we show the result of applying the theory to a model of the type (1) with synaptic interactions: Jij = ~ ~i~j + .iN [cos(~ )Xij +sin(~ )Yij ] Xij = Xji, Yij = -Yji (independent random Gaussian variables) (describing a nominated pattern being stored on a 'messy' synaptic background). The parameter w controls the degree of synaptic symmetry (e.g. w = 0: symmetric, w = 7r: anti-symmetric). Equation (4) applied to the observable D[(, h](u) gives: 8 ~ 8 mDt[C h] = J2[1-(O"tanh(f3H))Dt] 8h2Dt[(,h] + 8h A [(,h;Dt] + :h {DdCh] [h-Jo(tanh(f3H ))Dt]} 1 1 +2 [l+(tanh(f3h)] Dd--(, h] - 2 [l-(tanh(f3h)] DdC h] 258 E A. C. C. COOLEN, S. N. LAUGHfON, D. SHERRINGTON o .------,------.------.------.------.------~ - .2 -.4 -.6 "- "- .8 '~ ~---------- -_ 1 L-____ -L ______ L-____ ~ ______ ~ ____ ~ ______ ~ o 2 4 6 t Figure 3: Comparison of simulations (N = 8000, solid line), simple two-parameter theory (RS stable: dotted line, RS unstable: dashed line) and advanced theory (solid line), for the w = a (symmetric background) model, with Jo = 0, f3 = 00. Note that the two solid lines are almost on top of each other at the scale shown. ". 0.5 0.0 E -0.5 -0.5 o 2 4 6 o 2 4 6 t t Figure 4: Advanced theory versus N = 5600 simulations in the w = ~7r (asymmetric background) model, with f3 = 00 and J = 1. Solid: simulations; dotted: solving the RS diffusion equation. Modem Analytic Techniques to Solve the Dynamics of Recurrent Neural Networks 259 with (f(a,H))D = L:". JdH D[a,H]J(a, H). All complications are concentrated in the kernel A[C h; DJ, which is to be solved from a nontrivial set of equations emerging from the replica formalism. Some results of solving these equations numerically are shown in figures 3 and 4 (for details of the calculations and more elaborate comparisons with simulations we refer to (Laughton, Coolen and Sherrington, 1995; Coolen, Laughton and Sherrington, 1995)). It is clear that the advanced theory quite convincingly describes the transients of the simulation experiments, including the hitherto unexplained slowing down, for symmetric and nonsymmetric models. 6 DISCUSSION In this paper we have described novel techniques for studying the dynamics of recurrent neural networks near saturation. The simplest two-parameter theory (exact for t = 0, for t --+ 00 and for 0: --+ 0), which employs as dynamic order parameters the overlap with a pattern to be recalled and the total 'energy' per neuron, already describes quite accurately the location of the transients in the order parameter plane. The price paid for simplicity is that it overestimates the relaxation speed. A more advanced version of the theory, which describes the evolution of the joint distribution for neuron states and post-synaptic potentials, is mathematically more involved, but predicts the dynamical data essentially perfectly, as far as present applications allow us conclude. Whether this latter version is either exact, or just a very good approximation, still remains to be seen. In this paper we have restricted ourselves to models with binary neural variables, for reasons of simplicity. The theories generalise in a natural way to models with analogue neurons (here, however, already the simple version will generally involve order parameter functions as opposed to a finite number of order parameters). Ongoing work along these lines includes, for instance, the analysis of analogue and spherical attractor networks and networks of coupled oscillators near saturation. References B. Derrida, E. Gardner and A. Zippelius (1987), Europhys. Lett. 4: 167-173 A.C.C. Coolen and D. Sherrington (1993), in J.G. Taylor (ed.), Mathematical Approaches to Neural Networks, 293-305. Amsterdam: Elsevier. S. Amari and K. Maginu (1988), Neural Networks 1: 63-73 H. Nishimori and T. Ozeki (1993), J. Phys. A 26: 859-871 H. Rieger, M. Schreckenberg and J. Zittartz (1988), Z. Phys. B 72: 523-533 H. Horner, D. Bormann, M. Frick, H. Kinzelbach and A. Schmidt (1989), Z. Phys. B 76: 381-398 A.C.C. Coolen and D. Sherrington (1994), Phys. Rev. E 49(3): 1921-1934 H. Nishimori and T. Ozeki (1994), J. Phys. A 27: 7061-7068 A.C.C. Coolen and S. Franz (1994), J. Phys. A 27: 6947-9954 A.C.C. Coolen and D. Sherrington (1994), J. Phys. A 27: 7687-7707 S.N. Laughton and A.C.C. Coolen (1995), Phys. Rev. E 51: 2581-2599 S.N. Laughton, A.C.C. Coolen and D. Sherrington (1995), J. Phys. A (in press) A.C.C. Coolen, S.N. Laughton and D. Sherrington (1995), Phys. Rev. B (in press)
|
1995
|
62
|
1,109
|
A Realizable Learning Task which Exhibits Overfitting Siegfried Bos Laboratory for Information Representation, RIKEN, Hirosawa 2-1, Wako-shi, Saitama, 351-01, Japan email: boes@zoo.riken.go.jp Abstract In this paper we examine a perceptron learning task. The task is realizable since it is provided by another perceptron with identical architecture. Both perceptrons have nonlinear sigmoid output functions. The gain of the output function determines the level of nonlinearity of the learning task. It is observed that a high level of nonlinearity leads to overfitting. We give an explanation for this rather surprising observation and develop a method to avoid the overfitting. This method has two possible interpretations, one is learning with noise, the other cross-validated early stopping. 1 Learning Rules from Examples The property which makes feedforward neural nets interesting for many practical applications is their ability to approximate functions, which are given only by examples. Feed-forward networks with at least one hidden layer of nonlinear units are able to approximate each continuous function on a N-dimensional hypercube arbitrarily well. While the existence of neural function approximators is already established, there is still a lack of knowledge about their practical realizations. Also major problems, which complicate a good realization, like overfitting, need a better understanding. In this work we study overfitting in a one-layer percept ron model. The model allows a good theoretical description while it exhibits already a qualitatively similar behavior as the multilayer perceptron. A one-layer perceptron has N input units and one output unit. Between input and output it has one layer of adjustable weights Wi, (i = 1, ... ,N). The output z is a possibly nonlinear function of the weighted sum of inputs Xi, i.e. z = g(h) , with 1 N h = I1tT L Wi Xi . vN i=l (1) A Realizable Learning Task Which Exhibits Overfitting 219 The quality of the function approximation is measured by the difference between the correct output z* and the net's output z averaged over all possible inputs. In the supervised learning scheme one trains the network using a set of examples ;fll (JL = 1, ... , P), for which the correct output is known. It is the learning task to minimize a certain cost function, which measures the difference between the correct output z~ and the net's output Zll averaged over all examples. Using the mean squared error as a suitable measure for the difference between the outputs, we can define the training error ET and the generalization error Ea as (2) The development of both errors as a function of the number P of trained examples is given by the learning curves. Training is conventionally done by gradient descend. For theoretical purposes it is very useful to study learning tasks, which are provided by a second network, the so-called teacher network. This concept allows a more transparent definition of the difficulty of the learning task. Also the monitoring of the training process becomes clearer, since it is always possible to compare the student network and the teacher network directly. Suitable quantities for such a comparison are, in the perceptron case, the following order parameters, N q:= IIWII = 2:(Wi )2. (3) i=l Both have a very transparent interpretation, r is the normalized overlap between the weight vectors of teacher and student, and q is the norm of the student's weight vector. These order parameters can also be used in multilayer learning, but their number increases with the number of all possible permutations between the hidden units of teacher and student. 2 The Learning Task Here we concentrate on the case in which a student perceptron has to learn a mapping provided by another perceptron. We choose identical networks for teacher and student. Both have the same sigmoid output function, i.e. g*(h) = g(h) = tanh( "Ih). Identical network architectures of teacher and student are realizable tasks. In principle the student is able to learn the task provided by the teacher exactly. Unrealizable tasks can not be learnt exactly, there remains always a finite error. If we use uniformally distributed random inputs ;f and weights W, the weighted sum h in (1) can be assumed as Gaussian distributed. Then we can express the generalization error (2) by the order parameters (3), Ea= JDZ1 JDz2~{tanh["IZll-tanh[q(rzl+~Z2)]r, (4) with the Gaussian measure J 1 +00 dz (Z2) Dz:= -- exp -- 00 ../2i 2 (5) From equation (4) we can see how the student learns the gain "I of the teachers output function. It adjusts the norm q of its weights. The gain "I plays an important role since it allows to tune the function tanhbh) between a linear function b « 1) and a highly nonlinear function b » 1). Now we want to determine the learning curves of this task. 220 s.B6s 3 Emergence of Overfitting 3.1 Explicit Expression for the Weights Below the storage capacity of the perceptron, i.e. a = 1, the minimum of the training error ET is zero. A zero training error implies that every example has been learnt exactly, thus (6) The weights with minimal norm that fulfill this condition are given by the Pseudoinverse (see Hertz et al. 1991), P Wi = 2: h~ (C-l)~v xf, (7) ~,v=l Note, that the weights are completely independent of the output function g(h) = g*(h). They are the same as in the simplest realizable case, linear perceptron learns linear perceptron. 3.2 Statistical Mechanics The calculation of the order parameters can be done by a method from statistical mechanics which applies the commonly used replica method. For details about the replica approach see Hertz et al. (1991). The solution of the continuous perceptron problem can be found in Bas et al. (1993). Since the results of the statistical mechanics calculations are exact only in the thermodynamic limit, i.e. N ~ 00, the variable a is the more natural measure. It is defined as the fraction of the number of patterns P over the system size N, i.e. a := PIN. In the thermodynamic limit N and P are infinite, but a is still finite. Normally, reasonable system sizes, such as N ~ 100, are already well described by this theory. Usually one concentrates on the zero temperature limit, because this implies that the training error ET accepts its absolute minimum for every number of presented examples P. The corresponding order parameters for the case, linear perceptron learns linear student, are q='Yva, r =va. (8) The zero temperature limit can also be called exhaustive training, since the student net is trained until the absolute minimum of ET is reached. For small a and high gains 'Y, i.e levels of nonlinearity, exhaustive training leads to overfitting. That means the generalization error Ea(a) is not, as it should, monotonously decreasing with a. It is one reason for overfitting, that the training follows too strongly the examples. The critical gain 'Yc, which determines whether the generalization error Ea ( a) is increasing or decreasing function for small values of a, can be determined by a linear approximation. For small a, both order parameters (3) are small, and the student's tanh-function in (4) can be approximated by a linear function. This simplifies the equation (4) to the following expression, Ea(f) = Ea(O) - i [2H(r) - 'Y 1, with H( 'Y):= J Dz tanh(rz) z. (9) Since the function H(r) has an upper bound, i.e. J2/7r, the critical gain is reached if 'Yc = 2H{rc). The numerical solution gives 'Yc = 1.3371. If 'Y is higher, the slope of Ea(a) is positive for small a. In the following considerations we will use always the gain 'Y = 5 as an example, since this is an intermediate level of nonlinearity. A Realizable Learning Task Which Exhibits Overfitting 1.0 0.8 0.6 0.4 0.2 ...... -- '- .- . -'-.- --.---.- .-. -'-. - '- ' - . -.-. -'-. w _ _ • _ .. __ -- -- -.. -- -.. -- -- ----- .. _- --0.0 0.0 0.2 0.4 0.6 PIN -'-.100.0 10.0 5.0 2.0 1.0 0.5 0.8 221 1.0 Figure 1: Learning curves E ( 0:) for the problem, tanh- perceptron learns tanhperceptron, for different values of the gain,. Even in this realizable case, exhaustive training can lead to overfitting, if the gain , is high enough. 3.3 How to Understand the Emergence of Overfitting Here the evaluation of the generalization error in dependence of the order parameters rand q is helpful. Fig. 2 shows the function EG(r, q) for r between 0 and 1 and q between 0 and 1.2,. The exhaustive training in realizable cases follows always the line q( r) = ,r independent of the actual output function. That means, training is guided only by the training error and not by the generalization error. If the gain , is higher than ,e, the line EG = EG(O, 0) starts with a lower slope than q(r) = ,r, which results in overfitting. 4 How to Avoid Overfitting From Fig. 2 we can guess already that q increases too fast compared to r. Maybe the ratio between q and r is better during the training process. So we have to develop a description for the training process first. 4.1 Training Process We found already that the order parameters for finite temperatures (T > 0) of the statistical mechanics approach are a good description of the training process in an unrealizable learning task (Bos 1995). So we use the finite temperature order parameters also in this task. These are, again taken from the task 'linear perceptron learns linear percept ron' , ( ) = J(~) (1 + 0:) a - 20: q 0:, a, 2 ' a a - 0: r(o:, a) = (0:) a2 0: a (1+0:)a-20:' (10) with the temperature dependent variable a:= 1 + [,8(Q - q)]-l . (11) 222 6.0 5.0 4.0 q 3.0 2.0 ".1.0 . . . . /local minZ i abs. mici. ./ local m~. ........ ... ... ........ ........ S.BOS --- ...... ~: ...... -. 0.0 ~~==~-~-:=::: .... !:. ==±~===--L..::===~·= ··· ···~ ·· ·· ·3· ·· ·· · 0.0 0.2 0.4 0.6 0.8 1.0 r Figure 2: Contour plot of EG(r,q) defined by (4), the generalization error as a function of the two order parameters. Starting from the minimum EG = 0 at (r, q) = (1,5) the contour lines for EG = 0.1,0.2, ... , 0.8 are given (dotted lines). The dashed line corresponds to EG(O,O) = 0.42. The solid lines are parametric curves of the order parameters (r, q) for certain training strategies. The straight line illustrates exhaustive training, the lower ones the optimal training, which will be explained in Fig. 3. Here the gain I = 5. The zero temperature limit corresponds to a = 1. We will show now that the decrease of the temperature dependent parameter a from 00 to 1, describes the evolution of the order parameters during the training process. In the training process the natural parameter is the number of parallel training steps t. In each parallel training step all patterns are presented once and all weights are updated. Fig. 3 shows the evolution of the order parameters (10) as parametric curves (r,q). The exhaustive learning curve is defined by a = 1 with the parameter 0: (solid line). For each 0: the training ends on this curve. The dotted lines illustrate the training process, a runs from infinity to 1. Simulations of the training process have shown that this theoretical curve is a good description, at least after some training steps. We will now use this description of the training process for the definition of an optimized training strategy. 4.2 Optimal temperature The optimized training strategy chooses not a = 1 or the corresponding temperature T = 0, but the value of a (Le. temperature), which minimizes the generalization error EG. In the lower solid curve indicating the parametric curve (r, q) the value of a is chosen for every 0:, which minimizes EG. The function EG(a) has two minima between 0: = 0.5 and 0.7. The solid line indicates always the absolute minimum. The parametric curves corresponding to the local minima are given by the double dashed and dash-dotted lines. Note, that the optimized value a is always related to an optimized temperature through equation (11). But the parameter a is also related to the number of training steps t. A Realizable Learning Task Which Exhibits Overfilling 6.0 5.0 4.0 q 3.0 2.0 1.0 0.0 0.0 local min. abs. min. local min. simulation I--t--l 0.2 0.4 r 223 0.6 0.8 1.0 Figure 3: Training process. The order parameters (10) as parametric curves (r,q) with the parameters a and a. The straight solid line corresponds to exhaustive learning, i.e. a = 1 (marks at a = 0.1,0.2, ... 1.0). The dotted lines describe the training process for fixed a. Iterative training reduces the parameter a from 00 to 1. Examples for a = 0.1,0.2,0.3,0.4,0.9,0.99 are given. The lower solid line is an optimized learning curve. To achieve this curve the value of a is chosen, which minimizes EG absolutely. Between a ~ 0.5 and 0.7 the error EG has two minima; the double- dashed and dash-dotted lines indicate the second, local minimum of EG. Compare with Fig. 2, to see which is the absolute and which the local minimum of EG. A naive early stopping procedure ends always in the minimum with the smaller q, since it is the first minimum during the training process (see simulation indicated with errorbars). 4.3 Early Stopping Fig. 3 and Fig. 2 together indicate that an earlier stopping of the training process can avoid the overfitting. But in order to determine the stopping point one has to know the actual generalization error during the training. Cross-validation tries to provide an approximation for the real generalization error. The cross-validation error Ecv is defined like ET , see (2), on a set of examples, which are not used during the training. Here we calculate the optimum using the real generalization error, given by rand q, to determine the optimal point for early stopping. It is a lower bound for training with finite cross-validation sets. Some preliminary tests have shown that already small cross- validation sets approximate the real EG quite well. Training is stopped, when EG increases. The resulting curve is given by the error bars in Fig. 3. The errorbars indicate the standard deviation of a simulation with N = 100 averaged over 50 trials. In Fig. 4 the same results are shown as learning curves EG(a). There one can see clearly that the early stopping strategy avoids the overfitting. 5 Summary and outlook In this paper we have shown that overfitting can also emerge in realizable learning tasks. The calculation of a critical gain and the contour lines in Fig. 2 imply, that 224 0.5 0.4 0.3 EO 0.2 0.1 0.0 0.0 exh. local min. abs. min. local min. simulation ~ 0.2 S.BOS 0.4 0.6 0.8 1.0 PIN Figure 4: Learning curves corresponding to the parametric curves in Fig. 3. The upper solid line shows again exhaustive training. The optimized finite temperature curve is the lower solid line. From 0: = 0.6 exhaustive and optimal training lead to identical results (see marks). The simulation for early stopping (errorbars) finds the first minimum of EG. the reason for the overfitting is the nonlinearity of the problem. The network adjusts slowly to the nonlinearity of the task. We have developed a method to avoid the overfitting, it can be interpreted in two ways. Training at a finite temperature reduces overfitting. It can be realized, if one trains with noisy examples. In the other interpretation one learns without noise, but stops the training earlier. The early stopping is guided by cross-validation. It was observed that early stopping is not completely simple, since it can lead to a local minimum of the generalization error. One should be aware of this possibility, before one applies early stopping. Since multilayer perceptrons are built of nonlinear perceptrons, the same effects are important for multilayer learning. A study with large scale simulations (Miiller et al. 1995) has shown that overfitting occurs also in realizable multilayer learning tasks. Acknowledgments I would like to thank S. Amari and M. Opper for stimulating discussions, and M. Herrmann for hints concerning the presentation. References S. Bos. (1995) Avoiding overfitting by finite temperature learning and crossvalidation. International Conference on Artificial Neural Networks '95 Vo1.2, p.111. S. Bos, W. Kinzel & M. Opper. (1993) Generalization ability of perceptrons with continuous outputs. Phys. Rev. E 47:1384-1391. J. Hertz, A. Krogh & R. G. Palmer. (1991) Introduction to the Theory of Neural Computation. Reading: Addison-Wesley. K. R. Miiller, M. Finke, N. Murata, K. Schulten & S. Amari. (1995) On large scale simulations for learning curves, Neural Computation in press.
|
1995
|
63
|
1,110
|
Simulation of a Thalamocortical Circuit for Computing Directional Heading in the Rat Hugh T. Blair* Department of Psychology Yale University New Haven, CT 06520-8205 tadb@minerva.cis.yale.edu Abstract Several regions of the rat brain contain neurons known as head-direction celis, which encode the animal's directional heading during spatial navigation. This paper presents a biophysical model of head-direction cell acti vity, which suggests that a thalamocortical circuit might compute the rat's head direction by integrating the angular velocity of the head over time. The model was implemented using the neural simulator NEURON, and makes testable predictions about the structure and function of the rat head-direction circuit. 1 HEAD-DIRECTION CELLS As a rat navigates through space, neurons called head-direction celis encode the animal's directional heading in the horizontal plane (Ranck, 1984; Taube, Muller, & Ranck, 1990). Head-direction cells have been recorded in several brain areas, including the postsubiculum (Ranck, 1984) and anterior thalamus (Taube, 1995). A variety of theories have proposed that head-direction cells might play an important role in spatial learning and navigation (Brown & Sharp, 1995; Burgess, Recce, & O'Keefe, 1994; McNaughton, Knierim, & Wilson, 1995; Wan, Touretzky, & Redish, 1994; Zhang, 1995). 1.1 BASIC FIRING PROPERTIES A head-direction cell fires action potentials only when the rat's head is facing in a particular direction with respect to the static surrounding environment, regardless of the animal's location within that environment. Head-direction cells are not influenced by the position of the rat's head with respect to its body, they are only influenced by the direction of the *Also at the Yale Neuroengineering and Neuroscience Center (NNC), 5 Science Park North, New Haven, CT 06511 Simulation of Thalamocortical Circuit for Computing Directional Heading in Rats 153 G> ~ g> , 0 ~ 05 )( '" E ;! o 0 0~~90~--:-::180:-"--::27:70---~360 Head Direction 360,0 270 90 180 Figure I: Directional Tuning Curve of a Head-Direction Cell head with respect to the stationary reference frame of the spatial environment. Each headdirection cell has its own directional preference, so that together, the entire population of cells can encode any direction that the animal is facing. Figure 1 shows an example of a head-direction cell's directional tuning curve, which plots the firing rate of the celI as a function of the rat's momentary head direction. The tuning curve shows that this cell fires maximalIy when the rat's head is facing in a preferred direction of about 160 degrees. The cell fires less rapidly for directions close to 160 degrees, and stops firing altogether for directions that are far from 160 degrees. 1.2 THE VELOCITY INTEGRATION HYPOTHESIS McNaughton, Chen, & Markus (1991) have proposed that head-direction cells might rely on a process of dead-reckoning to calculate the rat's current head direction, based on the previous head direction and the angular velocity at which the head is turning. That is, head-direction cells might compute the directional position of the head by integrating the angular velocity of the head over time. This velocity integration hypothesis is supported by three experimental findings. First, several brain regions that are associated with headdirection cells contain angular velocity cells, neurons that fire in proportion to the angular head velocity (McNaughton et al., 1994; Sharp, in press). Second, some head-direction cells in postsubiculum are modulated by angular head velocity, such that their peak firing rate is higher if the head is turning in one direction than in the other (Taube et al., 1990). Third, it has recently been found that head-direction cells in the anterior thalamus, but not the postsubiculum, anticipate the future direction of the rat's head (Blair & Sharp, 1995). 1.3 ANTICIPATORY HEAD-DIRECTION CELLS Blair and Sharp (1995) discovered that head-direction cells in the anterior thalamus shift their directional preference to the left during clockwise turns, and to the right during counterclockwise turns. They showed that this shift occurs systematically as a function of head velocity, in a way that alIows these cells anticipate the future direction of the rat's head. To illustrate this, consider a cell that fires whenever the head will be facing a specific direction, 9, in the near future. How would such a cell behave? There are three cases to consider. First, imagine that the rat's head is turning clockwise, approaching the direction 9 from the left side. In this case, the anticipatory cell must fire when the head is facing to the left of 9, because being to the left of 9 and turning clockwise predicts arrival at 9 in the near future. Second, when the head is turning counterclockwise and approaching 9 from the right side, the anticipatory cell must fire when the head is to the right of 9. Third, if the head is still, then the cell should only fire if the head is presently facing 9. In summary, an anticipatory head direction cell should shift its directional preference to the left during clockwise turns, to the right during counterclockwise turns, and not at all when the head is still. This behavior can be formalized by the equation !leV) = 9 - V't, [1] 154 H. T. BLAIR where ~ denotes the cell's preferred present head direction. v denotes the angular velocity of the head. 8 denotes the future head direction that the cell anticipates, and 't is a constant time delay by which the cell's activity anticipates arrival at 8. Equation 1 assumes that ~ is measured in degrees. which increase in the clockwise direction. and that v is positive for clockwise head turns. and negative for counterclockwise head turns. Blair & Sharp (1995) have demonstrated that Equation 1 provides a good approximation of head-direction cell behavior in the anterior thalamus. 1.3 ANTICIPATORY TIME DELAY (r) Initial reports suggested that head-direction cells in the anterior thalamus anticipate the future head direction by an average time delay of't = 40 msec, whereas postsubicular cells encode the present head direction, and therefore "anticipate" by 't = 0 msec (Blair & Sharp, 1995; Taube & Muller, 1995). However, recent evidence suggests that individual neurons in the anterior thalamus may be temporally tuned to anticipate the rat's future head-direction by different time delays between 0-100 msec, and that postsubicular cells may "lag behind" the present head-direction by about to msec (Blair & Sharp, 1996). 2 A BIOPHYSICAL MODEL This section describes a biophysical model that accounts for the properties of head-direction cells in postsubiculum and anterior thalamus. by proposing that they might be connected to form a thalamocortical circuit. The next section presents simulation results from an implementation of the model, using the neural simulator NEURON (Hines, 1993). 2.1 NEURAL ELEMENTS Figure 2 illustrates a basic circuit for computing the rat's head-direction. The circuit consists of five types of cells: 1) Present Head-Direction (PHD) Cells encode the present direction of the rat's head, 2) Anticipatory Head-Direction (AHD) Cells encode the future direction of the rat's head, 3) Angular-Velocity (AV) Cells encode the angular velocity of the rat's head (the CLK AV Cell is active during clockwise turns, and the CNT AV Cell is active during counterclockwise turns), 4) the Angular Speed (AS) Cell fires in inverse proportion to the angular speed of the head, regardless of the turning direction (that is, the AS Cell fires at a lower rate during fast turns, and at a higher rate during slow turns), 5) Angular-Velocity Modulated Head-Direction (AVHD) Cells are head-direction cells that fire AHDCells RTN A~~~IS Excitatory ~ Inhibitory --~,;;",I".· AS Cell MB I ABBREVIADONS AT = Anterior Thalamus MB. Mammillary Bodi .. PS = P08tsubiculum RS • Rempl"'ill Cortex R1N = Reticul.11III. Nu. Figure 2: A Model of the Rat Head-Direction System Simulation of Thalamocortical Circuit for Computing Directional Heading in Rats 155 only when the head is turning in one direction and not the other (the CLK AVHD Cell fires in its preferred direction only when the head is turning clockwise, and the CNT AVHD Cell fires in its preferred direction only when the head turns counterclockwise). 2.2 FUNCTIONAL CHARACTERISTICS In the model, AHD Cells directly excite their neighbors on either side, but indirectly inhibit these same neighbors via the AVHD Cells, which act as inhibitory interneurons. AHD Cells also send excitatory feedback connections to themselves (omitted from Figure 2 for clarity), so that once they become active. they remain active until they are turned off by inhibitory input (the rate of firing can also be modulated by inhibitory input). When the rat is not turning its head. the cell representing the current head direction fires constantly, both exciting and inhibiting its neighbors. In the steady-state condition (Le., when the rat is not turning its head), lateral inhibition exceeds lateral excitation, and therefore activity does not spread in either direction through the layer of AHD Cells. However. when the rat begins turning its head, some of the AVHD Cells are turned off, allowing activity to spread in one direction. For example. during a clockwise head tum. the CLK AV Cell becomes active, and inhibits the layer of CNT AVHD Cells. As a result, AHD Cells stop inhibiting their right neighbors, so activity spreads to the right through the layer of AHD Cells. Because AHD Cells continue to inhibit their neighbors to the left, activity is shut down in the leftward direction, in the wake of the activity spreading to the right. The speed of propagation through the AHD layer is governed by the AS Cell. During slow head turns, the AS Cell fires at a high rate, strongly inhibiting the AHD Cells, and thereby slowing the speed of propagation. During fast head turns, the AS Cell fires at a low rate, weakly inhibiting the AHD Cells, allowing activity to propagate more quickly. Because of inhibition from AS cells, AHD cells fire faster when the head is turning than when it is still (see Figure 4), in agreement with experimental data (Blair & Sharp, 1995). AHD Cells send a topographic projection to PHD Cells, such that each PHD Cell receives excitatory input from an AHD Cell that anticipates when the head will soon be facing in the PHD Cell's preferred direction. AHD Cell activity anticipates PHD Cell activity because there is a transmission delay between the AHD and PHD Cells (assumed to be 5 msec in the simulations presented below). Also, the weights of the connections from AHD Cells to PHD Cells are small, so each AHD Cell must fire several action potentials before its targeted PHD Cell can begin to fire. The time delay between AHD and PHD Cells accounts for anticipatory firing, and corresponds to the 1: parameter in Equation I. 2.3 ANATOMICAL CHARACTERISTICS Each component of the model is assumed to reside in a specific brain region. AHD and PHD Cells are assumed to reside in anterior thalamus (AT) and postsubiculum (PS), respectively. AS Cells have been observed in PS (Sharp, in press) and retrosplenial cortex (RS) (McNaughton, Green, & Mizumori, 1986), but the model predicts that they may also be found in the mammillary bodies (MB), since MB receives input from PS and RS (Shibata, 1989), and MB projects to ATN. AVHD Cells have been observed in PS (Taube et ai., 1990), but the model predicts that they may aiso be found in the reticular thalamic nucleus (RTN), because RTN receives input from PSIRS (Lozsadi, 1994), and RTN inhibits AT. It should be noted that lateral excitation between ATN cells has not been shown, so this feature of the model may be incorrect. Table 1 summarizes anatomical evidence. 3 SIMULATION RESULTS The model illustrated in Figure 2 has been implemented using the neural simulator NEURON (Hines. 1993). Each neural element was represented as a single spherical compart156 H. T. BLAIR Table 1: Anatomical Features of the Model FEATURE OF MODEL PHD Cells in PSIRS AHD Cells in AT AV Cells in PSIRS AT projects to PS AT projects to RTN PSIRS projects to RTN AVHD Cells in RTN AS Cells in MB REFERENCE Chen et aI., 1990; Ranck, 1984 Blair & Sharp, 1995 McNaughton et aI., 1994; Sharp, in press van Groen & Wyss, 1990 Shibata, 1992 Lozsadi, 1994 PREDICTION OF MODEL PREDICTION OF MODEL ment, 30 Jlm in diameter, with RC time constants ranging between 15 and 30 msec. Synaptic connections were simulated using triggered alpha-function conductances. The results presented here demonstrate the behavior of the model, and compare the properties of the model with experimental data. To begin each simulation, a small current was injected in to one of the AHD Cells, causing it to initiate sustained firing. This cell represented the simulated rat's initial head direction. Head-turning behavior was simulated by injecting current into the AV and AS Cells, with an amplitude that yielded firing proportional to the desired angular head velocity. 3.1 ACTIVITY OF HEAD-DIRECTION CELLS Figure 3 presents a simple simulation, which illustrates the behavior of head-direction cells in the model. The simulated rat begins by facing in the direction of 0 degrees. Over the course of 250 msec, the rat quickly turns its head 60 degrees to the right, and then returns to the initial starting position of 0 degrees. The average velocity of the head in this simulation was 480 degrees/sec, which is similar to the speed at which an actual rat performs a fast head tum (Blair & Sharp, 1995). Over the course of the simulation, neural activation propagates from the O-degree cell to the 60-degree cell, and then back to the 0degree cell. 3.2 COMPARISON WITH EXPERIMENTAL DATA To examine how well the model reproduces firing properties of PS and AT cells, another simple simulation was performed. The firing rate the model's PHD and AHD Cells was examined while the simulated rat performed several 360-degree revolutions in both the clockwise and counterclockwise directions. Results are summarized in Figure 4, which ACTIVITY OF PHD CELLS ANIMAL 15c.lll_WlM"-__ -----"'~~ 30CelII----MMIM'-_--M\.W~ oW CelII----~~ ___ -M~~ 60c.III=:::;~~~~~:::;::==::, 50 100 1 Turning Right A'--;Tu....,mIng~Le~1t ---:, Time (msec) , BEHAVIOR WSOIIIt ••.......... ". AVlrlgl Angular Velocity = 480" 'sec Figure 3: Simulation Example Simulation of Thalamocortical Circuit for Computing Directional Heading in Rats 157 Cil12.0 ----------., ~ 10.0 . Z' ~ 8.0 ' CD 6.0 I 15» .:i 40 : g 2.0 i ,; 0.0 i O<>"T (exper.,..ntaI Getal 0-0 Ps (.)(~'"*"". ,*a) .n _AT (modol dolo' .••••• .. Ps (model ~. ) ••••• [} ...•..........•.•....• -o • • io .2.0 ,'--_ ___ _ _ -----' N 0 100 200 300 400 500 Angular Head Velocity (deglsec) N'25.0 r-, --~----e. ~ ~ 20.0 ~ o.-/' ..... ·····~ 0)15.0 , .= Li: 10.0 , • g, I Cii 5.0 , ~ 0.0 i~ _______ -' o 100 200 300 400 500 [}----------------.----.-o Angular Head Velocity (degJsec) Figure 4: Compared Properties of Real and Simulated Head-Direction Cells compares simulation data with experimental data. The experimental data in Figure 4 shows averaged results for 21 cells recorded in AT, and 19 cells recorded in PS. Because AT cells anticipate the future head direction, they exhibit an angular separation between their clockwise and counterclockwise directiQnal preference, whereas as no such separation occurs for PS cells (see section 2.4). For AT cells, the magnitude of the angular separation is proportional to angular head velocity, with greater separation occurring for fast turns, and less separation for slow turns (see Eq. 1). The left panel of Figure 4 shows that the model's PHD and AHD Cells exhibit a similar pattern of angular separation. Blair & Sharp (1995) reported that the firing rates of AT and PS cells differ in two ways: 1) AT cells fire at a higher rate than PS cells, and 2) AT cells have a higher rate during fast turns than during slow turns, whereas PS cells fire at the same rate, regardless of turning speed. In Figure 4 (right panel), it can be seen that the model reproduces these findings. 4 DISCUSSION AND CONCLUSIONS In this paper, I have presented a neural model of the rat head-direction system. The model includes neural elements whose firing properties are similar to those of actual neurons in the rat brain. The model suggests that a thalamocortical circuit might compute the directional position of the rat's head, by integrating angular head velocity over time. 4.1 COMPARISON WITH OTHER MODELS McNaughton et al. (1991) proposed that neurons encoding head-direction and angular velocity might be connected to form a linear associative mapping network. Skaggs et al. (1995) have refined this idea into a theoretical circuit, which incorporates head-direction and angular velocity cells. However, the Skaggs et al. (1995) circuit does not incorporate anticipatory head-direction cells, like those found in AT. A model that does incorporate anticipatory cells has been developed by Elga, Redish, & Touretzky (unpublished manuscript). Zhang (1995) has recently presented a theoretical analysis of the head-direction circuit, which suggests that anticipatory head-direction cells might be influenced by both the angular velocity and angular acceleration of the head, whereas non-anticipatory cells may be influenced by the angular velocity only, and not the angular acceleration. 4.2 LIMITATIONS OF THE MODEL In its current form, the model suffers some significant limitations. For example, the directional tuning curves of the model's head-direction cells are much narrower than those of actual head-direction cells. Also, in its present form, the model can accurately track the rat's head-direction over a rather limited range of angular head velocities. These limitations are presently being addressed in a more advanced version of the model. 158 H. T. BLAIR Acknowledgments This work was supported by NRSA fellowship number 1 F31 MH11102-01Al from NIMH. a Yale Fellowship. and the Yale Neuroengineering and Neuroscience Center (NNC). I thank Michael Hines. Patricia Sharp. and Steve Fisher for their assistance. References Blair. H.T .. & Sharp. P.E. (1995). Anticipatory head-direction cells in anterior thalamus: Evidence for a thalamocortical circuit that mtegrates angular head velocity to compute head direction. Journal of Neuroscience, IS, 6260-6270. Blair, H.T .• & Sharp (1996). Temporal Tuning of Anticipatory Head-Direction Cells in the Anterior Thalamus of the Rat. Submitted. Brown. M. & Sharp. P.E. (1995). Simulation of spatial learning in the morris water maze by a. neural network model of the hippocampal formation and nucleus accumbens. Hippocampus, 5. 171-188. Burgess, N .• Recce. M .• & O'Keefe. J. (1994). A model of hippocampal function. Neural Networks, 7. 1065-1081. Elga. AN .• Redish, AD .• & Touretzky. D.S. (1995). A model of the rodent head-direction system. Unyublished Manuscript. Hines. M. (1993). NEURON: A program for simulation of nerve equations. In F. Eckman (Ed.). Neural Systems: Analysis and Modeling, Norwell. MA : Kluwer Academic Publishers. pp. 127-136. Lozsadi. D.A. (1994). Organization of cortical afferents to the rostral, limbic sector of the rat thalamic reticular nucleus. The Journal of Comparative Neurology, 341, 520-533. McNaughton. B.L.. Chen. L.L.. & Markus. E.1. (1991). Dead reckoning, landmark learning. and the sense of direction: a neurophysiological and computational hypothesis. Journal of Cognitive Neuroscience, 3, 190-202. McNaughton, B.L.. Green. E.1 .• & Mizumori, S.1.y. (1986). Representation of body motion trajectory by rat sensory motor cortex neurons. Society for Neuroscience Abstracts. 12,260. McNaughton, B.L.. Knierim. J.J .• & Wilson. M.A (1995). Vector encoding and the vestibular foundations of spatial cognition: neurophysiological and computational mechanisms. In M. Gazzaniga (Ed.). The Cognitive Neurosciences. Cambndge: MIT Press. McNaughton. B.L.. Mizumori, S.Y.1 .• Barnes. C.A .• Leonard. B.J .• MarqUiS. M .• & Green. B.J. (1994). Coritcal representation of motion during unrestrained spatial navigaton in the rat. Cerebral Cortex, 4, 27-39. Ranck, J.B. (1984). Head-direction cells in the deep ceUlayers of dorsal presubiculum in freely moving rats. Society for Neuroscience Abstracts, 12, 1524. Shibata. H. (1989). Descending projections to the mammillary nuclei in the rat. as studied by retrograde and anterograde transport of wheat germ agglutinin-horseradish peroxidase. The Journal of Comparative Neurology, 285. 436-452. Shibata. H. (1992). Topographic organization of subcortical projections to the anterior thalamic nuclei in the rat. The Journal of Comparative Neurology, 323, 117-127. Sharp, P.E. (in press). Multiple spatiallbehavioral corrrelates for cells in the rat postsubiculum: multiple regression analysis and comparison to other hippocampal areas. Cerebral Cortex. Skaggs, W.E .• Knierim. J.1 .• Kudrimoti. H.S., & McNaughton, B.L. (1995). A model of the neural basis of the rat's sense of direction. In G. Tesauro. D.S. Touretzky, & T.K. Leen (Eds.), Advances in Neural Information Processing Systems 7. MIT Press. Taube. 1.S. (1995). Head-direction cells recorded in the anterior thalamic nuclei of freelymoving rats. Journal of Neuroscience, 15, 70-86. Taube. J.S .• & Muller. R.O. (1995). Head-direction cell activity in the anterior thalamus. but not the postsubiculum, predicts the animal's future directional heading. Society for Neuroscience Abstracts. 21. 946. Taube. J.S., Muller. R.U .• & Ranck, J.B. (1990). Head-direction cells recorded from the postsubiculum in freely moving rats, I. Description and quantitative analysis. Jounral of Neuroscience, 10, 420-435. van Groen. T., & Wyss, J.M. (1990). The postsubicular cortex in the rat: characterization of the fourth region of subicular cortex and its connections. Journal of Comparative Neurology, 216. 192-210. Wan, H.S .• Touretzky. D.S .• & Redish. D.S. (1994). A rodent navigation model that combines place code. head-direction, and path integration information. Society for Neuroscience Abstracts, 20, 1205. Zhang, K. (1995). Representation of spatial orientation by the intrinsic dynamics of the head-direction cell ensemble: A theory. Submitted.
|
1995
|
64
|
1,111
|
Clustering data through an analogy to the Potts model Marcelo Blatt, Shai Wiseman and Eytan Domany Department of Physics of Complex Systems, The Weizmann Institute of Science, Rehovot 76100, Israel Abstract A new approach for clustering is proposed. This method is based on an analogy to a physical model; the ferromagnetic Potts model at thermal equilibrium is used as an analog computer for this hard optimization problem. We do not assume any structure of the underlying distribution of the data. Phase space of the Potts model is divided into three regions; ferromagnetic, super-paramagnetic and paramagnetic phases. The region of interest is that corresponding to the super-paramagnetic one, where domains of aligned spins appear. The range of temperatures where these structures are stable is indicated by a non-vanishing magnetic susceptibility. We use a very efficient Monte Carlo algorithm to measure the susceptibility and the spin spin correlation function. The values of the spin spin correlation function, at the super-paramagnetic phase, serve to identify the partition of the data points into clusters. Many natural phenomena can be viewed as optimization processes, and the drive to understand and analyze them yielded powerful mathematical methods. Thus when wishing to solve a hard optimization problem, it may be advantageous to apply these methods through a physical analogy. Indeed, recently techniques from statistical physics have been adapted for solving hard optimization problems (see e.g. Yuille and Kosowsky, 1994). In this work we formulate the problem of clustering in terms of a ferromagnetic Potts spin model. Using the Monte Carlo method we estimate physical quantities such as the spin spin correlation function and the susceptibility, and deduce from them the number of clusters and cluster sizes. Cluster analysis is an important technique in exploratory data analysis and is applied in a variety of engineering and scientific disciplines. The problem of partitionaZ clustering can be formally stated as follows. With everyone of i = 1,2, ... N patterns represented as a point Xi in a d-dimensional metric space, determine the partition of these N points into M groups, called clusters, such that points in a cluster are more similar to each other than to points in different clusters. The value of M also has to be determined. Clustering Data through an Analogy to the Potts Model 417 The two main approaches to partitional clustering are called parametric and nonparametric. In parametric approaches some knowledge of the clusters' structure is assumed (e.g. each cluster can be represented by a center and a spread around it). This assumption is incorporated in a global criterion. The goal is to assign the data points so that the criterion is minimized. A typical example is variance minimization (Rose, Gurewitz, and Fox, 1993). On the other hand, in non-parametric approaches a local criterion is used to build clusters by utilizing local structure of the data. For example, clusters can be formed by identifying high-density regions in the data space or by assigning a point and its K -nearest neighbors to the same cluster. In recent years many parametric partitional clustering algorithms rooted in statistical physics were presented (see e.g. Buhmann and Kiihnel , 1993). In the present work we use methods of statistical physics in non-parametric clustering. Our aim is to use a physical problem as an analog to the clustering problem. The notion of clusters comes very naturally in Potts spin models (Wang and Swendsen, 1990) where clusters are closely related to ordered regions of spins. We place a Potts spin variable Si at each point Xi (that represents one of the patterns), and introduce a short range ferromagnetic interaction Jij between pairs of spins, whose strength decreases as the inter-spin distance Ilxi - Xj" increases. The system is governed by the Hamiltonian (energy function) 1i = - L hj D8,,8j <i,j> Si = 1 . .. q , (1) where the notation < i, j > stands for neighboring points i and j in a sense that is defined later. Then we study the ordering properties of this inhomogeneous Potts model. As a concrete example, place a Potts spin at each of the data points of fig. 1. ~~--~------~--------~--------~------~--------~------~--~ ·30 ·20 -10 10 20 30 Figure 1: This data set is made of three rectangles, each consisting of 800 points uniformly distributed, and a uniform rectangular background of lower density, also consisting of 800 points. Points classified (with Tclus = 0.08 and () = 0.5) as belonging to the three largest clusters are marked by crosses, squares and x's. The fourth cluster is of size 2 and all others are single point clusters marked by triangles. At high temperatures the system is in a disordered (paramagnetic) phase. As the temperature is lowered, larger and larger regions of high density of points (or spins) exhibit local ordering, until a phase transition occurs and spins in the three rectangular high density regions become completely aligned (i. e. within each region all Si take the same value - super-paramagnetic phase). The aligned regions define the clusters which we wish to identify. As the temperature 418 M. BLATT, S. WISEMAN, E. DOMANY is further lowered, a pseudo-transition occurs and the system becomes completely ordered (ferromagnetic). 1 A mean field model To support our main idea, we analyze an idealized set of points where the division into natural classes is distinct. The points are divided into M groups. The distance between any two points within the same group is d1 while the distance between any two points belonging to different groups is d2 > d1 (d can be regarded as a similarity index). Following our main idea, we associate a Potts spin with each point and an interaction J1 between points separated by distance d1 and an h between points separated by d2 , where a ~ J2 < J1• Hence the Hamiltonian (1) becomes; 1{ = - ~ L L 6~; ,~j - ~ L L 6s; ,sj si = 1, ... , q , (2) /10 i<j /1o<V i ,j where si denotes the ith spin (i = 1, ... , ~) of the lJth group (lJ = 1, ... , M). From standard mean field theory for the Potts model (Wu, 1982) it is possible to show that the transition from the ferromagnetic phase to the paramagnetic phase is at Tc = 2M (qJ.)~Og(q-l) [J1 + (M - 1)h] . The average spin spin correlation function, 6~,,~ j at the paramagnetic phase is t for all points Xi and Xj; i. e. the spin value at each point is independent of the others. The ferromagnetic phase is further divided into two regions. At low temperatures, with high probability, all spins are aligned; that is 6~.,sJ ~ 1 for all i and j. At intermediate temperatures, between T* and Tc, only spins of the same group lJ are aligned with high probability; 6~" ~'-: ~ 1, .' J while spins belonging to different groups, Jl and lJ, are independent; 6~1" s~ ~ 1 . • ' 1 q The spin spin correlation function at the super-paramagnetic phase can be used to decide whether or not two spins belong to the same cluster. In contrast with the mere inter-point distance, the spin spin correlation function is sensitive to the collective behavior of the system and is therefore a suitable quantity for defining collective structures (clusters). The transition temperature T* may be calculated and shown to be proportional to J2 ; T* = a(N, M, q) h. In figure 2 we present the phase diagram, in the (~, ~) plane, for the case M = 4, N = 1000 and q = 6. paramagnetic /' 1e-01 ~ _____ --,-____ ~~ "1 / super-paramagnet~s-,'-' .., 1e"()2 ferromagnetic / "",,; f:: ./,' , ...... ', .. ' 1e-03 ;" .. , .-' 1e..()4~--~~~~--~~----~~~ 1e..()S 1e"()4 1e-03 le-02 le..()1 1e+OO J2JJl Figure 2: Phase diagram of the mean field Potts model (2) for the case M = 4, N = 1000 and q = 6. The critical temperature Tc is indicated by the solid line, and the transition temperature T*, by the dashed line. The phase diagram fig. 2 shows that the existence of natural classes can manifest itself in the thermodynamic properties of the proposed Potts model. Thus our approach is supported, provided that a correct choice of the interaction strengths is made. Clustering Data through an Analogy to the Potts Model 419 2 Definition of local interaction In order to minimize the intra-cluster interaction it is convenient to allow an interaction only bet.ween "neighbors". In common \ .... ith other "local met.hods" , we assume that there is a 'local length scale' '" a, which is defined by the high density regions and is smaller than the typical distance between points in the low density regions. This property can be expressed in the ordering properties of the Potts system by choosing a short range interaction. Therefore we consider that each point interacts only with its neighbors with interaction strength __ 1 (!lXi-Xj!l2) Jij J ji - R exp 2a 2 . (3) Two points, Xi and Xj, are defined as neighbors if they have a mutual neighborhood value J{; that is, if Xi is one of the J{ nearest neighbors of Xj and vice-versa. This definition ensures that hj is symmetric; the number of bonds of any site is less than J{. We chose the "local length scale", a, to be the average of all distances Ilxi - Xj II between pairs i and j with a mutual neighborhood value J{. R is the average number of neighbors per site; i. e it is twice the number of non vanishing interactions, Jij divided by the number of points N (This careful normalization of the interaction strength enables us to estimate the critical temperature Tc for any data sample). 3 Calculation of thermodynanlic quantities The ordering properties of the system are reflected by the susceptibility and the Spill spin correlation functioll D'<"'<J' where -.. -. stands for a thermal average. These quantities can be estimated by averaging over the configurations genel'ated by a Monte Carlo procedure. We use the Swendsen-Wang (Wang and Swendsen, 1990) Monte Carlo algorithm for the Potts model (1) not only because of its high efficiency, but also because it utilizes the SW clusters. As will be explained the SW clusters are strongly connected to the clusters we wish to identify. A layman's explanation of the method is as follows. The SW procedure stochastically identifies clusters of aligned spins, and then flips whole clusters simultaneously. Starting from a given spin configuration, SW go over all the bonds between neighboring points, and either "freeze" or delete them. A bond connecting two neighboring sites i and j, is deleted with probability p~,j = exp( -* 63 .. 3 J and frozen with probability p? = 1 p~,j. Having gone over all the bonds, all spins which have a path of frozen bonds connecting them are identified as being in the same SW cluster. Note t.hat, according to the definition of p~,j, only spins of the same value can be frozen in the same SW cluster. Now a new spin configuration is generated by drawing, for each cluster, randomly a value s = 1, ... q, which is assigned to all its spins. This procedure defines one Monte Carlo step and needs to be iterated in order to obtain thermodynamic averages. At temperatures where large regions of correlated spins occur, local methods (e. g. Metropolis), which flip one spin at a time, become very slow. The SVl method overcomes this difficulty by flipping large clusters of aligned spins simult.aneously. Hence the SW method exhibits much smaller autocorrelation times than local methods. The strong connection between the SW clusters and the ordering properties of the Pot.ts spins is manifested in the relation -6-- (<1- 1)710+ 1 _".,8) q (4) 420 M. BLATI, S. WISEMAN, E. DOMANY where nij = 1 whenever Si and Sj belong to the same SW-cluster and nij = 0 otherwise. Thus, nij is the probability that Si and Sj belong to the same SW-cluster. The r.h.s. of (4) has a smaller variance than its l.h.s., so that the probabilities nij provide an improved estimator of the spin spin correlation function. 4 Locating the super-paramagnetic phase In order to locate the temperature range in which the system IS III the superparamagnetic phase we measure the susceptibility of the system which is proportional to the variance of the magnetization N2 X = T (m2 - m ) . (5) The magnetization, m, is defined as qNmax/N -1 m=-----q-1 (6) where NJ.' is the number of spins with the value J.l. In the ferromagnetic phase the fluctuations of the magnetization are negligible, so the susceptibility, X, is small. As the temperature is raised, a sudden increase of the susceptibility occurs at the transition from the ferromagnetic to the super-paramagnetic phase. The susceptibility is non-vanishing only in the superparamagnetic phase, which is the only phase where large fluctuations in the magnetization can occur. The point where the susceptibility vanishes again is an upper bound for the transition temperature from the super-paramagnetic to the paramagnetic phase. 5 The clustering procedure Our method consists of two main steps. First we identify the range of temperatures where the clusters may be observed (that corresponding to the super-paramagnetic phase) and choose a temperature within this range. Secondly, the clusters are identified using the information contained in the spin spin correlation function at this temperature. The procedure is summarized here, leaving discussion concerning the choice of the parameters to a later stage. (a) Assign to each point Xi a q-state Potts spin variable Si. q was chosen equal to 20 in the example that we present in this work. (b) Find all the pairs of points having mutual neighborhood value K. We set K = 10. (c) Calculate the strength of the interactions using equation (3). (d) Use the SW procedure with the Hamiltonian (1) to calculate the susceptibility X for various temperatures. The transition temperature from the paramagnetic phase _ 1 can be roughly estimated by Tc ~ 410;(1~A)' (e) Identify the range of temperatures of non-vanishing X (the super-paramagnetic phase). Identify the temperature Tmax where the susceptibility X is maximal, and the temperature Tvanish, where X vanishes at the high temperature side. The optimal temperature to identify the clusters lies between these two temperatures. As a rule of thumb we chose the "clustering temperature" Tcltl~ = Tvan .. ~+Tma.r but the results depend only weakly on Tclu~, as long as T cltls is in the super-paramagnetic range, Tmax < Tcltl~ < Tvani~h. Clustering Data through an Analogy to the Potts Model 421 (f) At the clustering temperature Tclu s , estimate the spin spin correlation, o s "s J ' for all neigh boring pairs of points Xi and Xj, using (4) . (g) Clusters are identified according to a thresholding procedure. The spin spin correlation function 03. ,3J of points Xi and Xj is compared with a threshold, (); if OS,,3J > () they are defined as "friends". Then all mutual friends (including fl'iends of friends, etc) are assigned to the same cluster. We chose () = 0.5. In order to show how this algorithm works, let us consider the distribution of points presented in figure 1. Because of the overlap of the larger sparse rectangle with the smaller rectangles, and due to statistical fluctuations, the three dense rectangles actually contain 883, 874 and 863 points. Going through steps (a) to (d) we obtained the susceptibility as a function of the temperature as presented in figure 3. The susceptibility X is maximal at T max = 0.03 and vanishes at Tvanish = 0.13. In figure 1 we present the clusters obtained according to steps (f) and (g) at Tclus = 0.08. The size of the largest clusters in descending order is 900, 894, 877, 2 and all the rest are composed of only one point. The three biggest clusters correspond to the clusters we are looking for, while the background is decomposed into clusters of size one. 0.035 0030 0025 0020 0015 0 010 0 005 0000 0.00 0.02 0.04 0.06 0 ,08 0.10 012 o.te 016 T Figure 3: The susceptibility density x;;. as a function of t.he t.emperature. Let us discuss the effect of the parameters on the procedure. The number of Potts states, q, determines mainly the sharpness of the transition and the critical temperature. The higher q, the sharper the transition. On the other hand, it is necessary to perform more statistics (more SW sweeps) as the value of q increases. From our simulations, we conclude that the influence of q is very weak. The maximal number of neighbors, f{, also affects the results very little; we obtained quite similar results for a wide range of f{ (5 ~ f{ ~ 20). No dramatic changes were observed in the classification, when choosing clustering temperatures Tc1u3 other than that suggested in (e). However this choice is clearly ad-hoc and a better choice should be found. Our method does not provide a natural way to choose a threshold () for the spin spin correlation function. In practice though, the classification is not very sensitive to the value of (), and values in the range 0.2 < () < 0.8 yield similar results. The reason is that the frequency distribution of the values of the spin spin correlation function exhibit.s t.wo peaks, one close to 1 and the other close to 1, while for intermediate values it is verv close q v t.o zero. In figure (4) we present the average size of the largest S\V cluster as a function of the temperature, along with the size of the largest cluster obtained by the thresholding procedUl'e (described in (7)) using three different threshold values () = 0.1, 0 . .5, o .~). Not.e the agreement. between the largest clust.er size defined by t.he threshold e = 0.5 and the average size of the largest SW cluster for all t.emperatures (This agreement holds for the smaller clusters as well) . It support.s our thresholding procedure as a sensible one at all temperatUl'es. 422 500 o~~~~~~~~~~~~~ 0.00 0.02 0.04 0.06 0.08 0.10 0.12 0.14 0.16 T 6 Discussion M. BLATT, S. WISEMAN, E. DOMANY Figure 4: Average size of the largest SW cluster as a function of the temperature, is denoted by the solid line. The triangles, x's and squares denote the size of the largest cluster obtained with thresholds () = 0.2, 0.5 and 0.9 respectively. Other methods that were proposed previously, such as Fukunaga's (1990) , can be formulated as a Metropolis relaxation of a ferromagnetic Potts model at T = O. The clusters are then determined by the points having the same spin value at the local minima of the energy at which the relaxation process terminates. Clearly this procedure depends strongly on the initial conditions. There is a high probability of getting stuck in a metastable state that does not correspond to the desired answer. Such a T = 0 method does not provide any way to distinguish between "good" and "bad" metastable states. We applied Fukunaga's method on the data set of figure (1) using many different initial conditions. The right answer was never obtained. In all runs, domain walls that broke a cluster into two or more parts appeared. Our method generalizes Fukunaga's method by introducing a finite temperature at which the division into clusters is stable. In addition, the SW dynamics are completely insensitive to the initial conditions and extremely efficient. Work in progress shows that our method is especially suitable for hierarchical clustering. This is done by identifying clusters at several temperatures which are chosen according to features of the susceptibility curve. In particular our method is successful in dealing with "real life" problems such as the Iris data and Landsat data. Acknowledgments We thank 1. Kanter for many useful discussions. This research has been supported by the US-Israel Bi-national Science Foundation (BSF) , and the Germany-Israel Science Foundation (GIF). References J .M. Buhmann and H. Kuhnel (1993); Vector quantization with complexity costs, IEEE Trans. Inf. Theory 39, 1133. K. Fukunaga (1990); Introd. to statistical Pattern Recognition, Academic Press. K. Rose, E. Gurewitz, and G.C. Fox (1993); Constrained clustering as an optimization method, IEEE Trans on Patt. Anal. and Mach. Intel. PAMI 15, 785. S. Wang and R.H. Swendsen (1990); Cluster Monte Carlo alg., Physica A 167,565. F.Y. Wu (1982) , The Potts model, Rev Mod Phys, 54, 235. A.L. Yuille and J.J. Kosowsky (1994); Statistical algorithms that converge, Neural Computation 6, 341 (1994).
|
1995
|
65
|
1,112
|
Learning Fine Motion by Markov Mixtures of Experts Marina Meilii Dept. of Elec. Eng. and Computer Sci. Massachussetts Inst. of Technology Cambridge, MA 02139 mmp@ai.mit.edu Michael I. J Ol'dan Dept.of Brain and Cognitive Sciences Massachussetts Inst. of Technology Cambridge, MA 02139 jordan@psyche.mit.edu Abstract Compliant control is a standard method for performing fine manipulation tasks, like grasping and assembly, but it requires estimation of the state of contact (s.o.c.) between the robot arm and the objects involved. Here we present a method to learn a model of the movement from measured data. The method requires little or no prior knowledge and the resulting model explicitly estimates the s.o.c. The current s.o.c. is viewed as the hidden state variable of a discrete HMM. The control dependent transition probabilities between states are modeled as parametrized functions of the measurement. We show that their parameters can be estimated from measurements at the same time as the parameters of the movement in each s.o.c. The learning algorithm is a variant of the EM procedure. The E step is computed exactly; solving the M step exactly is not possible in general. Here, gradient ascent is used to produce an increase in likelihood. 1 INTRODUCTION For a large class of robotics tasks, such as assembly tasks or manipulation of relatively light-weight objects, under appropriate damping of the manipulator the dynamics of the objects can be neglected. For these tasks the main difficulty is in having the robot achieve its goal despite uncertainty in its position relative to the surrounding objects. Uncertainty is due to inaccurate knowledge of the geometric shapes and positions of the objects, of their physical properties (surface friction coefficients), or to positioning errors in the manipulator. The standard solution to this problem is controlled compliance first introduced in (Mason, 1981). Under compliant motion, the task is performed in stages; in each stage the robot arm 1004 M. MElLA, M. I. JORDAN maintains contact with a selected surface or feature of the environment; the stage ends when contact with the feature corresponding to the next stage is made. Decomposing the given task into subtasks and specifying each goal or subgoal in terms of contact constraints has proven to be a particularly fertile idea, from which a fair number of approaches have evolved. But each of them have to face and solve the problem of estimating the state of contact (i.e. checking if the contact with the correct surface is achieved), a direct consequence of dealing with noisy measurements. Additionally, most approaches assume prior geometrical and physical knowledge of the environment. In this paper we present a method to learn a model of the environment which will serve to estimate the s.o.c. and to predict future positions from noisy measurements. It associates to each state of contact the coresponding movement model (m.m.); that is: a relationship between positions, nominal and actual velocities that holds over a domain of the position-nominal velocity space. The current m.m. is viewed as the hidden state variable of a discrete Hidden Markov Model (HMM) with transition probabilities that are parametrized functions of the measurement. We call this model Markov Mixture of Experts (MME) and show how its parameters can be estimated. In section 2 the problem is defined, section 3 introduces the learning algorithm, section 4 presents a simulated example and 5 discusses other aspects relevant to the implementation. 2 REACHABILITY GRAPHS AND MARKOV MIXTURES OF EXPERTS For any ensemble of objects, the space of all the relative degrees of freedom of the objects in the ensemble is called the configuration space (C-space). Every possible configuration of the ensemble is represented by a unique point in the C-space and movement in the real space maps into continuous trajectories in the C-space (Lozano-Perez, 1983). The sets of points corresponding to each state of contact create a partition over the C-space. Because trajectories are continuous, a point can move from a s.o.c. only to a neighboring s.o.c. This can be depicted by a directed graph with vertices representing states of contact and arcs for the possible transitions between them, called the reach ability graph. If no constraints on the velocities are imposed, then in the reachability graph each s.o.c. is connected to all its neighbours. But if the range of velocities is restricted, the connectivity of the graph decreases and the connections are generally non-symmetric. Figure 1 shows an example of a C-space and its reachability graph for velocities with only positive components. Ideally, in the absence of noise, the states of contact can be perfectly observed and every transition through the graph is thus deterministic. To deal with the uncertainty in the measurements, we will attach probabilities to the arcs of the graph in the following way: Let us denote by Qi the set of configurations corresponding to s.o.c. i and let the movement of a point x with uniform nominal velocity v for a time aT be given by x( t + aT) = r (x, v, aT); both x and v are vectors of same dimension as the C-space. Now, let x', v' be the noisy measurements of the true values x, v, x E Qj and P[x, vlx', v',j] the posterior distribution of (x , v) given the measurements and the s.o.c. Then, the probability of transition to a state i from a given state j in time T3 can be expressed as: P[ilx',v',j] = r P[x,vlx',v',j]dxdv = aij(x',V') (1) J{x ,vIXEQj ,rex ,v ,T.)EQ.} Defining the transition probability matrix A = [aji]rj=l and assuming measurement Learning Fine Motion by Markov Mixtures of Experts 1005 y x (a) (b) Figure 1: A configuration space (a) and its reachability graph (b). The nodes represent movement models: C is the free space, A and B are surfaces with static and dynamic friction, G represents jamming in the corner. The velocity V has positive components. noise P[x'lq = i, x E Qd leads to an HMM with output x having a continuous emission probability distribution and where the s.o.c. plays the role of a hidden state variable. Our main goal is to estimate this model from observed data. To give a general statement of the problem we will assume that all the position, velocity and force measurements are represented by the input vector u; the output vector y of dimensionality ny contains the future position (which our model will learn to predict). Observations are made at moments which are integer multiples of T$' indexed by t = 0,1, .. , T. If T$ is a constant sampling time the dependency of the transition probability on Ts can be ignored. For the purpose of the parameter estimation, the possible dependence between u(t) and yet + 1) will also be ignored, but it should be considered when the trained model is used for prediction. Throughout the following section we will also assume that the input-output dependence is described by a Gaussian conditional density p(y(t)lu(t), q(t) = k) with mean f(u(t),(h:) and variance E = (1'21. This is equivalent to assuming that given the S.O.c. all noise is additive Gaussian output noise, which is obviously an approximation. But this approximation will allow us to derive certain quantities in closed form in an effective way. The function feu, (he) is the m.m. associated with state of contact k (with Ok its parameter vector) and q is the selector variable representing it. Sometimes we will find it useful to partition the domain of a m.m. into subdomains and to represent it by a different function (i .e. a different set of parameters Ok) on each of the subdomains; then, the name movement model will be extended to them. The evolution of q is controlled by a Markov chain which depends on u and of a set of parameters W: aij(u(t), W) = Pr[q(t + 1) = ilq(t) = j, u(t)] t = 0, 1, ... with L aij(u, W) = 1 \:Iu, W, j = 1, . .. , m. (2) 1006 M. MElLA, M. I. JORDAN y u ,,------.1& q ! 'd 1-------' t .......................... .................. ............. _ ••••••••••••• ~ .. .............................. _ ...................... . Figure 2: The Markov Mixture of Experts architecture Fig. 2 depicts this architecture. It can be easily seen that this model generalizes the mixture of experts (ME) architecture (Jacobs, et al., 1991), to which it reduces in the case where aij are independent of j (the columns of A are all equal). It becomes the model of (Bengio and Frasconi, 1995) when A and f are neural networks. 3 AN EM ALGORITHM FOR MME To estimate the values of the unknown parameters (J"2, Wk, Ok, k = 1, ... ,m given the sequence of observations {(u(t), y(t))};=o, T> 0 the Expectation Maximization (EM) algorithm will be used. The states {q(t)};=o play the role of the unobserved variables. More about EM can be found in (Dempster et al., 1977) while aspects specific to this algorithm are in (Meila and Jordan, 1994). The E step computes the probability of each state and of every transition to occur at t E {O, ... , T} given the observations and an initial parameter set. This can be done efficiently by the forward-backward algorithm (Rabiner and Juang, 1986). Pr[q(t) = k I {(u(t), y(t))};=o, W, 0, (J"2] (3) Pr[q(t) = j, q(t + 1) = i I {(u(t), y(t))};=o , W, 0, (J"2] In the M step the new estimates of the parameters are found by maximizing the average complete log-likelihood J, which in our case has the form T-l m J(O, (J"2, W) = L L eij(t) lnaij(u(t), W)t=o i,j=l Since each parameter appears in only one term of J the maximization is equivalent to: T 0l:ew = argmin L 'n(t) lIy(t) - f( u(t), Ok)11 2 Ih t=o (5) Learning Fine Motion by Markov Mixtures of Experts 1007 T-l wnew = argmax L L~ij(t) In (aij(u(t), w)) W t=o ij (6) 1 T m ny(T + 1) ~ ~ ''}'k(t) Ily(t) - I(u(t), Ok )11 2 (7) There is no general closed form solution to (5) and (6). Their difficulty depends on the form of I and aij. The complexity of the m.m. is determined by the geometrical shape of the objects' surfaces. For planar surfaces and no rotational degrees of freedom I is linear in Ok. Then, (5) becomes a weighted least squares problem which can be solved in closed form. The functions in A depend both on the movement and of the noise models. Because the noise is propagated through non-linearities to the output, an exact form as in (1) may be hard to compute analytically. Moreover, a correct noise model for each of the possible uncertainties is rarely available (Eberman, 1995). A common practical approach is to trade accuracy for computability and to parametrize A in a form which is easy to update but deprived of physical meaning. In all the cases where maximization cannot be performed exactly, one can resort to Generalized EM by merely increasing J. In particular, gradient ascent in parameter space is a technique which can replace maximization. This modification will not affect the overall convergence of the EM iteration but can significantly reduce its speed. Because EM only finds local maxima of the likelihood, the initialization is important. If I( u, Ok) correspond to physical movement models, good initial estimates for their parameters can be available. The same applies to those components of W which bear physical significance. A complementary approach is to reduce the number of parameters by explicitly setting the probabilities of impossible transitions to O. 4 SIMULATION RESULTS Simulations have been run on the C-space shown in fig. 1. The inputs were the 4-dimensional vectors of position (x, y) and nominal velocity (Vx , Vy); the output was the predicted position. The coordinate range was [0, 10] and the admissible velocities were confined to the upper right quadrant (Vmax 2: Vx, Vy 2: Vmin > 0). The restriction in direction implied that the trajectories remain in the coordinate domain; it also appeared in the topology of the reachability graph, which has no transition to the free space from another state. This model was implemented by a MME. The m.m. are linear in the parameters, corresponding to the piecewise linearity of the true model. To implement the transition matrix A we used a bank of gating net-works, one for each s.o.c., consisting of 2 layer perceptrons with softmax1 output. There are 230 free parameters in the gating networks and 64 in the m.m. The training set included N = 5000 data points, in sequences of length T ~ 6, all starting in free space. The starting position of the sequence and the nominal velocities at each step were picked randomly. We found that a more uniform distribution of the data points over the states of contact is necessary for successful learning. Since this is not expected to happen in applications (where, e.g., sticking occurs less often than sliding) , the obtained models were tested also on a distribution that 1 () exp(WTx) The softmax function is given by: softmax. x = Z ! T ,i = 1, .. m with Wj , x jexp(Wj x) vectors of the same dimension. 1008 M. MElLA, M. I. JORDAN Table 1: Performance of MME versus ME (a) Model Prediction Standard Error (MSE) 1/2 Test set Trammg distributIon Umform V distribution noise level 0 .1 .2 .3 .4 0 .1 .2 .3 .4 MME,(1' =.2 .024 .113 .222 .332 .443 .023 .11 .219 .327 .437 MME,(1' =0 .003 .114 .228 .343 .456 .010 .109 .218 .327 .435 ME, (1' = .2 .052 .133 .25 .37 .493 .044 .129 .247 .367 .488 ME, (1' =0 .047 .131 .25 .37 .49 .034 .126 .245 .366 .488 (b) State Misclassification Error [%] Test set Trammg distribution Umform V distribution noise level 0 .1 . ~ .3 .4 U .1 .~ .3 .4 MME, (1' =.2 5.15 5.2 5.5 5.9 6.4 3.45 3.5 3.8 4.2 4.6 MME, (1' =0 .78 1.40 2.35 3.25 4.13 .89 1.19 1.70 2.30 2.88 ME, (1' =.2 6.46 6.60 7.18 7.73 8.13 3.85 3.90 4.38 4.99 5.65 ME, (1' =0 6.25 6.45 6.98 7.61 8.15 3.84 3.98 4.53 5.05 5.70 was uniform over velocities (and consequently, highly non-uniform over states of contact). Gaussian noise with (1'=0.2 or 0 was added to the (x, y) training data. In the M step, the parameters of the gating networks were updated by gradient ascent. For the m.m.least squares estimation was used. To ensure that models and gates are correctly coupled, initial values for () are chosen around the true values. As discussed in the previous section, this is not an unrealistic assumption. W was initialized with small random values. Each simulation was run until convergence. We used two criteria to measure the performance of the learning algorithm: square root of prediction MSE and hidden state misdassificaton. The results are summarized in table 1. The test set size is 50,000 in all cases. Input noise is Gaussian with levels between 0 and 0.4. Comparisons were made with a ME model with the same number of states. The simulations show that the MME architecture is tolerant to input noise, although it is not taking it into account explicitly. The MME consistently outperforms the ME model in both prediction and state estimation accuracy. 5 DISCUSSION An algorithm to estimate the parameters of composite movement models in the presence of noisy measurements has been presented. The algorithm exploits the physical decomposability of the problem and the temporal relationship between the data points to produce estimates of both the model's parameters and the s.o.c. It requires only imprecise initial knowledge about the geometry and physical properties of the system. Prediction via MME The trained model can be used either as an estimator for the state of contact or as a forward model in predicting the next position. For the former goal the forward part of the forward-backward algorithm can be used to implement a recursive estimator or the methods in (Eberman, 1995) can be used. The obtained 'Yk(t) , combined with the outputs of the movement models, will produce a predicted output y. An improved posterior estimate of y can be obtained Learning Fine Motion by Markov Mixtures of Experts 1009 by combining f) with the current measurement. Scaling issues. Simulations have shown that relatively large datasets are required for training even for a small number of states. But, since the states represent physical entities, the model will inherit the geometrical locality properties thereof. Thus, the number of possible transitions from a state will be bounded by a small constant when the number of states grows, keeping the data complexity linear in m. As a version of EM, our algorithm is batch. It follows that parameters are not adapted on line. In particular, the discretization time T& must be fixed prior to training. But small changes in Ts can be accounted for by rescaling the velocities V. For the other changes, inasmuch as they are local, relearning can be confined to those components of the architecture which are affected. References Bengio, Y. and Frasconi, P. (1995). An input output HMM architecture. In G. Tesauro, D. Touretzky, & T. Leen (Eds.), Neural Information Processing Sys. tems 7, Cambridge, MA: MIT Press, pp. 427-435. Dempster, A. P., Laird, N. M., and Rubin, D. B. (1977). Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, B, 39:1- 38. Eberman, B. S. (1995). A sequential decision approach to sensing manipulation contact features. PhD thesis, M.I.T., Dept. of Electrical Engineering. Jacobs, R. A., Jordan, M. 1., Nowlan, S., & Hinton, G. E. (1991). Adaptive mixtures of local experts. Neural Computation, 3, 1-12. Lozano-Perez, T. (1983). Spatial planning: a configuration space approach. IEEE Transactions on Computers. Mason, M. T. (1981). Compliance and force control for computer controlled manipulation. IEEE Trans. on Systems, Man and Cybernetics. Meila, M. and Jordan, M. 1. (1994). Learning the parameters of HMMs with auxilliary input. Technical Report 9401, MIT Computational Cognitive Science, Cambridge, MA. Rabiner, R. L. and Juang, B. H. (1986). An introduction to hidden Markov models. ASSP Magazine, 3(1):4-16.
|
1995
|
66
|
1,113
|
Stable Linear Approximations to Dynamic Programming for Stochastic Control Problems with Local Transitions Benjamin Van Roy and John N. Tsitsiklis Laboratory for Information and Decision Systems Massachusetts Institute of Technology Cambridge, MA 02139 e-mail: bvr@mit.edu, jnt@mit.edu Abstract We consider the solution to large stochastic control problems by means of methods that rely on compact representations and a variant of the value iteration algorithm to compute approximate costto-go functions. While such methods are known to be unstable in general, we identify a new class of problems for which convergence, as well as graceful error bounds, are guaranteed. This class involves linear parameterizations of the cost-to- go function together with an assumption that the dynamic programming operator is a contraction with respect to the Euclidean norm when applied to functions in the parameterized class. We provide a special case where this assumption is satisfied, which relies on the locality of transitions in a state space. Other cases will be discussed in a full length version of this paper. 1 INTRODUCTION Neural networks are well established in the domains of pattern recognition and function approximation, where their properties and training algorithms have been well studied. Recently, however, there have been some successful applications of neural networks in a totally different context - that of sequential decision making under uncertainty (stochastic control). Stochastic control problems have been studied extensively in the operations research and control theory literature for a long time, using the methodology of dynamic programming [Bertsekas, 1995]. In dynamic programming, the most important object is the cost-to-go (or value) junction, which evaluates the expected future 1046 B. V. ROY, 1. N. TSITSIKLIS cost to be incurred, as a function of the current state of a system. Such functions can be used to guide control decisions. Dynamic programming provides a variety of methods for computing cost-to- go functions. Unfortunately, dynamic programming is computationally intractable in the context of many stochastic control problems that arise in practice. This is because a cost-to-go value is computed and stored for each state, and due to the curse of dimensionality, the number of states grows exponentially with the number of variables involved. Due to the limited applicability of dynamic programming, practitioners often rely on ad hoc heuristic strategies when dealing with stochastic control problems. Several recent success stories - most notably, the celebrated Backgammon player of Tesauro (1992) - suggest that neural networks can help in overcoming this limitation. In these applications, neural networks are used as compact representations that approximate cost- to-go functions using far fewer parameters than states. This approach offers the possibility of a systematic and practical methodology for addressing complex stochastic control problems. Despite the success of neural networks in dynamic programming, the algorithms used to tune parameters are poorly understood. Even when used to tune the parameters of linear approximators, algorithms employed in practice can be unstable [Boyan and Moore, 1995; Gordon, 1995; Tsitsiklis and Van Roy, 1994]. Some recent research has focused on establishing classes of algorithms and compact representation that guarantee stability and graceful error bounds. Tsitsiklis and Van Roy (1994) prove results involving algorithms that employ feature extraction and interpolative architectures. Gordon (1995) proves similar results concerning a closely related class of compact representations called averagers. However, there remains a huge gap between these simple approximation schemes that guarantee reasonable behavior and the complex neural network architectures employed in practice. In this paper, we motivate an algorithm for tuning the parameters of linear compact representations, prove its convergence when used in conjunction with a class of approximation architectures, and establish error bounds. Such architectures are not captured by previous results. However, the results in this paper rely on additional assumptions. In particular, we restrict attention to Markov decision problems for which the dynamic programming operator is a contraction with respect to the Euclidean norm when applied to functions in the parameterized class. Though this assumption on the combination of compact representation and Markov decision problem appears restrictive, it is actually satisfied by several cases of practical interest. In this paper, we discuss one special case which employs affine approximations over a state space, and relies on the locality of transitions. Other cases will be discussed in a full length version of this paper. 2 MARKOV DECISION PROBLEMS We consider infinite horizon, discounted Markov decision problems defined on a finite state space S = {I, .. . , n} [Bertsekas, 1995]. For every state i E S, there is a finite set U(i) of possible control actions, and for each pair i,j E S of states and control action u E U (i) there is a probability Pij (u) of a transition from state i to state j given that action u is applied. Furthermore, for every state i and control action u E U (i), there is a random variable Ciu which represents the one-stage cost if action u is applied at state i. Let f3 E [0,1) be a discount factor. Since the state spaces we consider in this paper Stable Linear Approximations Programming for Stochastic Control Problems 1047 are finite, we choose to think of cost-to-go functions mapping states to cost- to-go values in terms of cost-to-go vectors whose components are the cost-to-go values of various states. The optimal cost-to-go vector V* E !Rn is the unique solution to Bellman's equation: Vi*= min. (E[CiU]+.BLPij(U)Vj*), ViES. (1) uEU(t) jES If the optimal cost-to-go vector is known, optimal decisions can be made at any state i as follows: u*=arg min. (E[CiU]+.BLPij(U)l--j*), ViES. uEU(t) jES There are several algorithms for computing V* but we only discuss the value iteration algorithm which forms the basis of the approximation algorithm to be considered later on. We start with some notation. We define the dynamic programming operator as the mapping T : !Rn r-t !Rn with components Ti : !Rn r-t !R defined by Ti(V) = min. (E[CiU]+.BLPij(U)Vj), ViES. (2) uEU(t) jES It is well known and easy to prove that T is a maximum norm contraction. In particular , IIT(V) - T(V')lloo :s; .BIIV - V'lIoo, The value iteration algorithm is described by V(t + 1) = T(V(t)), where V (0) is an arbitrary vector in !Rn used to initialize the algorithm. It is easy to see that the sequence {V(t)} converges to V*, since T is a contraction. 3 APPROXIMATIONS TO DYNAMIC PROGRAMMING Classical dynamic programming algorithms such as value iteration require that we maintain and update a vector V of dimension n. This is essentially impossible when n is extremely large, as is the norm in practical applications. We set out to overcome this limitation by using compact representations to approximate cost-to-go vectors. In this section, we develop a formal framework for compact representations, describe an algorithm for tuning the parameters of linear compact representations, and prove a theorem concerning the convergence properties of this algorithm. 3.1 COMPACT REPRESENTATIONS A compact representation (or approximation architecture) can be thought of as a scheme for recording a high-dimensional cost-to-go vector V E !Rn using a lowerdimensional parameter vector wE !Rm (m «n). Such a scheme can be described by a mapping V : !Rm r-t !Rn which to any given parameter vector w E !Rm associates a cost-to-go vector V (w). In particular, each component Vi (w) of the mapping is the ith component of a cost-to-go vector represented by the parameter vector w. Note that, although we may wish to represent an arbitrary vector V E !Rn, such a scheme allows for exact representation only of those vectors V which happen to lie in the range of V. In this paper, we are concerned exclusively with linear compact representations of the form V(w) = Mw, where M E !Rnxm is a fixed matrix representing our choice of approximation architecture. In particular, we have Vi(w) = Miw, where Mi (a row vector) is the ith row of the matrix M. 1048 B. V. ROY, J. N. TSITSIKLIS 3.2 A STOCHASTIC APPROXIMATION SCHEME Once an appropriate compact representation is chosen, the next step is to generate a parameter vector w such that V{w) approximates V*. One possible objective is to minimize squared error of the form IIMw V*II~. If we were given a fixed set of N samples {( iI, ~:), (i2' Vi;), ... , (i N, ~:)} of an optimal cost-to-go vector V*, it seems natural to choose a parameter vector w that minimizE's E7=1 (Mij w ~;)2. On the other hand, if we can actively sample as many data pairs as we want, one at a time, we might consider an iterative algorithm which generates a sequence of parameter vectors {w(t)} that converges to the desired parameter vector. One such algorithm works as follows: choose an initial guess w(O), then for each t E {O, 1, ... } sample a state i{t) from a uniform distribution over the state space and apply the iteration (3) where {a(t)} is a sequence of diminishing step sizes and the superscript T denotes a transpose. Such an approximation scheme conforms to the spirit of traditional function approximation - the algorithm is the common stochastic gradient descent method. However, as discussed in the introduction, we do not have access to such samples of the optimal cost-to-go vector. We therefore need more sophisticated methods for tuning parameters. One possibility involves the use of an algorithm similar to that of Equation 3, replacing samples of ~(t) with TiCt) (V(t)). This might be justified by the fact that T(V) can be viewed as an improved approximation to V*, relative to V. The modified algorithm takes on the form (4) Intuitively, at each time t this algorithm treats T(Mw(t)) as a "target" and takes a steepest descent step as if the goal were to find a w that would minimize IIMwT(Mw(t))II~. Such an algorithm is closely related to the TD(O) algorithm of Sutton (1988). Unfortunately, as pointed out in Tsitsiklis and Van Roy (1994), such a scheme can produce a diverging sequence {w(t)} of weight vectors even when there exists a parameter vector w* that makes the approximation error V* - Mw* zero at every state. However, as we will show in the remainder of this paper, under certain assumptions, such an algorithm converges. 3.3 MAIN CONVERGENCE RESULT Our first assumption concerning the step size sequence {a(t)} is standard to stochastic approximation and is required for the upcoming theorem. Assumption 1 Each step size a(t) is chosen prior to the generation of i(t), and the sequence satisfies E~o a(t) = 00 and E~o a 2 (t) < 00. Our second assumption requires that T : lRn t-+ lRn be a contraction with respect to the Euclidean norm, at least when it operates on value functions that can be represented in the form Mw, for some w. This assumption is not always satisfied, but it appears to hold in some situations of interest, one of which is to be discussed in Section 4. Assumption 2 There exists some {3' E [0, 1) such that IIT(Mw) - T(Mw')112 ::; {3'IIMw - Mw'112, Vw,w' E lRm. Stable Linear Approximations to Programming for Stochastic Control Problems 1049 The following theorem characterizes the stability and error bounds associated with the algorithm when the Markov decision problem satisfies the necessary criteria. Theorem 1 Let Assumptions 1 and 2 hold, and assume that M has full column rank. Let I1 = M(MT M)-l MT denote the projection matrix onto the subspace X = {Mwlw E ~m}. Then, (a) With probability 1, the sequence w(t) converges to w*, the unique vector that solves: Mw* = I1T(Mw*). (b) Let V* be the optimal cost-to-go vector. The following error bound holds: IIMw* - V*1I2 ~ (1 ;!~ynllI1V* - V*lloo. 3.4 OVERVIEW OF PROOF Due to space limitations, we only provide an overview of the proof of Theorem 1. Let s : ~m f-7 ~m be defined by s(w) = E [( Miw - Ti(Mw(t)))MT] , where the expectation is taken over i uniformly distributed among {I, .. . , n}. Hence, E[w(t + l)lw(t), a(t)] = w(t) - a(t)s(w(t)), where the expectation is taken over i(t). We can rewrite s as s(w) = ~(MTMW - MTT(MW)) , and it can be thought of as a vector field over ~m. If the sequence {w(t)} converges to some w, then s ( w) must be zero, and we have MTMw MTT(Mw) Mw = I1T(Mw). Note that III1T(Mw) - I1T(Mw')lb ~ {j'IIMw - Mw'112, Vw,w' E ~m, due to Assumption 2 and the fact that projection is a nonexpansion of the Euclidean norm. It follows that I1Te) has a unique fixed point w* E ~m, and this point uniquely satisfies Mw* = I1T(Mw*). We can further establish the desired error bound: IIMw* - V*112 < IIMw* - I1T(I1V*) 112 + III1T(I1V*) - I1V*112 + III1V* - V*112 < {j'IIMw* - V*112 + IIT(I1V*) - V*112 + III1V* - V*1I2 < t3'IIMw* - V*112 + (1 + mv'nIII1V* - V*lloo, and it follows that Consider the potential function U(w) = ~llw w*II~. We will establish that (\1U(w))T s(w) 2 ,U(w), for some, > 0, and we are therefore dealing with a 1050 B. V. ROY, J. N. TSITSIKLIS "pseudogradient algorithm" whose convergence follows from standard results on stochastic approximation [Polyak and Tsypkin, 1972J. This is done as follows: (\7U(w)f s(w) ~ (w - w*) T MT (Mw - T(Mw)) ~ (w - w*) T MT(Mw - IIT(Mw) - (J - II)T(MW)) = ~(MW-Mw*)T(MW-IIT(MW)), where the last equality follows because MTrr = MT. Using the contraction assumption on T and the nonexpansion property of projection mappings, we have IlIIT(Mw) - Mw*112 IIIIT(Mw) - rrT(Mw*)112 ::; ,6'IIMw - Mw*1I2' and applying the Cauchy-Schwartz inequality, we obtain (\7U(W))T s(w) > 1 -(IIMw - Mw*ll~ -IIMw - Mw*1121IMw* - IIT(Mw)112) n !:.(l - ,6')IIMw - Mw*II~· n > Since M has full column rank, it follows that (\7U(W))T s(w) ~ 1'U(w), for some fixed l' > 0, and the proof is complete. 4 EXAMPLE: LOCAL TRANSITIONS ON GRIDS Theorem 1 leads us to the next question: are there some interesting cases for which Assumption 2 is satisfied? We describe a particular example here that relies on properties of Markov decision problems that naturally arise in some practical situations. When we encounter real Markov decision problems we often interpret the states in some meaningful way, associating more information with a state than an index value. For example, in the context of a queuing network, where each state is one possible queue configuration, we might think of the state as a vector in which each component records the current length of a particular queue in the network. Hence, if there are d queues and each queue can hold up to k customers, our state space is a finite grid zt (Le., the set of vectors with integer components each in the range {O, ... ,k-l}). Consider a state space where each state i E {I, ... , n} is associated to a point xi E zt (n = k d ), as in the queuing example. We might expect that individual transitions between states in such a state space are local. That is, if we are at a state xi the next visited state x j is probably close to xi in terms of Euclidean distance. For instance, we would not expect the configuration of a queuing network to change drastically in a second. This is because one customer is served at a time so a queue that is full can not suddenly become empty. Note that the number of states in a state space of the form zt grows exponentially with d. Consequently, classical dynamic programming algorithms such as value iteration quickly become impractical. To efficiently generate an approximation to the cost-to-go vector, we might consider tuning the parameters w E Rd and a E R of an affine approximation ~(w, a) = wT xi + a using the algorithm presented in the previous section. It is possible to show that, under the following assumption Stable Linear Approximations to Programming for Stochastic Control Problems 1051 concerning the state space topology and locality of transitions, Assumption 2 holds with f3' = .; f32 + k~3' and thus Theorem 1 characterizes convergence properties of the algorithm. Assumption 3 The Markov decision problem has state space S = {1, ... , k d }, and each state i is uniquely associated with a vector xi E zt with k ~ 6(1 - (32)-1 + 3. A ny pair xi, x j E zt of consecutively visited states either are identical or have exactly one unequal component, which differs by one. While this assumption may seem restrictive, it is only one example. There are many more candidate examples, involving other approximation architectures and particular classes of Markov decision problems, which are currently under investigation. 5 CONCLUSIONS We have proven a new theorem that establishes convergence properties of an algorithm for generating linear approximations to cost-to-go functions for dynamic programming. This theorem applies whenever the dynamic programming operator for a Markov decision problem is a contraction with respect to the Euclidean norm when applied to vectors in the parameterized class. In this paper, we have described one example in which such a condition holds. More examples of practical interest will be discussed in a forthcoming full length version of this paper. Acknowledgments This research was supported by the NSF under grant ECS 9216531, by EPRI under contract 8030-10, and by the ARO. References Bertsekas, D. P. (1995) Dynamic Programming and Optimal Control. Athena Scientific, Belmont, MA. Boyan, J. A. & Moore, A. W. (1995) Generalization in Reinforcement Learning: Safely Approximating the Value Function. In J. D. Cowan, G. Tesauro, and D. Touretzky, editors, Advances in Neural Information Processing Systems 7. Morgan Kaufmann. Gordon, G. J. (1995) Stable Function Approximation in Dynamic Programming. Technical Report: CMU-CS-95-103, Carnegie Mellon University. Polyak, B. T. & Tsypkin, Y. Z., (1972) Pseudogradient Adaptation and Training Algorithms. A vtomatika i Telemekhanika, 3:45-68. Sutton, R. S. (1988) Learning to Predict by the Method of Temporal Differences. Machine Learning, 3:9-44. Tesauro, G. (1992) Practical Issues in Temporal Difference Learning. Machine Learning, 8:257-277. Tsitsiklis, J. & Van Roy, B. (1994) Feature-Based Methods for Large Scale Dynamic Programming. Technical Report: LIDS-P-2277, Laboratory for Information and Decision Systems, Massachusetts Institute of Technology. Also to appear in Machine Learning.
|
1995
|
67
|
1,114
|
Generalisation of A Class of Continuous Neural Networks John Shawe-Taylor Dept of Computer Science, Royal Holloway, University of London, Egham, Surrey TW20 OEX, UK Email: johnCdcs.rhbnc.ac . uk Jieyu Zhao* IDSIA, Corso Elvezia 36, 6900-Lugano, Switzerland Email: jieyuCcarota.idsia.ch Abstract We propose a way of using boolean circuits to perform real valued computation in a way that naturally extends their boolean functionality. The functionality of multiple fan in threshold gates in this model is shown to mimic that of a hardware implementation of continuous Neural Networks. A Vapnik-Chervonenkis dimension and sample size analysis for the systems is performed giving best known sample sizes for a real valued Neural Network. Experimental results confirm the conclusion that the sample sizes required for the networks are significantly smaller than for sigmoidal networks. 1 Introduction Recent developments in complexity theory have addressed the question of complexity of computation over the real numbers. More recently attempts have been made to introduce some computational cost related to the accuracy of the computations [5]. The model proposed in this paper weakens the computational power still further by relying on classical boolean circuits to perform the computation using a simple encoding of the real values. Using this encoding we also show that Teo circuits interpreted in the model correspond to a Neural Network design referred to as Bit Stream Neural Networks, which have been developed for hardware implementation [8]. With the perspective afforded by the general approach considered here, we are also able to analyse the Bit Stream Neural Networks (or indeed any other adaptive system based on the technique), giving VC dimension and sample size bounds for PAC learning. The sample sizes obtained are very similar to those for threshold networks, *Work performed while at Royal Holloway, University of London 268 1. SHAWE-TAYLOR, J. ZHAO despite their being derived by very different techniques. They give the best bounds for neural networks involving smooth activation functions, being significantly lower than the bounds obtained recently for sigmoidal networks [4, 7]. We subsequently present simulation results showing that Bit Stream Neural Networks based on the technique can be used to solve a standard benchmark problem. The results of the simulations support the theoretical finding that for the same sample size generalisation will be better for the Bit Stream Neural Networks than for classical sigmoidal networks. It should also be stressed that the approach is very general - being applicable to any boolean circuit - and by its definition employs compact digital hardware. This fact motivates the introduction of the model, though it will not play an important part in this paper. 2 Definitions and Basic Results A boolean circuit is a directed acyclic graph whose nodes are referred to as gates, with a single output node of out-degree zero. The nodes with in-degree zero are termed input nodes. The nodes that are not input nodes are computational nodes. There is a boolean function associated with each computational node of arity equal to its in-degree. The function computed by a boolean network is determined by assigning (input) values to its input nodes and performing the function at each computational node once its input values are determined. The result is the value at the output node. The class TCo is defined to be those functions that can be computed by a family of polynomially sized Boolean circuits with unrestricted fanin and constant depth, where the gates are either NOT or THRESHOLD. In order to use the boolean circuits to compute with real numbers we use the method of stochastic computing to encode real numbers as bit streams. The encoding we will use is to consider the stream of binary bits, for which the l's are generated independently at random with probability p, as representing the number p. This is referred to as a Bernoulli sequence of probability p. In this representation, the multiplication of two independently generated streams can be achieved by a simple AND gate, since the probability of a Ion the output stream is equal to P1P2, where Pl is the probability of a 1 on the first input stream and P2 is the probability of a 1 on the second input stream. Hence, in this representation the boolean circuit consisting of a single AND gate can compute the product of its two inputs. More background information about stochastic computing can be found in the work of Gaines [1]. The analysis we provide is made by treating the calculations as exact real valued computations. In a practical (hardware) implementation real bit streams would have to be generated [3] and the question of the accuracy of a delivered result arlses. In the applications considered here the output values are used to determine a binary value by comparing with a threshold of 0.5. Unless the actual output is exactly 1 or ° (which can happen), then however many bits are collected at the output there is a slight probability that an incorrect classification will be made. Hence, the number of bits required is a function of the difference between the actual output and 0.5 and the level of confidence required in the correctness of the classification. Definition 1 The real function computed by a boolean circuit C, which computes the boolean function fe : {O, l}n -* {O, I}, is the function ge : [0, lr -* [0,1], Generalisation of a Class of Continuous Neural Networks 269 obtained by coding each input independently as a Bernoulli sequence and interpreting the output as a similar sequence. Hence, by the discussion above we have for the circuit C consisting of a single AND gate, the function ge is given by ge(:l:1, :1:2) = :1:1:1:2. We now give a proposition showing that the definition of real computation given above is well-defined and generalises the Boolean computation performed by the circuit. Proposition 2 The bit stream on the output of a boolean circuit computing a real function is a Bernoulli sequence. The real function ge computed by an n input boolean circuit C can be expressed in terms of the corresponding boolean function fe as follows: n o:E{O,1}" i=1 In particular, gel{o,)}" = fe · Proof: The output bit stream is a Bernoulli sequence, since the behaviour at each time step is independent of the behaviour at previous time sequences, assuming the input sequences are independent. Let the probability of a 1 in the output sequence be p. Hence, ge (:I:) = p. At any given time the input to the circuit must be one of the 2n possible binary vectors a. P:l(a) gives the probability of the vector a occurring. Hence, the expected value of the output of the circuit is given in the proposition statement, but by the properties of a Bernoulli sequence this value is also p. The final claim holds since Po: (a) = 1, w hile Po: (a') = 0 for a # a' .• Hence, the function computed by a circuit can be denoted by a polynomial of degree n, though the representation given above may involve exponentially many terms. This representation will therefore only be used for theoretical analysis. 3 Bit Stream Neural Networks In this section we describe a neural network model based on stochastic computing and show that it corresponds to taking TCo circuits in the framework considered in Section 2. A Stochastic Bit Stream Neuron is a processing unit which carries out very simple operations on its input bit streams. All input bit streams are combined with their corresponding weight bit streams and then the weighted bits are summed up. The final total is compared to a threshold value. If the sum is larger than the threshold the neuron gives an output 1, otherwise O. There are two different versions of the Stochastic Bit Stream Neuron corresponding to the different data representations. The definitions are given as follows. Definition 3 (AND-SBSN): A n-input AND version Stochastic Bit Stream Neuron has n weights in the range [-1,1 j and n inputs in the range [0,1 j, which are all unipolar representations of Bernoulli sequences. An extra sign bit is attached to each weight Bernoulli sequence. The threshold 9 is an integer lying between -n to n which is randomly generated according to the threshold probability density function ¢( 9). The computations performed during each operational cycle are 270 J. SHAWE-TA YLOR, J. ZHAO (1) combining respectively the n bits from n input Bernoulli sequences with the corresponding n bits from n weight Bernoulli sequences using the AND operation. (2) assigning n weight sign bits to the corresponding output bits of the AND gate, summing up all the n signed output bits and then comparing the total with the randomly generated threshold value. If the total is not less than the threshold value, the AND-SBSN outputs 1, otherwise it outputs O. We can now present the main result characterising the functionality of a Stochastic Bit Stream Neural Network as the real function of an Teo circuit. Theorem 4 The functionality of a family of feedforward networks of Bit Stream Neurons with constant depth organised into layers with interconnections only between adjacent layers corresponds to the function gc for an TCo circuit C of depth twice that of the network. The number of input streams is equal to the number of network inputs while the number of parameters is at most twice the number of weights. Proof: Consider first an individual neuron. We construct a circuit whose real functionality matches that of the neuron. The circuit has two layers. The first consists of a series of AND gates. Each gate links one input line of the neuron with its corresponding weight input. The outputs of these gates are linked into a threshold gate with fixed threshold 2d for the AND-SBSN, where d is the number of input lines to the neuron. The threshold distribution of the AND SBSN is now simulated by having a series of 2d additional inputs to the threshold gate. The number of additional input streams required to simulate the threshold depends on how general a distribution is allowed for the threshold. We consider three cases: 1. If the threshold is fixed (i.e. not programmable), then no additional inputs are required, since the actual threshold can be suitably adapted. 2. If the threshold distribution is always focussed on one value (which can be varied), then an additional flog2(2d)1 (rlog2(d)l) inputs are required to specify the binary value of this number. A circuit feeding the corresponding number of 1 's to the threshold gate is not hard to construct. 3. In the fully general case any series of 2d + 1 (d + 1) numbers summing to one can be assigned as the probabilities of the possible values 4>(0),4>(1), ... , 4>(t), where t 2d for the AND SBSN. We now construct a circuit which takes t input streams and passes the I-bits to the threshold gate of all the inputs up to the first input stream carrying a O. No fUrther input is passed to the threshold gate. In other words Threshold gate receives s q. Input streams 1, ... , s have bit 1 and bits of input either s = t or input stream s + 1 has input o. We now set the probability p, of stream s as follows; PI p, 1 - 4>(0) 1 2:;~~ 4>( i) 1 2:;~g 4>( i) for s = 2, ... , t With these values the probability of the threshold gate receiving s bits is 4>( s) as required. Generalisation of a Class of Continuous Neural Networks 271 This completes the replacement of a single neuron. Clearly, we can replace all neurons in a network in the same manner and construct a network with the required properties provided connections do not 'shortcut' layers, since this would create interactions between bits in different time slots. _ 4 VC Dimension and Sample Sizes In order to perform a VC Dimension and sample size analysis of the Bit Stream Neural Networks described in the previous section we introduce the following general framework. Definition 5 For a set Q of smooth functions f : R n x Rl -+ R, the class F is defined as F = Fg = {fw Ifw{x) = f{x, w), f E Q}. The corresponding classification class obtained by taking a fixed set of s of the functions from Q, thresholding the corresponding functions from F at 0 and combining them (with the same parameter vector) in some logical formula will be denoted H,{F). We will denote H1{F) by H{F). In our case we will consider a set of circuits C each with n + l input connections, n labelled as the input vector and l identified as parameter input connections. Note that if circuits have too few input connections, we can pad them with dummy ones. The set g will then be the set Q=Qe={gc!CEC}, while Fgc will be denoted by Fe. We now quote some of the results of [7] which uses the techniques of Karpinski and MacIntyre [4] to derive sample sizes for classes of smoothly parametrised functions. Proposition 6 [7} Let Q be the set of polynomials P of degree at most d with P : R n x Rl -+ Rand F = Fg = {PwIPw{x) = p{x, w),p E g}. Hence, there are l adjustable parameters and the input dimension is n . Then the VC-dimension of the class H,{Fe) is bounded above by log2{2{2d)l) + 1711og2{s). Corollary 7 For a set of circuits C, with n input connections and l parameter connections, the VC-dimension of the class H,{Fe) is bounded above by Proof: By Proposition 2 the function gc computed by a circuit C with t input connections has the form t gc{x) = L P;e(a)fc{a), where P;e{a) = II xfi{l- xd1- cxi ). i=l Hence, gc( x) is a polynomial of degree t. In the case considered the number t of input connections is n + l. The result follows from the proposition. _ 272 J. SHAWE-TAYLOR. 1. ZHAO Proposition 8 [7] Let 9 be the set of polynomials P of degree at most d with p: 'Rn X 'Rl -+ 'R and F = Fg = {PwIPw(x) = p(x, w),p E g}. Hence, there are l adjustable parameters and the input dimension is n. If a function h E H.(F) correctly computes a function on a sample of m inputs drawn independently according to a fixed probability distribution, where m ~ "",(e, 0) = e(1 ~ y'€) [Uln ( 4e~) + In (2l/(~ - 1)) 1 then with probability at least 1 - 0 the error rate of h will be less than E on inputs drawn according to the same distribution. Corollary 9 For a set of circuits C, with n input connections and l parameter connections, If a function h E H.(Fc) correctly computes a function on a sample of m inputs drawn independently according to a fixed probability distribution, where m ~ "",(e, 0) = e(1 ~ y€) [Uln ( 4eJs~n +l)) + In Cl/(~ - 1)) 1 then with probability at least 1 - 0 the error rate of h will be less than E on inputs drawn according to the same distribution. Proof: As in the proof of the previous corollary, we need only observe that the functions gC for C E C are polynomials of degree at most n + l. • Note that the best known sample sizes for threshold networks are given in [6]: m ~ "",(e, 0) = e(1 ~ y'€) [2Wln (6~) + In (l/(lo- 1)) 1 ' where W is the number of adaptable weights (parameters) and N is the number of computational nodes in the network. Hence, the bounds given above are almost identical to those for threshold networks, despite the underlying techniques used to derive them being entirely different. One surprising fact about the above results is that the VC dimension and sample sizes are independent of the complexity of the circuit (except in as much as it must have the required number of inputs). Hence, additional layers of fixed computation cannot increase the sample complexity above the bound given). 5 Simulation Results The Monk's problems which were the basis of a first international comparison of learning algorithms, are derived from a domain in which each training example is represented by six discrete-valued attributes. Each problem involves learning a binary function defined over this domain, from a sample of training examples of this function. The 'true' concepts underlying each Monk's problem are given by: MONK-I: (attributet = attribute2) or (attribute5 = 1) MONK-2: (attributei = 1) for EXACTLY TWO i E {I, 2, ... , 6} MONK-3: (attribute5 = 3 and attribute4 = 1) or (attribute5 =1= 4 and attribute2 =1= 3) Generalisation of a Class of Continuous Neural Networks 273 There are 124, 169 and 122 samples in the training sets of MONK-I, MONK-2 and MONK-3 respectively. The testing set has 432 patterns. The network had 17 input units, 10 hidden units, 1 output unit, and was fully connected. Two networks were used for each problem. The first was a standard multi-layer perceptron with sigmoid activation function trained using the backpropagation algorithm (BP Network). The second network had the same architecture, but used bit stream neurons in place of sigmoid ones (BSN Network). The functionality of the neurons was simulated using probability generating functions to compute the probability values of the bit streams output at each neuron. The backpropagation algorithm was adapted to train these networks by computing the derivative of the output probability value with respect to the individual inputs to that neuron [8]. Experiments were performed with and without noise in the training examples. There is 5% additional noise (misclassifications) in the training set of MONK-3. The results for the Monk's problems using the moment generating function simulation are shown as follows: BP Network BSN Network training testing training testing MONK-l 100% 86.6% 100% 97.7% MONK-2 100% 84.2% 100% 100% MONK-3 97.1% 83.3% 98.4% 98.6% It can be seen that the generalisation of the BSN network is much better than that of a general multilayer backpropagation network. The results on MONK-3 problem is extremely good. The results reported by Hassibi and Stork [2] using a sophisticated weight pruning technique are only 93.4% correct for the training set and 97.2% correct for the testing set. References [1] B. R. Gaines, Stochastic Computing Systems, Advances in Information Systems Science 2 (1969) pp37-172. [2] B. Hassibi and D.G. Stork, Second order derivatives for network pruning: Optimal brain surgeon, Advances in Neural Information Processing System, Vol 5 (1993) 164-171. [3] P. Jeavons, D.A. Cohen and J. Shawe-Taylor, Generating Binary Sequences for Stochastic Computing, IEEE Trans on Information Theory, 40 (3) (1994) 716-720. [4] M. Karpinski and A. MacIntyre, Bounding VC-Dimension for Neural Networks: Progress and Prospects, Proceedings of EuroCOLT'95, 1995, pp. 337-341, Springer Lecture Notes in Artificial Intelligence, 904. [5] P. Koiran, A Weak Version of the Blum, Shub and Smale Model, ESPRIT Working Group NeuroCOLT Technical Report Series, NC-TR-94-5, 1994. [6] J. Shawe-Taylor, Threshold Network Learning in the Presence of Equivalences, Proceedings of NIPS 4, 1991, pp. 879-886. [7] J. Shawe-Taylor, Sample Sizes for Sigmoidal Networks, to appear in the Proceedings of Eighth Conference on Computational Learning Theory, COLT'95, 1995. [8] John Shawe-Taylor, Peter Jeavons and Max van Daalen, "Probabilistic Bit Stream Neural Chip: Theory", Connection Science, Vol 3, No 3, 1991.
|
1995
|
68
|
1,115
|
Visual gesture-based robot guidance with a modular neural system E. Littmann, Abt. Neuroinformatik, Fak. f. Informatik Universitat Ulm, D-89069 Ulm, FRG enno@neuro.informatik.uni-ulm.de A. Drees, and H. Ritter AG Neuroinformatik, Techn. Fakultat Univ. Bielefeld, D-33615 Bielefeld, FRG andrea,helge@techfak.uni-bielefeld.de Abstract We report on the development of the modular neural system "SEEEAGLE" for the visual guidance of robot pick-and-place actions. Several neural networks are integrated to a single system that visually recognizes human hand pointing gestures from stereo pairs of color video images. The output of the hand recognition stage is processed by a set of color-sensitive neural networks to determine the cartesian location of the target object that is referenced by the pointing gesture. Finally, this information is used to guide a robot to grab the target object and put it at another location that can be specified by a second pointing gesture. The accuracy of the current system allows to identify the location of the referenced target object to an accuracy of 1 cm in a workspace area of 50x50 cm. In our current environment, this is sufficient to pick and place arbitrarily positioned target objects within the workspace. The system consists of neural networks that perform the tasks of image segmentation, estimation of hand location, estimation of 3D-pointing direction, object recognition, and necessary coordinate transforms. Drawing heavily on the use of learning algorithms, the functions of all network modules were created from data examples only. 1 Introduction The rapidly developing technology in the fields of robotics and virtual reality requires the development of new and more powerful interfaces for configuration and control of such devices. These interfaces should be intuitive for the human advisor and comfortable to use. Practical solutions so far require the human to wear a device that can transfer the necessary information. One typical example is the data glove [14, 12]. Clearly, in the long run solutions that are contactless will be much more desirable, and vision is one of the major modalities that appears especially suited for the realization of such solutions. In the present paper, we focus on a still restricted but very important task in robot control, the guidance of robot pick-and-place actions by unconstrained human pointing gestures in a realistic laboratory environment. The input of target locations by 904 E. LITTMANN, A. DREES, H. RITTER pointing gestures provides a powerful, very intuitive and comfortable functionality for a vision-based man-machine interface for guiding robots and extends previous work that focused on the detection of hand location or the discrimination of a small, discrete number of hand gestures only [10, 1, 2, 8]. Besides two color cameras, no special device is necessary to evaluate the gesture of the human operator. A second goal of our approach is to investigate how to build a neural system for such a complex task from several neural modules. The development of advanced artificial neural systems challenges us with the task of finding architect.ures for the cooperat.ion of multiple functional modules such that. part of the structure of the overall system can be designed at a useful level of abstraction, but at the same t.ime learning can be used to create or fine-tune the functionality of parts of t.he system on the basis of suit.able training examples. To approach this goal requires to shift the focus from exploring t.he properties of single networks to exploring the propert.ies of entire systems of neural networks. The work on "mixtures of experts" [3, 4] is one important contribution along these lines. While this is a widely applicable and powerful approach, there clearly is a need to go beyond the exploration of strictly hierarchical systems and to gain experience with architectures t.hat admit more complex types of information flow as required e.g. by the inclusion of feat.ures such as control of focal attention or reent.rant processing branches. The need for such features arose very naturally in the context of the task described above, and in the following sect.ion we will report our results wit.h a system architecture that is crucially based on the exploitation of such elements. 2 System architecture Our system, described in fig. 1, is situated in a complex laboratory environment. A robot arm with manipulator is mounted at one side of a table with several objects of different color placed on it. A human operator is positioned at the next side to the right of the robot. This scenery is watched by two cameras from the other two sides from high above. The cameras yield a stereo color image of t.he scene (images 10). The operator points with one hand at one of the objects on the table. On the basis of the image information, the object is located and the robot grabs it. Then, the operator points at another location, where the robot releases the object. 1 The syst.em consists of several hardware components: a PUMA 560 robot arm with six axes and a three-fingered manipulator 2; two single-chip PULNIX color cameras; two ANDRox vision boards with software for data acquisition and processing; a work space consisting of a table with a black grid on a yellow surface. Robot and person refer to the same work space. Bot.h cameras must show both the human hand and the table with the objects. Within this constraint, the position of the cameras can be chosen freely as long as they yield significantly different views. An important prerequisite for the recognition of the pointing direction is the segmentation of the human hand from the background scenery. This task is solved by a LLM network (Sl) trained to yield a probability value for each image pixel to belong to the hand region. The training is based on t.he local color information. This procedure has been investigated in [7]. An important feature of the chosen method is the great reliability and robustness of both the classification performance and the localization accuracy of the searched object. Furthermore, the performance is quite constant over a wide range of image resolutions. This allows a fast two-step procedure: First, the images are segmented in low resolution (Sl: 11 -+ A1) and the hand position is extracted. Then, a small 1 In analogy to the sea eagle who watches its prey from high above, shoots down to grab the prey, and then flies to a safe place to feed, we nicknamed our system "SEE-EAGLE". 2Development by Prof. Pfeiffer, TV Munich Visual Gesture-based Robot Guidance with a Modular Neural System 905 Fig. 1: System architecture. From two color camera images 10 we extract the hand position (11 I> Sl I> A1 (pixel coord.) I> P1 I> cartesian hand coord.). In a subframe centered on the hand location (12) we determine the pointing direction (12 I> S2 I> A2 (pixel coord.) I> G I> D I> pointing angles). Pointing direction and hand location define a cartesian target location that is mapped to image coord. that define the centers of object subframes (10 I> P2 I> 13). There we determine the target object (13 I> S3 I> A3) and map the pixel coord. of its centers to world coord. (A3 I> P3 I> world target loc.). These coordinates are used to guide the robot R to the target object. 906 E. LITTMANN. A. DREES. H. RlTIER subframe (12) around the estimated hand position is processed in high resolution by another dedicated LLM network (S2: 12 -t A2). For details of the segmentation process, refer to [6]. The extraction of hand information by LLMs on the basis of Gabor masks has already been studied for hand posture [9] and orientation [5]. The method is based on a segmented image containing the hand only (A2). This image is filtered by 36 Gabor masks that are arranged on a 3x3 grid with 4 directions per grid position and centered on the hand. The filter kernels have a radius of 10 pixels, the distance between the grid points is 20 pixels. The 36 filter responses (G) form the input vector for a LLM network (D). Further details of the processing are reported in [6]. The network yields the pointing direction of the hand (D: 12 -t G -t pointing direction). Together with the hand position which is computed by a parametrized self-organizing map ("PSOM", see below and [11, 13]) (P1: Al -t cartesian hand position), a (cartesian) target location in the workspace can be calculated. This location can be retransformed by the PSOM into pixel coordinates (P2: cartesian target location -t target pixel coordinates). These coordinates define the center of an "attention region" (13) that is searched for a set of predefined target objects. This object recognition is performed by a set of LLM color segmentation networks (S3: 13 -t A3), each previously trained for one of the defined targets. A ranking procedure is used to determine the target object. The pixel coordinates ofthe target in the segmented image are mapped by the PSOM to world coordinates (P3: A3 -t cartesian target position). The robot R now moves to above these world coordinates, moves vertically down, grabs whatever is there, and moves upward again. Now, the system evaluates a second pointing gesture that specifies the place where to place the object. This time, the world coordinates calculated on the basis of the pointing direction from network D and the cartesian hand location from PSOM PI serve directly as target location for the robot. For our processing we must map corresponding pixels in the stereo images to cartesian world coordinates. For these transformations, training data was generated with aid of the robot on a precise sampling grid. We automatically extract the pixel coordinates of a LED at the tip of the robot manipulator from both images. The seven-dimensional feature vector serves as training input for an PSOM network [11]. By virtue of its capability to represent a transformation in a symmetric, "multiway" -fashion, this offers the additional benefit that both the camera-to-world mapping and its inverse can be obtained with a single network trained only once on a data set of 27 calibration positions of the robot. A detailed description for such a procedure can be found in [13]. 3 Results 3.1 System performance The accuracy of the current system allows to estimate the pointing target to an accuracy of 1 ± 0.4 cm (average over N = 7 objects at randomly chosen locations in the workspace) in a workspace area of 50x50 cm. In our current environment, this is sufficient to pick and place any of the seven defined target objects at any location in the workspace. This accuracy can only be achieved if we use the object recognition module described in sec. 2. The output of the pointing direction module approximates the target location with an considerably lower accuracy of 3.6± 1.6 cm. 3.2 Image segmentation The problem to evaluate these preprocessing steps has been discussed previously [7], especially the relation of specifity and sensitivity of the network for the given task. As the pointing recognition is based on a subframe centered on the hand center, it is very sensitive to deviations from this center so that a good localization accuracy Visual Gesture-based Robot Guidance with a Modular Neural System 907 is even more important than the classification rate. The localization accuracy is calculated by measuring the pixel distance between the centers determined manually on the original image and as the center of mass in the image obtained after application of the neural network. Table 1 provides quantitative results. On the whole) the two-step cascade of LLM networks yields for 399 out of 4 00 images an activity image precisely centered on the human hand. Only in one image) the first LLM net missed the hand completely) due to a second hand in the image that could be clearly seen in this view. This image was excluded from further processing and from the evaluation of the localization accuracy. Camera A Camera B Pixel deviatIOn NRMSE Pixel deViatIOn NRMSE Person A 0.8 ± 1.2 0.03 ± 0.06 0.8 ± 2.2 0.03 ± 0.09 Person H 1.3 ± 1.4 0.06 ± 0.11 2.2 ± 2.8 0.11 ± 0.21 Table 1: Estimation error of the hand localization on the test set. Absolute error in pixels and normalized error for both persons and both camera images. 3.3 Recognition performance One major problem in recognizing human pointing gestures is the variability of these gestures and their measurement for the acquisition of reliable training information. Different persons follow different strategies where and how to point (fig. 2 (center) and (right». Therefore) we calculate this information indirectly. The person is told to point at a certain grid position with known world coordinates. From the camera images we extract the pixel positions of the hand center and map them to world coordinates using the PSOM net (PI in fig. 1). Given these coordinates the angles of the intended pointing vector with the basis vectors of the world coordinate system can be calculated trigonometrically. These angles form the target vector for the supervised training of a LLM network (D in fig. 1). After training) the output of the net is used to calculate the point where the pointing vector intersects the table surface. For evaluation of the network performance we measure the Euclidian distance between this point and the actual grid point where the person intended to point at. Fig. 3 (left) shows the mean euclidean error MEE of the estimated target position as a function of the number of learning steps. The error on the training set can be considerably reduced) whereas on the test set the improvement stagnates after some 500 training steps. If we perform even more training steps the performance might actually suffer from overfitting. The graph compares training and test results achieved on images obtained by two different ways of determining the hand center. The "manual" curves show the performance that can be achieved if the Gabor masks are manually centered on the hand. For the "neuronal)) curves) the center of mass calculated in the fine-segmented and postprocessed subframe was used. This allows us to study the influence of the error of the segmentation and localization steps on the pointing recognition. This influence is rather small. The MEE increases from 17 mm for the optimal method to 19 mm for the neural method) which is hardly visible in practice. The curves in fig. 3 (center) are obtained if we apply the networks to images of another person. The MEE is considerably larger but a detailed analysis' shows that part of this deviation is due to systematic differences in the pointing strategy as shown in fig. 2 (right). Over a wide range, the number of nodes used for the LLM network has only minor influence on the performance. While obviously the performance on the training set can be arbitrarily improved by spending more nodes, the differences in the MEE on the test set are negligible in a range of 5 to 15 nodes. Using more nodes is problematic as the training data consists of 50 examples only. If not indicated otherwise) we use LLM networks with 10 nodes. Further results) 908 E. LIITMANN. A. DREES. H. RIITER Fig. 2: The table grid points can be reconstructed according to the network output. The target grid is dotted. Reconstruction of training grid (left) and test grid (center) for one person, and of the test grid for another person (right). MER MEB on test oet of unknown perron 30 m ..... aI,trainn :l~ neuronal, train 70 4 Fig. 3: The euclidean error of manual, test 20 ~ 68 ----~--.--66 estimated target point calcue I~ e £ ~ £ 64 lated using the network out10 62 -~. 60 put depends on the prepro~58 cessing (left), and the person 0 56 100 250 sao 1000 2SOO SOOO 100 :l~ sao 1000 2SOO SOOO (center). train.., itHabonr trairq IteratioN comparing the pointing recognition based on only one of the camera images, indicate that the method works better if the camera takes a lateral view rather than a frontal view. All evaluations were done for both persons. The performance was always very similar. 4 Discussion While we begin to understand many properties of neural networks at the single network level, our insight into principled ways of how to build neural systems is still rather limited. Due to the complexity of this task, theoretical progress is (and probably will continue to be) very slow. What we can do in the mean time, however, is to experiment with different design strategies for neural systems and try to "evolve" useful approaches by carefully chosen case studies. The current work is an effort along these lines. It is focused on a challenging, practically important vision task with a number of generic features that are shared with vision tasks for which biological vision systems were evolved. One important issue is how to achieve robustness at the different processing levels of the system. There are only very limited possibilities to study this issue in simulations, since practically nothing is known about the statistical properties of the various sources of error that occur when dealing with real world data. Thus, a real implementation that works with actual data is practically the only way to study the robustness issue in a realistic fashion. Therefore, the demonstrated integration of several functional modules that we had developed previously in more restricted settings [7, 6] was a non-trivial test of the feasability of having these functions cooperate in a larger, modular system. It also gives confidence that the scaling problem can be dealt with successfully if we apply modular neural nets. A related and equally important issue was the use of a processing strategy in which earlier processing stages incrementally restrict the search space for the subsequent stages. Thus, the responsibility for achieving the goal is not centralized in any single module and subsequent modules have always the chance to compensate for limited errors of earlier stages. This appears to be a generally useful strategy for achieving Visual Gesture-based Robot Guidance with a Modular Neural System 909 robustness and for cutting computational costs that is related to the use of "focal attention" , which is clearly an important element of many biological vision systems. A third important point is the extensive use of learning to build the essential constituent functions of the system from data examples. We are not yet able to train the assembled system as a whole. Instead, different modules are trained separately and are integrated only later. Still, the experience gained with assembling a complex system via this "engineering-type" of approach will be extremely valuable for gradually developing the capability of crafting larger functional building blocks by learning methods. We conclude that carefully designed experiments with modular neural systems that are based on the use of real world data and that focus on similar tasks for which also biological neural systems were evolved can make a significant contribution in tackling the challenge that lies ahead of us: to develop a reliable technology for the construction of large-scale artificial neural systems that can solve complex tasks in real world environments. Acknowledgements We want to thank Th. Wengerek (robot control), J. Walter (PSOM implementation), and P. Ziemeck (image acquisition software). This work was supported by BMFT Grant No. ITN9104AO. References [1] T. J. Darell and A. P. Pentland. Classifying hand gestures with a view-based distributed representation. In J. D. Cowan, G. Tesauro, and J. Alspector, editors, Neural Information Processing Systems 6, pages 945-952. Morgan Kaufman, 1994. [2] J. Davis and M. Shah. Recognizing hand gestures. In J.-O. Eklundh, editor, Computer Vision ECCV '94, volume 800 of Lecture Notes in Computer Science, pages 331340. Springer-Verlag, Berlin Heidelberg New York, 1994. [3] R.A. Jacobs, M.1. Jordan, S.J. Nowlan, and G.E. Hinton. Adaptive mixtures of local experts. Neural Computation, 3:79- 87, 1991. [4] M.1. Jordan and R.A. Jacobs. Hierarchical mixtures of experts and the EM algorithm. Neural Computation, 6(2):181-214, 1994. [5] F. Kummert, E. Littmann, A. Meyering, S. Posch, H. Ritter, and G. Sagerer. A hybrid approach to signal interpretation using neural and semantic networks. In Mustererkennung 1993, pages 245-252. Springer, 1993. [6] E. Littmann, A. Drees, and H. Ritter. Neural recognition of human pointing gestures in real images. Submitted to Neural Processing Letters, 1996. [7] E. Littmann and H. Ritter. Neural and statistical methods for adaptive color segmentation a comparison. In G. Sagerer, S. Posch, and F. Kummert, editors, Mustererkennung 1995, pages 84-93. Springer-Verlag, Heidelberg, 1995. [8] C. Maggioni. A novel device for using the hand as a human-computer interface. In Proceedings HC1'93 Human Control Interface, Loughborough, Great Britain, 1993. [9] A. Meyering and H. Ritter. Learning 3D shape perception with local linear maps. In Proc. of the lJCNN, volume IV, pages 432-436, Baltimore, MD, 1992. [10] Steven J. Nowlan and John C. Platt. A convolutional neural network hand tracker. In Neural Information Processing Systems 7. Morgan Kaufman Publishers, 1995. [11] H. Ritter. Parametrized self-organizing maps for vision learning tasks. In P. Morasso, editor, ICANN '94. Springer-Verlag, Berlin Heidelberg New York, 1994. [12] K. Viiiina.nen and K. Bohm. Gesture driven interaction as a human factor in virtual environments - an approach with neural networks. In R. Earnshaw, M. Gigante, and H. Jones, editors, Virtual reality systems, pages 93-106. Academic Press, 1993. [13] J. Walter and H. Ritter. Rapid learning with parametrized self-organizing maps. Neural Computing, 1995. Submitted. [14] T. G. Zimmermann, J. Lanier, C. Blanchard, S. Bryson, and Y. Harvill. A hand gesture interface device. In Proc. CHI+GI, pages 189-192, 1987.
|
1995
|
69
|
1,116
|
hnproved Silicon Cochlea • uSIng Compatible Lateral Bipolar Transistors Andre van Schalk, Eric Fragniere, Eric Vittoz MANTRA Center for Neuromimetic Systems Swiss Federal Institute of Technology CH-IOI5 Lausanne email: vschaik@di.epfl.ch Abstract Analog electronic cochlear models need exponentially scaled filters. CMOS Compatible Lateral Bipolar Transistors (CLBTs) can create exponentially scaled currents when biased using a resistive line with a voltage difference between both ends of the line. Since these CLBTs are independent of the CMOS threshold voltage, current sources implemented with CLBTs are much better matched than current sources created with MOS transistors operated in weak inversion. Measurements from integrated test chips are shown to verify the improved matching. 1. INTRODUCTION Since the original publication of the "analog electronic cochlea" by Lyon and Mead in 1988 [I], several other analog VLSI models have been proposed which try to capture more of the details of the biological cochlear function [2],[3],[4]. In spite of the differences in their design, all these models use filters with exponentially decreasing cutoff frequencies. This exponential dependency is generally obtained using a linear decreasing voltage on the gates of MOS transistors operating in weak-inversion. In weak-inversion, the drain current of a saturated MOS transistor depends exponentially on its gate voltage. The linear decreasing voltage is easily created using a resistive poly silicon line; if there is a voltage difference between the two ends of the line, the voltage on the line will decrease linearly all along its length. 672 A. V AN SCHAlK. E. FRAGNIl1RE. E. VlrrOZ The problem of using MOS transistors in weak-inversion as current sources is that their drain currents are badly matched. An RMS mismatch of 12% in the drain current of two identical transistors with equal gate and source voltages is not exceptional [5], even when sufficient precautions, such as a good layout, are taken. The main cause of this mismatch is a variation of the threshold voltage between the two transistors. Since the threshold voltage and its variance are technology parameters, there is no good way to reduce the mismatch once the chip has been fabricated. One can avoid this problem using Compatible Lateral Bipolar Transistors (CLBTs) [6] for the current sources. They can be readily made in a CMOS substrate, and their collector current also depends exponentially on their base voltage, while this current is completely independent of the CMOS technology's threshold Voltage. The remaining mismatch is due to geometry mismatch of the devices, a parameter which is much better controlled than the variance of the threshold voltage. Therefore, the use of CLBTs can yield a large improvement in the regularity of the spacing of the cochlear filters. This regularity is especially important in a cascade of filters like the cochlea, since one filter can distort the input signal of all the following filters. We have integrated an analog electronic cochlea as a cascade of second-order lOW-pass filters, using CLBTs as exponentially scaled current sources. The design of this cochlea is based on the silicon cochlea described in [7], since a number of important design issues, such as stability, dynamic range, device mismatch and compactness, have already been addressed in this design. In this paper, the design of [7] is briefly presented and some remaining possible improvements are identified. These improvements, notably the use of Compatible Lateral Bipolar Transistors as current sources, a differentiation that does not need gain correction and temperature independent biasing of the cut-off frequency, are then discussed in more detail. Finally, measurement results of a test chip will be presented and compared to the design without CLBTs. 2. THE ANALOG ELECTRONIC COCHLEA The basic building block for the filters in all analog electronic cochlear models is the transconductance amplifier, operated in weak inversion. For input voltages smaller than about 60 mV pp, the amplifier can be approximated as a linear transconductance: with transconductance gm given by: 10 gm = 2nUT (1) (2) where Io is the bias current, n is the slope factor, and the thermal voltage UT = kT/q = 25.6 mV at room temperature. This linear range is usually the input range used in the cochlear filters, yielding linear filters. In [7], a transconductance amplifier having a wider linear input range is proposed. This allows larger input signals to be used, up to about 140 m Vpp. Furthermore, the wide range transconductance amplifier can be used to eliminate the large-signal instability shown to be present in the original second-order section [7]. This second-order section will be discussed in more detail in section 3.2. Improved Silicon Cochlea Using Compatible Lateral Bipolar Transistors 673 The traditional techniques to improve matching [5], as for instance larger device sizes for critical devices and placing identical devices close together with identical orientation, are also discussed in [7] with respect to the implementation of the cochlear filter cascade. The transistors generating the bias current 10 of the transconductance amplifiers in the second-order sections were identified as the most critical devices, since they have the largest effect on the cut-off frequency and the quality factor of each section. Therefore, extra area had to be devoted to these bias transistors. A further improvement is obtained in [7] by using a single resistive line to bias both the transconductance amplifiers controlling the cut-off frequency and the transconductance amplifier controlling the quality factor. The quality factor Q is then changed by varying the source of the transistor which biases the Q control amplifier. Instead of using two tilted resistive lines, this scheme uses only one tilted resistive line and a non-tilted Q control line, and therefore doesn't need to rely on an identical tilt on both resistive lines. 3. IMPROVED ANALOG ELECTRONIC COCHLEA The design discussed in the previous section already showed a substantial improvement over the first analog electronic cochlea by Lyon and Mead. However, several improvements remain possible. 3.1 VT VARIATION The bias transistors have been identified as the major source of mismatch of the cochlea's parameters. This mismatch is mainly due to variation of the threshold voltage VT of the MOS transistors. Since the drain current of a saturated MOS transistor in weak-inversion depends exponentially on the difference between its gate-source voltage and its threshold voltage, small variations in VT introduce large variations in the drain current of these transistors, and since both the cut-off frequency and the quality factor of the filters are proportional to these drain currents, large parameter variations are generated by small V T variations. This problem can be circumvented by the use of CMOS Compatible Lateral Bipolar transistors as bias transistors. A CMOS Compatible Lateral Bipolar Transistor is obtained if the drain or source junction of a MOS transistor is forward-biased in order to inject minority carriers into the local substrate. If the gate voltage is negative enough (for an n-channel device), then no current can flow at the surface and the operation is purely bipolar [6]. Fig. 1 shows the major flows of current carriers in this mode of operation, with the source, drain and well terminals renamed emitter E, collector C and base B. VBC<O :fG C ISub ~ holes -... p ........ electrons n Fig. 1. : Bipolar operation of the MOS transistor: carrier flows and symbol. 674 A. V AN SCHAlK. E. FRAGNIERE. E. VITIOZ Since there is no p+ buried layer to prevent injection to the substrate, this lateral npn bipolar transistor is combined with a vertical npn. The emitter current IE is thus split into a base current IB, a lateral collector current Ic and a substrate collector current Isub• Therefore, the common-base current gain ex. = -IdlE cannot be close to 1. However, due to the very small rate of recombination inside the well and to the high emitter efficiency, the common-emitter current gain ~ = IeIlB can be large. Maximum values of ex. and ~ are obtained in concentric structures using a minimum size emitter surrounded by the collector and a minimum lateral base width. For VCE = VBE-VBC larger than a few hundred millivolts, this transistor is in active mode and the collector current is given, as for a normal bipolar transistor, by fu k=~e~ W where ISb is the specific current in bipolar mode, proportional to the cross-section of the emitter to collector flow of carriers. Since k is independent of the MOS transistor threshold voltage V T, the main source of mismatch of distributed MOS current sources is suppressed, when o....BTs are used to create the current sources. VC.c D __ ...... --'=-_B c::::J lEI _ 0+ poIy-Si p+ (b) Fig. 2. o....BT cascode circuit (a) and its layout (b). A disadvantage of the CLBT is its low Early voltage, i.e., the device has a low output resistance. Therefore, it is preferable to use a cascode circuit as shown in fig. 2. This yields an output resistance several hundred times larger than that of the single o....BT, whereas the area penalty, in a layout as shown in fig 2b, is acceptable. Another disadvantage of the CLBTs, when biased using a resistive line, is their base current, which introduces an additional voltage drop on the resistive line. However, since the cut-off frequencies in the cochlea are controlled by the output current of the CLBTs and since these cut-off frequencies are relatively small, typically 20 kHz, the output current of the CLBTs will be small. If the common-emitter current gain ~ is much larger than 1, the base current of these o....BTs will be very small, and the voltage error introduced by the small base currents will be negligible. Furthermore, since the cut-off frequencies of the cochlea will typically span 2 decades with an exponentially decreasing cut-off frequency from the beginning to the end, only the first few filters will have any noticeable influence on the current drawn from the resistive line. 3.2 DIFFERENTIATION The stabilized second-order section of [7] uses two wide range transconductance amplifiers (A 1 and A2 in fig. 3) with equal bias current and equal capacitive load, to control the cut-off frequency. A basic transconductance amplifier (A3) is used in a Improved Silicon Cochlea Using Compatible Lateral Bipolar Transistors 675 feedback path to control the quality factor of the filter. The voltage VOU1 at the output of each second-order stage represents the basilar membrane displacement. Since the output of the biological cochlea is proportional to the velocity of the basilar membrane, the output of each second-order stage has to be differentiated. In [7] this is done by creating a copy of the output current Lru- of amplifier A2 at every stage. Since the voltage on a capacitor is proportional to the integral of the current onto the capacitor, Idit is effectively proportional to the basilar membrane velocity. Yet, with equal displacement amplitudes, velocity will be much larger for high frequencies than for low frequencies, yielding output signals with an amplitude that decreases from the beginning of the cochlea to the end. This can be corrected by normalizing Lru- to give equal amplitude at every output. A second resistive line with identical tilt controlling the gain of the current mirrors that create the copies of Idit at each stage is used for this purpose in [7]. However, if using a single resistive line for the control of the cut-off frequencies and the quality factor improves the performance of the chip, the same is true for the control of the current mirror gain. fromprev. section Fig. 3. One section of the cochlear cascade, with differentiator. An alternative solution, which does not need normalization, is to take the difference between VOuI and VI (see fig. 3). This can be shown to be equivalent to differentiating V Out. with OdB gain at the cut-off frequency for all stages. This can be easily done with a combination of 2 transconductance amplifiers. These amplifiers can have a large bias current, so they can also be used to buffer the cascade voltages before connecting them to the output pins of the chip, to avoid charging the cochlear cascade with the extra capacitance introduced by the output pins. 3.3 TEMPERATURE SENSITIVITY The cut-off frequency of the first and the last low-pass filter in the cascade can be set by applying voltages to both ends of the resistive line, and the intermediate filters will have a cut-off frequency decreasing exponentially from the beginning to the end. Yet, if we apply directly a voltage to the ends of the resistive line, the actual cut-off frequency obtained will depend on the temperature, since the current depends exponentially on the applied voltage normalized to the thermal voltage Ur (see(3). It is therefore better to create the voltages at both ends of the resistive line on-chip using a current biasing a CLBT with its base connected to its collector (or its drain connected to its gate if aMOS transistor is used). If this gate voltage is buffered, so that the current through the resistive line is not drawn from the input current, the bias currents of the first and last filter, and thus the cut-off frequency of all filters can be set, independent of temperature. 676 A. V AN SCHAlK, E. FRAGNIERE, E. VITTOZ 3.4 THE IMPROVED SILICON COCHLEA The improved silicon cochlea is shown in figure 4. It uses the cochlear sections shown in figure 3, CLBTs as the bias transistors of each filter, and one resistive line to bias all CLBTs. The resistive line is biased using two bipolar current mirror structures and two voltage buffers, which allow temperature independent biasing of the cut-off frequencies of the cochlea. A similar structure is used to create the voltage source V q to control, independent of temperature, the actual quality factor of each section. The actual bipolar current mirror implemented uses the cascode structure shown in figure 2a, however this is not shown in figure 4 for clarity. Vdiffl Fig 4. The improved silicon cochlea 4. TEST RESULTS The proposed silicon cochlea has been integrated using the ECPD15 technology at ES2 (Grenoble, France), containing 104 second-order stages, on a 4.77mm X 3.21mm die. Every second stage is connected to a pin, so its output voltage can be measured. In fig. 5, the frequency response curves after on-chip derivation are shown for the output taps of both the cochlea described in [7] (left), and our version (right). This clearly shows the improved regularity of the cut-off frequencies and the gain obtained using CLBTs. The drop-off in gain for the higher frequency stages (right) is a border effect, since at the beginning of the cochlea no accumulation of gain has yet taken place. In the figure on the left this is not visible, since the first nine outputs are not presented. -20 ·30 F~(Hz) 10000 10 ~ ~ 0 ·10 ·20 ·30 F~(Hz) Fig.5. Measured frequency responses at the different taps. 20000 In fig. 6 we show the cut-off frequency versus tap number of both chips. Ideally, this should be a straight line on a log-linear scale, since the cut-off frequency decreases Improved Silicon Cochlea Using Compatible Lateral Bipolar Transistors 677 exponentially with tap number. This also clearly shows the improved regularity using CLBTs as current sources. lOOOO·r-------------------, 200·~----------------~~ 10 15 20 25 30 o 10 20 30 40 50 Fig.6. Cut-off frequency (Hz) versus tap number for both silicon cochleae. 5. CONCLUSIONS Since the biological cochlea functions as a distributed filter, where the natural frequency decreases exponentially with the position along the basilar membrane, analog electronic cochlear models need exponentially scaled filters. The output current of a Compatible Lateral Bipolar Transistor depends exponentially on the base-emitter voltage. It is therefore easy to create exponentially scaled current sources using CLBTs biased with a resistive polysilicon line. Because the CLBTs are insensitive to variations of the CMOS threshold voltage VT, current sources implemented with CLBTs are much better matched than current sources using MaS transistors in weak inversion. Regularity is further improved using an on-chip differentiation that does not need a second resistive line to correct its gain, and therefore doesn't depend on identical tilt on both resistive lines. Better independence of temperature can be obtained by fixing the frequency domain of the cochlea using bias currents instead of voltages. Acknowledgments The authors would like to thank Felix Lustenberger for simulation and layout of the chip. We are also indebted to Lloyd Watts for allowing us to use his measurement data. References [1] R.F. Lyon and C.A. Mead, "An analog electronic cochlea," IEEE Trans. Acoust .• Speech. Signal Processing, vol. 36, pp. 1119-1134, July 1988. [2] R.F. Lyon, "Analog implementations of auditory models," Proc. DARPA Workshop Speech and Natural Language. San Mateo, CA:Morgan Kaufmann, 1991. [3] W. Liu, et. al., "Analog VLSI implementation of an auditory periphery model," Advances Res. VLSI, Proc. 1991 Santa Cruz Con/., MIT Press, 1991, pp. 153-163. [4] L. Watts, "Cochlear Mechanics: Analysis and Analog VLSI," Ph.D. thesis, California Institute of Technology, Pasadena, 1992. [5] E. Vittoz, "The design of high-performance analog circuits on digital CMOS chips," IEEE 1. Solid-State Circuits, vol. SC-20, pp. 657-665, June 1985. [6] E. Vittoz, "MaS transistors operated in the lateral bipolar mode and their application in CMOS technology," IEEE 1. Solid-State Circuits, vol. SC-24, pp. 273-279, June 1983. [7] L. Watts, et. al., "Improved implementation of the silicon cochlea," IEEE 1. SolidState Circuits, vol. SC-27, pp. 692-700, May 1992.
|
1995
|
7
|
1,117
|
Symplectic Nonlinear Component Analysis Lucas C. Parra Siemens Corporate Research 755 College Road East, Princeton, NJ 08540 lucas@scr.siemens.com Abstract Statistically independent features can be extracted by finding a factorial representation of a signal distribution. Principal Component Analysis (PCA) accomplishes this for linear correlated and Gaussian distributed signals. Independent Component Analysis (ICA), formalized by Comon (1994), extracts features in the case of linear statistical dependent but not necessarily Gaussian distributed signals. Nonlinear Component Analysis finally should find a factorial representation for nonlinear statistical dependent distributed signals. This paper proposes for this task a novel feed-forward, information conserving, nonlinear map - the explicit symplectic transformations. It also solves the problem of non-Gaussian output distributions by considering single coordinate higher order statistics. 1 Introduction In previous papers Deco and Brauer (1994) and Parra, Deco, and Miesbach (1995) suggest volume conserving transformations and factorization as the key elements for a nonlinear version of Independent Component Analysis. As a general class of volume conserving transformations Parra et al. (1995) propose the symplectic transformation. It was defined by an implicit nonlinear equation, which leads to a complex relaxation procedure for the function recall. In this paper an explicit form of the symplectic map is proposed, overcoming thus the computational problems. 438 L. C.PARRA In order to correctly measure the factorization criterion for non-Gaussian output distributions, higher order statistics has to be considered. Comon (1994) includes in the linear case higher order cumulants of the output distribution. Deco and Brauer (1994) consider multi-variate, higher order moments and use them in the case of nonlinear volume conserving transformations. But the calculation of multicoordinate higher moments is computational expensive. The factorization criterion for statistical independence can be expressed in terms of minimal mutual information. Considering only volume conserving transformations allows to concentrate on single coordinate statistics, which leads to an important reduction of computational complexity. So far, this approach (Deco & Schurman, 1994; Parra et aI., 1995) has been restricted to second order statistic. The present paper discusses the use of higher order cumulants for the estimation of the single coordinate output distributions. The single coordinate entropies measured by the proposed technique match the entropies of the sampled data more accurately. This leads in turns to better factorization results. 2 Statistical Independence More general than decorrelation used in PCA the goal is to extract statistical independent features from a signal distribution p(x). We look for a deterministic transformation on ~n: y = f(x) which generates a factorial representation p(y) = It p(Yd, or at least a representation where the individual coordinates P(Yi) of the output variable yare "as factorial as possible". This can be accomplished by minimizing the mutual information M I[P(y)]. n o ::; M I[P(y)] = L H[P(Yi)] - H[P(y)], (1) i=l since M I[P(y)] = 0 holds if p(y) is factorial. The mutual information can be used as a measure of "independence". The entropies H in the definition (1) are defined as usual by H[P(y)] = - J~oop(y)lnp(y)dy. As in linear PCA we select volume conserving transformations, but now without restricting ourselves to linearity. In the noise-free case of reversible transformations volume conservation implies conservation of entropy from the input x to the output y, i.e. H[P(y)] = H[P(x)] = canst (see Papoulis, 1991). The minimization of mutual information (1) reduces then to the minimization of the single coordinate output entropies H[P(Yi)]. This substantially simplifies the complexity of the problem, since no multi-coordinate statistics is required. 2.1 Measuring the Entropy with Cumulants With an upper bound minimization criterion the task of measuring entropies can be avoided (Parra et aI., 1995): (2) Symplectic Nonlinear Component Analysis Edgeworth appIOlClmatlOr'l to second and fanh order O.B,----~--~-~--~-___, 0.7 0.6 ~0 .5 ~ 04 >~ 03 ~ Q.. 0.2 0., o ~ . : .O.~~--=---,!------=---~----: dQ(y1)/dY1 -----~> )i 439 1 Figure 1: LEFT: Doted line: exponential distribution with additive Gaussian noise sampled with 1000 data points. (noise-variance/decay-constant = 0.2). Dashed line: Gaussian approximation equivalent to the Edgeworth approximation to second order. Solid line: Edgeworth approximation including terms up to fourth order. RIGHT: Structure of the volume conserving explicit symplectic map. The minimization of the individual output coordinate entropies H(P(Yi)] simplifies to the minimization of output variances (Ti. For the validity of that approach it is crucial that the map y = f(x) transforms the arbitrary input distribution p(x) into a Gaussian output distribution. But volume conserving and continuous maps can not transform arbitrary distributions into Gaussians. To overcome this problem one includes statistics - higher than second order - to the optimization criterion. Comon (1994) suggests to use the Edgeworth expansion of a probability distribution. This leads to an analytic expression of the entropy in terms of measurable higher order cumulants. Edgeworth expands the multiplicative correction to the best Gaussian approximation of the distribution in the orthonormal basis of Hermite polynomials hcr(y). The expansion coefficients are basically given by the cumulants Ccr of distribution p~y). The Edgeworth expansions reads for a zero-mean distribution with variance (T , (see Kendall & Stuart, 1969) p(y) 2 -l-e-~ f(y) -j2;(J (3) Note, that by truncating this expansion at a certain order, we obtain an approximation Papp(Y), which is not strictly positive. Figure 1, left shows a sampled exponential distribution with additive Gaussian noise. By cutting expansion (3) at fourth order, and further expanding the logarithm in definition of entropy up to sixth order, Comon (1994) approximates the entropy by, 440 L.C.PARRA 1 1 c§ 1 c~ 7 c~ 1 c~ C4 H(P(Y)app] ~ 2"ln(271'e) + In 0' - 120'6 - 480'8 - 480'12 + 8" 0'60'4 (4) We suggest to use this expression to minimize the single coordinate entropies in the definition of the mutual information (1). 2.2 Measuring the Entropy by Estimating an Approximation Note that (4) could only be obtained by truncating the expansion (3). It is therefore limited to fourth order statistic, which might be not enough for a satisfactory approximation. Besides, the additional approximation of the logarithm is accurate only for small corrections to the best Gaussian approximation, i.e. for fey) ~ 1. For distributions with non-Gaussian tails the correction terms might be rather large and even negative as noted above. We therefore suggest alternatively, to measure the entropy by estimating the logarithm of the approximated distribution In Papp (y) with the given data points Yv and using Edgeworth approximation (3) for Papp (y), 1 N 1 N H(P(y)] ~ - N L lnpapp (Yv) = canst + In 0' - N LIn f(yv) (5) v=1 v=1 Furthermore, we suggest to correct the truncated expansion Papp by setting fapp (y) -+ 0 for all fapp (y) < O. For the entropy measurement (5) there is in principle no limitation to any specific order. In table 1 the different measures of entropy are compared. The values in the row labeled 'partition' are measured by counting the numbers n(i) of data points falling in equidistant intervals i of width D.y and summing -pC i)D.y lnp(i) over all intervals, with p(i)D.y = n(i)IN. This gives good results compared to the theoretical values only because of the relatively large sampling size. These values are presented here in order to have an reliable estimate for the case of the exponential distribution, where cumulant methods tend to fail. The results for the exponential distribution show the difficulty of the measurement proposed by Comon, whereas the estimation measurement given by equation (5) is stable even when considering (for this case) unreliable 5th and 6th order cumulants. The results for the symmetric-triangular and uniform distribution demonstrate the insensibility of the Gaussian upper bound for the example of figure 2. A uniform squared distribution is rotated by an angle a. On the abscissa and ordinate a triangular or uniform distribution are observed for the different angles a = II/4 or a = 0 respectively. The approximation of the single coordinate entropies with a Gaussian measure is in both cases the same. Whereas measurements including higher order statistics correctly detect minimal entropy (by fixed total information) for the uniform distribution at a = O. 3 Explicit Symplectic Transformation Different ways of realizing a volume conserving transformation that guarantees H(P(x)] = H(P(x)] have been proposed (Deco & Schurman, 1994; Parra et aI., Symplectic Nonlinear Component Analysis 441 11easured entropy of Gauss uniform triangular exponential sampled distributions symmetric + Gauss noise partition 1.35 ± .02 .024 ± .006 .14 ± .02 1.31 ± .03 Gaussian upper bound (2) 1.415 ± .02 .18 ± .016 .18 ± .02 1.53 ± .04 Coman, eq. (4) 1.414 ± .02 .14 ± .015 .17 ± .02 3.0 ± 2.5 Estimate (5) - 4th order 1.414 ± .02 .13 ± .015 .17±.02 1.39 ± .05 Estimate (5) - 6th order 1.414 ± .02 .092 ± .001 .16 ± .02 1.3 ± .5 theoretical value 1.419 .0 .153 Table 1: Entropy values for different distributions sampled with N = 1000 data points and the different estimation methods explained in the text. The standard deviations are obtained by multiple repetition of the experiment. 1995). A general class of volume conserving transformations are the symplectic maps (Abraham & Marsden, 1978). An interesting and for our purpose important fact is that any symplectic transformation can be expressed in terms of a scalar function. And in turn any scalar function defines a symplectic map. In (Parra et al., 1995) a non-reflecting symplectic transformation has been presented. But its implicit definition results in the need of solving a nonlinear equation for each data point. This leads to time consuming computations which limit in practice the applications to low dimensional problems (n~ 10). In this work reflecting symplectic transformations with an explicit definition are used to define a "feed-forward" volume conserving maps. The input and output space is divided in two partitions x = (Xl, X2) and Y = (Yl, Y2), with Xl, X2, Yl , Y2 E ?Rn / 2 . (6) The structure of this symplectic map is represented in figure 1, right. Two scalar functions P : ?Rn / 2 1-+ ?R and Q : ?Rn / 2 1-+ ?R can be chosen arbitrarily. Note that for quadratic functions equation (6) represents a linear transformation. In order to have a general transformation we introduce for each of these scalar functions a 3-layer perceptron with nonlinear hidden units and a single linear output unit: (7) The scalar functions P and Q are parameterized by the network parameters Wl, W2 E Rm and Wl, W 2 E Rm x Rn/2. The hidden-unit, nonlinear activation function 9 applies to each component of the vectors WlYl and W2X2 respectively. Because of the structure of equation (6) the output coordinates Yl depend only additively on the input coordinates Xl. To obtain a more general nonlinear dependence a second symplectic layer has to be added. To obtain factorial distributions the parameters of the map have to be trained. The approximations of the single coordinate entropies (4) or (5) are inserted in the mutual information optimization criterion (1). These approximations are expressed through moments in terms of the measured output data points. Therefore, the 442 O,B,.---~-~-~-~-~-~-~---, 0,6 0,4 0,2 -0.2 -0.4 -0.6 , , . :.': :' ... " , ..... :.,' , . , -~~,B---0~,6-~-0~.4---0~,2-~--0~.2--0~.4--0~.6-~0,B L.C.PARRA Figure 2: Sampled 2-dimensional squared uniform distribution rotated by 7l" /4. Solid lines represent the directions found by any of the higher order techniques explained in the text. Dashed lines represent directions calculated by linear PCA. (This result is arbitrary and varies with noise) . gradient of these expressions with respect to parameters ofthe map can be computed in principle. For that matter different kinds of averages need to be computed. Even though, the computational complexity is not substantially increased compared with the efficient minimum variances criterion (2), the complexity of the algorithm increases considerably. Therefore, we applied an optimization algorithm that does not require any gradient information. The simple stochastic and parallel update algorithm ALOPEX (Unnikrishnan & Venugopal, 1994) was used. 4 Experiments As explained above, finding the correct statistical independent directions of a rotated two dimensional uniform distribution causes problems for techniques which include only second order statistic. The statistical independent coordinates are simply the axes parallel to the edges of the distribution (see figure 2). A rotation i. e. a linear transformation suffices for this task. The covariance matrix of the data is diagonal for any rotation of the squared distribution and, hence, does not provide any information about the correct orientation of the square. It is well known, that PCA fails to find in the case of non-Gaussian distributions the statistical independent coordinates. Similarly the Gaussian upper bound technique (2)is not capable to minimize the mutual information in this case. Instead, with anyone of the higher order criteria explained in the previous section one finds the appropriate coordinates for any linearly transformed multi-dimensional uniform distribution. This has been observed empirically for a series of setups. The symplectic map was restricted in this experiments to linea1;ity by using square scalar functions. The second example shows that the proposed technique in fact finds nonlinear relations between the input coordinates. An one-dimensional signal distributed according to the distribution of figure 1 was nonlinearly transformed into a twoSymplectic Nonlinear Component Analysis . '.: <~., .' . 443 . : ' .. ; Figure 3: Symplectic map trained with 4th and 2nd order statistics corresponding to the equations (5) and (2) respectively. Left: input distribution. The line at the center of the distribution gives the nonlinear transformed noiseless signal distributed according to the distribution shown in figure 1. Center and Right: Output distribution of the symplectic map corresponding to the 4th order (right) and 2nd order (center) criterion. dimensional signal and corrupted with additive noise, leading to the distribution shown in figure 3, left. The task of finding statistical independent coordinates has been tackled by an explicit symplectic transformation with. n = 2 and m = 6. On figure 3 the different results for the optimization according to the Gaussian upper bound criterion (2) and the approximated entropy criterion (5) are shown. Obviously considering higher order statistics in fact improves the result by finding the better representation of the nonlinear dependency. Reference Abraham, R., & Marsden, J . (1978). Foundations of Mechanics The BenjaminCummings Publishing Company, Inc., London. Comon, P. (1994). Independent component analysis, A new concept Signal Processing, 36, 287- 314. Deco, G., & Brauer, W. (1994). Higher Order Statistical Decorrelation by Volume Concerving Nonlinear Maps. Neural Networks, ? submitted. Deco, G., & Schurman, B. (1994). Learning Time Series Evolution by Unsupervised Extraction of Correlations. Physical Review E, ? submitted. Kendall, M. G., & Stuart, A. (1969). The Advanced Theory of Statistics (3 edition)., Vol. 1. Charles Griffin and Company Limited, London. Papoulis, A. (1991). Probability, Random Variables, and Stochastic Processes. Third Edition, McGraw-Hill, New York. Parra, L., Deco, G., & Miesbach, S. (1995). Redundancy reduction with information-preserving nonlinear maps. Network, 6(1), 61-72. Unnikrishnan, K., P., & Venugopal, K., P. (1994). Alopex: A Correlation-Based Learning Algorithm for Feedforward and Recurrent Neural Networks. Neural Computation, 6(3), 469- 490.
|
1995
|
70
|
1,118
|
Independent Component Analysis of Electroencephalographic Data Scott Makeig Naval Health Research Center P.O. Box 85122 San Diego CA 92186-5122 scott~cplJmmag.nhrc.navy.mil Tzyy-Ping Jung Naval Health Research Center and Computational Neurobiology Lab The Salk Institute, P.O. Box 85800 San Diego, CA 92186-5800 jung~salk.edu Anthony J. Bell Computational Neurobiology Lab The Salk Institute, P.O. Box 85800 San Diego, CA 92186-5800 tony~salk.edu Terrence J. Sejnowski Howard Hughes Medical Institute and Computational Neurobiology Lab The Salk Institute, P.O. Box 85800 San Diego, CA 92186-5800 terry~salk.edu Abstract Because of the distance between the skull and brain and their different resistivities, electroencephalographic (EEG) data collected from any point on the human scalp includes activity generated within a large brain area. This spatial smearing of EEG data by volume conduction does not involve significant time delays, however, suggesting that the Independent Component Analysis (ICA) algorithm of Bell and Sejnowski [1] is suitable for performing blind source separation on EEG data. The ICA algorithm separates the problem of source identification from that of source localization. First results of applying the ICA algorithm to EEG and event-related potential (ERP) data collected during a sustained auditory detection task show: (1) ICA training is insensitive to different random seeds. (2) ICA may be used to segregate obvious artifactual EEG components (line and muscle noise, eye movements) from other sources. (3) ICA is capable of isolating overlapping EEG phenomena, including alpha and theta bursts and spatially-separable ERP components, to separate ICA channels. (4) N onstationarities in EEG and behavioral state can be tracked using ICA via changes in the amount of residual correlation between ICA-filtered output channels. 146 S. MAKEIG, A. l . BELL, T.-P. lUNG, T. l. SEJNOWSKI 1 Introduction 1.1 Separating What from Where in EEG Source Analysis The joint problems of EEG source segregation, identification, and localization are very difficult, since the problem of determining brain electrical sources from potential patterns recorded on the scalp surface is mathematically underdetermined. Recent efforts to identify EEG sources have focused mostly on verforming spatial segregation and localization of source activity [4]. By applying the leA algorithm of Bell and Sejnowski [1], we attempt to completely separate the twin problems of source identification (What) and source localization (Where). The leA algorithm derives independent sources from highly correlated EEG signals statistically and without regard to the physical location or configuration of the source generators. Rather than modeling the EEG as a unitary output of a multidimensional dynamical system, or as "the roar of the crowd" of independent microscopic generators, we suppose that the EEG is the output of a number of statistically independent but spatially fixed potential-generating systems which may either be spatially restricted or widely distributed. 1.2 Independent Component Analysis Independent Component Analysis (leA) [1, 3] is the name given to techniques for finding a matrix, Wand a vector, w, so that the elements, u = (Ul .. . uNF, of the linear transform u = Wx + W of the random vector, x = [Xl ... xNF, are statistically independent. In contrast with decorrelation techniques such as Principal Components Analysis (peA) which ensure that {UiUj} = 0, Vij, ICA imposes the much stronger criterion that the multivariate probability density function (p .d.f.) of u factorizes: fu(u) = n::l fu.(ud . Finding such a factorization involves making the mutual information between the Ui go to zero: I(ui,uj) = O,Vij. Mutual information is a measure which depends on all higher-order statistics of the Ui while decorrelation only takes account of 2nd-order statistics. In (1], a new algorithm was proposed for carrying out leA. The only prior assumption is that the unknown independent components, Ui, each have the same form of cumulative density function (c.d.f.) after scaling and shifting, and that we know this form, call it Fu(u). ICA can then be performed by maximizing the entropy, H(y), of a non-linearly transformed vector: y = Fu(u) . This yields stochastic gradient ascent rules for adjusting Wand w: where y = (:ih ... YN F, the elements of which are: , a 0Yi (h ( )] Yi = -whic if y = Fu U 0Yi OUi _ Ofu(Ui) OFu(Ui) (1) (2) It can be shown that an leA solution is a stable point of the relaxation of eqs.(1-2). In practical tests on separating mixed speech signals, good results were found when using the logistic function, Yi = (1 + e-u• )-1, instead of the known c.d.f., Fu, of the speech signals. In this case Yi = 1 - 2Yi, and the algorithm has a simple form. These results were obtained despite the fact that the p.d.f. of the speech signals was not exactly matched by the gradient of the logistic function. In the experiments in this paper, we also used the speedup technique of prewhitening described in [2]. Independent Component Analysis of Electroencephalographic Data 147 1.3 Applying leA to EEG Data The leA technique appears ideally suited for performing source separation in domains where, (1) the sources are independent, (2) the propagation delays of the 'mixing medium' are negligible, (3) the sources are analog and have p.d.f.'s not too unlike the gradient of a logistic sigmoid, and (4) the number of independent signal sources is the same as the number of sensors, meaning if we employ N sensors, using the ICA algorithm we can separate N sources. In the case of EEG signals, N scalp electrodes pick up correlated signals and we would like to know what effectively 'independent brain sources' generated these mixtures. If we assume that the complexity of EEG dynamics can be modeled, at least in part, as a collection of a modest number of statistically independent brain processes, the EEG source analysis problem satisfies leA assumption (1). Since volume conduction in brain tissue is effectively instantaneous, leA assumption (2) is also satisfied. Assumption (3) is plausible, but assumption (4), that the EEG is a linear mixtures of exactly N sources, is questionable, since we do not know the effective number of statistically independent brain signals contributing to the EEG recorded from the scalp. The foremost problem in interpreting the output of leA is, therefore, determining the proper dimension of input channels, and the physiological and/or psychophysiological significance of the derived leA source channels. Although the leA model of the EEG ignores the known variable synchronization of separate EEG generators by common subcortical or corticocortical influences [5], it appears promising for identifying concurrent signal sources that are either situated too close together, or are too widely distributed to be separated by current localization techniques. Here, we report a first application of the ICA algorithm to analysis of 14-channel EEG and ERP recordings during sustained eyes-closed performance of an auditory detection task, and give evidence suggesting that the leA algorithm may be useful for identifying psychophysiological state transitions. 2 Methods EEG and behavioral data were collected to develop a method of objectively monitoring the alertness of operators of complex systems [8] . Ten adult volunteers participated in three or more half-hour sessions, during which they pushed one button whenever they detected an above-threshold auditory target stimulus (a brief increase in the level of the continuously-present background noise). To maximize the chance of observing alertness decrements, sessions were conducted in a small, warm, and dimly-lit experimental chamber, and subjects were instructed to keep their eyes closed. Auditory targets were 350 ms increases in the intensity of a 62 dB white noise background, 6 dB above their threshold of detectability, presented at random time intervals at a mean rate of 10/min, and superimposed on a continuous 39-Hz click train evoking a 39-Hz steady-state response (SSR). Short, and task-irrelevant probe tones of two frequencies (568 and 1098 Hz) were interspersed between the target noise bursts at 2-4 s intervals. EEG was collected from thirteen electrodes located at sites of the International 10-20 System, referred to the right mastoid, at a sampling rate of 312.5 Hz. A bipolar diagonal electrooculogram (EOG) channel was also recorded for use in eye movement artifact correction and rejection. Target Hits were defined as targets responded to within a 100-3000 ms window, while Lapses were targets not responded to. Two sessions each from three of the subjects were selected for analysis based on their containing at least 50 response Lapses. A continuous performance measure, local error rate, was computed by convolving the irregularly-sampled performance index time series (Hit=O/Lapse=l) with a 95 s smoothing window advanced for 1.64 s steps. 148 S. MAKEIG, A. l. BELL, T.-P. lUNG, T. 1. SEJNOWSKI The leA algorithm in eqs.(1-2) was applied to the 14 EEG recordings. The time index was permuted to ensure signal stationarity, and the 14-dimensional time point vectors were presented to a 14 ---. 14 leA network one at a time. To speed convergence, we first pre-whitened the data to remove first- and second-order statistics. The learning rate was annealed from 0.03 to 0.0001 during convergence. After each pass through the whole training set, we checked the amount of correlation between the leA output channels and the amount of change in weight matrix, and stopped the training procedure when, (1) the mean correlation among all channel pairs was below 0.05, and (2) the leA weights had stopped changing appreciably. 3 Results A small (4.5 s) portion of the resulting leA-transformed EEG time series is shown in Figure 1. As expected, correlations between the leA traces are close to zero. The dominant theta wave (near 7 Hz) spread across many EEG channels (left paneQ is more or less isolated to leA trace 1 (upper right), both in the epoch shown and throughout the session. Alpha activity (near 10 Hz) not obvious in the EEG data is uncovered in leA trace 2, which here and throughout the session contains alpha bursts interspersed with quiescent periods. Other leA traces (3-8) contain brief oscillatory bursts which are not easy to characterize, but clearly display different dynamics from the activity in leA trace 1 which dominates the raw EEG record. ICA trace 10 contains near-De changes associated with eye slow movements in the EOG and most frontal (Fpz) EEG channels. leA trace 13 contains mostly line noise (60 Hz), while ICA traces 9 and 14 have a broader high frequency (50-100 Hz) spectrum, suggesting that their source is likely to be high-frequency activity generated by scalp muscles. Apparently, the ICA source solution for this data does not depend strongly on learning rate or initial conditions. When the same portion of one session was used to train two leA networks with different random starting weights, data presentation orders, and learning rates, the two final ICA weight matrices were very close to one another. Filtering another segment of EEG data from the same session using each ICA matrix produced two ICA source transforms in which 11 of the 14 bestcorrelated output channel pairs correlated above 0.95 and none correlated less than 0.894. While ICA training minimized mutual information, and therefore also correlations between output channels during the initial (alert) leA training period, output data channels filtered by the same leA weight matrix became more correlated during the drowsy portion of the session, and then reverted to their initial levels of (de)correlation when the subject again became alert. Conversely, filtering the same session's data with an leA weight matrix trained on the drowsy portion of the session produced output channels that were more correlated during the alert portions of the session than during the drowsy training period. Presumably, these changes in residual correlation among ICA outputs reflect changes in the dynamics and topographic structure of the EEG signals in alert and drowsy brain states. An important problem in human electrophysiology is to determine a means of objectively identifying overlapping ERP subcomponents. Figure 3 (right paneQ shows an leA decomposition of (left paneQ ERPs to detected (Hit) and undetected (Lapse) targets by the same subject. leA spatial filtering produces two channels (S[I-2]) separating out the 39-Hz steady-state response (SSR) produced by the continuous 39-Hz click stimulation during the session. Note the stimulus-induced perturbation in SSR amplitude previously identified in [6]. Three channels (H[I-3]) pass time-limited components of the detected target response, while four others (1[1-4]) Independent Component Analysis of Electroencephalographic Data J 49 EEG leA Fz ~V~~~hI'o/A. Cz ~~MvN'N{\~v<wv'yJ\J\r"~ pz V&V\fM~IIjJ-r~ F3 .;vwvwvvvrv~~WV'JI~ F4 ~\0.,fvo/lf'1Vlf\,~~~ C3 VIV"'vVWv!l!vWN/W\'~~ C4 \MIV{lAtv!ifVV{\AfJV0~~ T4 ~~~~ P3 v"V'v""Nv\~"'-'V P4 M'-VVI/<{WY'V0,Ww~ Fpz ~\~I"IVlV';.~~ EOG~~~ 3 VfJV\.'\I\~~~'~ 4 'rIvV\.JJvvV'-r~·~, 5 i~'MI'\'V1fV{\tNN~10~~ 6 {.I'VVVVVvw....;;~~rwvr'ri(.,'r·Nvf 7 /<¥Yl1~'riiwNV~~~~ 8 ~v.AJvJw-~\Jn~ 9 >I*~vw~Y!"'~.fW'Iwi'fr."'I 10 ~~~~~ 11 1~IVVV~\fvv{iJYlw~~ 12/¥v\~~ 13~~~~~~~ 14 ~~\"W!~~~~~ -- 1 sec. Figure 1: Left: 4.5 seconds of 14-channel EEG data. Right: an le A transform of the same data, using weights trained on 6.5 minutes of similar data from the same seSSlOn. 150 Scalp ERPs Fz ~~~~...., + Cz ;., ~~P •. ..-.woiQfiS$i + pz -... + "-,;r::> Gtr~ --,.~ + Oz ~ .,. , Q ... ¢I,;;" '''iIi 04"'. + F3 ~ ""~~n;; + F4 :1 ~;f'O"IV'~~ 'tV; + C3 -., ' 1;so~ ; _ .... '''''''S + C4 ;;., ,~, coA'C>"",~ .... ,W Ii ::v; + T3 ~ 1"·" ,,~"" ~;t;!¥ ;Nft~h'" + T4 ;'+~~t ~~~."w + - .J P3 1 c ......... r::c;.,.~4&f' ! c 44c;c;e + P4 ~." 0« '4 ~~;. '" • j" ""*+ + FPZ~n~~~ •. + EOG;""1 :: -............ ~ ... '1* it~ + S. MAKEIG, A. J. BELL, T.-P. JUNG, T. J. SEJNOWSKI ICAERPs - I H2;c _'_~4W. ;11 + H3 ~ • ...,p.~~ Rep ~ L1 -..,~. v.. ..At '~"lf"'o ~ + L2 - +~~"4 = e ..... ~ + L3 ::"I-·~~ ~ + L4 -1$ f- ...... · A.E:::::;? <:::!> <P ~ + 81 ~'f'r4'i"'Qf4;.~TIN'I"m"'''''~''i + 82 ;,"1Y¥".-U'lt ~'tt~'.'~'~'~"'fH','t1 + U1 ;'~t.,,=,,,~~FW"" ~ .. _~ + U2 ~ I-- iiIIe<>"" ,;I6'O'~' t'4ac*'" , + U3 -.... ro1~C--_ 'k. • ... + U4 ~~ 111(ok i. tdt .. 'IWoio'" "M I~ C + US :+~[P'~Q""'~ "' ..... ~ .. _ o 0.2 0.4 0.6 0.8 1 sec Detected targets Undetected targets Figure 2: Left panel: Event-related potentials (ERPs) in response to undetected (bold traces) and detected (faint traces) noise targets during two half-hour sessions. Right panel: Same ERP signals filtered using an leA weight matrix trained on the ERP data. Independent Component Analysis of Electroencephalographic Data 151 components of the (larger) undetected target response. We suggest these represent the time course of the locus (either focal or distributed) of brain response activity, and may represent a solution to the longstanding problem of objectively dividing evoked responses into neurobiologically meaningful, temporally overlapping subcomponents. 4 Conclusions ICA appears to be a promising new analysis tool for human EEG and ERP research. It can isolate a wide range of artifacts to a few output channels while removing them from remaining channels. These may in turn represent the time course of activity in longlasting or transient independent 'brain sources' on which the algorithm converges reliably. By incorporating higher-order statistical information, ICA avoids the non-uniqueness associated with decorrelating decompositions. The algorithm also appears to be useful for decomposing evoked response data into spatially distinct subcomponents, while measures of nonstationarity in the ICA source solution may be useful for observing brain state changes. Acknowledgments This report was supported in part by a grant (ONR.Reimb.30020.6429) to the Naval Health Research Center by the Office of Naval Research. The views expressed in this article are those of the authors and do not reflect the official policy or position of the Department of the Navy, Department of Defense, or the U.S. Government. Dr. Bell is supported by grants from the Office of Naval Research and the Howard Hughes Medical Institute. References [1] A.J. Bell & T.J. Sejnowski (1995). An information-maximization approach to blind separation and blind deconvolution, Neural Computation 7:1129-1159. [2] A.J. Bell & T.J. Sejnowski (1995). Fast blind separation based on information theory, in Proc. Intern. Symp. on Nonlinear Theory and Applications (NOLTA), Las Vegas, Dec. 1995. [3] P. Comon (1994) Independent component analysis, a new concept? Signal processing 36:287-314. [4] A.M. Dale & M.1. Sereno (1993) EEG and MEG source localization: a linear approach. J. Cogn. Neurosci. 5:162. [5] R. Galambos & S. Makeig. (1989) Dynamic changes in steady-state potentials. In Erol Basar (ed.), Dynamics of Sensory and Cognitive Processing of the Brain, 102-122. Berlin:Springer-Verlag. [6] S. Makeig & R. Galambos. (1989) The CERP: Event-related perturbations in steady-state responses. In E. Basar & T.H. Bullock (ed.), Brain Dynamics: Progress and Perspectives, 375-400. Berlin:Springer-Verlag. [7] T-P. Jung, S. Makeig, M. Stensmo, & T. Sejnowski. Estimating alertness from the EEG power spectrum. Submitted for publication. [8] S. Makeig & M. Inlow (1993) Lapses in alertness: Coherence of fluctuations in performance and EEG spectrum. Electroencephalog. din. N europhysiolog. 86:23-35.
|
1995
|
71
|
1,119
|
Competence Acquisition in an Autonomous Mobile Robot using Hardware Neural Techniques. Geoff Jackson and Alan F. Murray Department of Electrical Engineering Edinburgh University Edinburgh, ER9 3JL Scotland, UK gbj@ee.ed.ac.uk,afm@ee.ed.ac.uk Abstract In this paper we examine the practical use of hardware neural networks in an autonomous mobile robot. We have developed a hardware neural system based around a custom VLSI chip, EPSILON III, designed specifically for embedded hardware neural applications. We present here a demonstration application of an autonomous mobile robot that highlights the flexibility of this system. This robot gains basic mobility competence in very few training epochs using an "instinct-rule" training methodology. 1 INTRODUCTION Though neural networks have been shown as an effective solution for a diverse range of real-world problems, applications and especially hardware implementations have been few and slow to emerge. For example in the DARPA neural networks study of 1988; of the 77 neural network applications investigated only 4 had resulted in field tested systems [Widrow, 1988]. Furthermore, none of these used dedicated neural network hardware. It is our view that this lack of tangible successes can be summarised by the following points: • Most neural applications will be served optimally by fast, generic digital computers . • Dedicated digital neural accelerators have a limited lifetime as "the fastest" , as standard computers develop so rapidly. lEdinburgh Pulse Stream Implemenation of a Learning Oriented Network. 1032 G. JACKSON, A. F. MURRAY • Analog neural VLSI is a niche technology, optimally applied at the interface between the real world and higher-level digital processing. This attitude has some profound implications with respect to the size, nature and constraints we place on new hardware neural designs. After several years of research into hardware neural network implementation, we have now concentrated on the areas in which analog neural network technology has an "edge" over well established digital technology. Within the pulse stream neural network research at the University of Edinburgh, the EPSILON chip's areas of strength can be summarised as: • Analog or digital inputs, digital outputs. • Modest size. • Scaleable and cascadeable design. • Compact, low power. This list points naturally and strongly to problems on the boundary of the real, analog world and digital processing, such as pre-processing/interpretation of analog sensor data. Here a modest neural network can act as an intelligent analog-to-digital converter presenting preprocessed information to its host. We are now engaged in a two pronged approach, whereby development of technology to improve the performance of pulse stream neural network chips is occurring concurrently with a search and development of applications to which this technology can be applied. The key requirements of this technological development are that devices must: • Work directly with analog signals. • Provide a moderate size network. • Have the potential for a fully integrated solution. In working with the above constraints and goals we have developed a new chip, EPSILON II, and a bus based processor card incorporating it. It is our aim to use this system to develop applications. As our first demonstration the EPSILON processor card has been mounted on an autonomous mobile robot. In this case the network utilises a mixture of analog and digital sensor information and performs a mapping between input/sensor space, a mixture of analog and digital signals, and output motor control. 2 THE EPSILON II CHIP The EPSILON II chip has been designed around the requirements of an application based system. It follows on from an earlier generation of pulse stream neural network chip, the EPSILON chip [Murray, 1992]. The EPSILON II chip represents neural states as a pulse encoded signal. These pulse encoded signals have digital signal levels which make them highly immune to noise and ideal for inter and intra-chip communication, facilitating efficient cascading of chips to form larger systems. The EPSILON II chip can take as inputs either pulse encoded signals or analog voltage levels, thus facilitating the fusing of analog and digital data in one system. Internally the chip is analog in nature allowing the synaptic multiplication function to be carried out in compact and efficient analog cells [J ackson, 1994]. Table 1 shows the principal specifications of the EPSILON II chip. The EPSILON II chip is based around a 32x32 synaptic matrix allowing efficient interfacing to digital systems. Several features of the device have been developed specifically for applications based usage. The first of these is a programmable input mode. This Competence Acquisition in an Autonomous Mobile Robot 1033 Table 1: EPSILON II Specifications EPSILON II Chip Specifications No. of state input pins 32 Input modes Analogt PW or PF Input mode programmability Bit programmable No. of state outputs 32 pinned out Output modes PW or PF Digital recovery of analog liP Yes - PW encoded No. of Synapses 1024 Additional autobias synapses 4 per output neuron Weight storage Dynamic Programmable activity voltage Yes Die size 6.9mm x 7mm allows each of the network inputs to be programmed as either a direct analog input or a digital pulse encoded input. We believe that this is vital for application based usage where it is often necessary to fuse real-world analog data with historical or control data generated digitally. The second major feature is a pulse recovery mode. This allows conversion of any analog input into a digital value for direct use by the host system. Both these features are utilised in the robotics application described in section 4 of this paper. 3 EPSILON PROCESSOR CARD The need to embed the EPSILON chip in a processor card is driven by several considerations. FirstlYt working with pulse encoded signals requires substantial processing to interface directly to digital systems. If the neural processor is to be transparent to the host system and is not to become a substantial processing overheadt then all pulse support operations must be carried out independently of the host system. SecondlYt to respond to further chip level advances and allow rapid prototyping of new applications as they emerget a certain amount of flexibility is needed in the system. It is with these points in mind that the design of the flexible EPSILON Processor Card (EPC) was undertaken. 3.1 DESIGN SPECIFICATION The EPC has been designed to meet the following specifications. The card must: • Operate on a conventional digital bus system. • Be transparent to the host processor t that is carry out all the necessary pulse encoding and decoding. • Carry out the refresh operations of the dynamic weights stored on the EPSILON chip. • Generate the ramp waveforms necessary for pulse width coding. • Support the operation of multiple EPCts. • Allow direct input of analog signals. As all data used and generated by the chip is effectively of 8-bit resolutiont the STE bUSt an industry standard 8-bit bUSt was chosen for the bus system. This is also cost 1034 G. JACKSON, A. F. MURRAY effective and allows the use of readily available support cards such as processors, DSP cards and analog and digital signal conditioning cards. To allow the transparency of operation the card must perform a variety of functions. A block diagram indicating these functions is shown in figure 1. FPGA ..................... ........................... __ ........ . · . · . · . · . · . · . 1--""':---1 Pulse to Dig. Conv. Dig. to Pulse Cony. Weight refresh Ctrl. Weight RAM Figure 1: EPSILON Processor Card A substantial amount of digital processing is required by the card, especially in the pulse conversion circuitry. To conform to the Eurocard standard size of the STE specification an FPGA device is used to "absorb" most of the digital logic. A twin mother/daughter board design is also used to isolate sensitive analog circuitry from the digital logic. The use of the FPGA makes the card extremely versatile as it is now easily reconfigurable to adapt to specialist application. The dotted box of figure 1 shows functions implemented by the FPGA device. An on board EPROM can hold multiple FPGA configurations such that the board can be reconfigured "on the fly" . All EPSILON support functions, such as ramp generation, weight refresh, pulse conversion and interface control are carried out on the card. Also the use of the FPGA means that new ideas are easily tested as all digital signal paths go via this device. Thus a card of new functionality can be designed without the need to design a new PCB. 3.2 SPECIALIST BUSES The digital pulse bus is buffered out under control of the FPGA to the neural bus along with two control signals. Handshaking between EPC's is done over these lines to allow the transfer of pulse stream data between processors. This implies that larger networks can be implemented with little or no increase in computation time or overhead. A separate analog bus is included to bring analog inputs directly onto the chip. 4 APPLICATIONS DEVELOPMENT The over-riding reason for the development of the EPC is to allow the easy development of hardware neural network applications. We have already indicated that we believe that this form of neural technology will find its niche where its advantages of direct sensor interface, compactness and cost-effectiveness are of prime importance. As a good and intrinsically interesting example of this genre of applications, we have chosen autonomous mobile robotic control as a first test for EPSILON II. The object of this demonstrator is not to advance the state-of-the-art in robotics. Competence Acquisition in an Autonomous Mobile Robot 1035 Rather it is to demonstrate analog neural VLSI in an appropriate and stimulating context. 4.1 "INSTINCT-RULE" ROBOT The "instinct-rule" robotic control philosophy is based on a software-controlled exemplarfrom the University's Department of Artificial Intelligence [Nehmzow, 1992]. The robot incorporates an EPC which interfaces all the analog sensor signals and provides the programmable neural link between sensor/input space and the motor drive actuators. ~ -"*,::::--,,, '" o en ffi -..,...,\--,,L en a) Controller Architecture. b) Instinct rule robot. Figure 2: "Instinct Rule" Robot The controller architecture is shown in figure 2. The neural network implemented on the EPC is the plastic element that determines the mapping between sensory data and motor actions. The majority of the monitor section is currently implemented on a host processor and monitors the performance of the neural network. It does this by regularly evaluating a set of instinct rules. These rules are simple behaviour based axioms. For example, we use two rules to promote simple obstacle avoidance competence in the robot, as listed in column one of table 2 Table 2: Instinct Rules Simple obstacle avoidance. Wall following l. Keep crash sensors inactive. l. Keep crash sensors inactive. 2. Move forward. 2. Keep side sensors active. 3. Move forward. If an instinct rule is violated the drive selector then chooses the next strongest output (motor action) from the neural network. This action is then performed to see if it relieves the violation. If it does, it is used as targets to train the neural network. If it does not, the next strongest action is tried. The mechanism to accomplish this will be described in more detail in section 4.2. Using this scheme the robot can be initialised with random weights (i.e. no mapping between sensors and motor control) and within a few epochs obtains basic obstacle avoidance competence. It is a relatively easy matter to promote more complex behaviour with the addition of other rules. For example to achieve a wall following behaviour a third 1036 G. JACKSON, A. F. MURRAY rule is introduced as shown in column two of table 2. Navigational tasks can be accomplished with the addition of a "maximise navigational signal" rule. An example of this is a light sensor mounted on the robot producing a behaviour to move towards a light source. Equally, a signal from a more complex, higher level, navigational system could be used. Thus the instinct rule controller handles basic obstacle avoidance competence and motor/sensory interface tasks leaving other resources free for intensive navigational tasks. 4.2 INSTINCT RULE EVALUATION USING SOMATIC TENSION The original instinct rule robot used binary sensor signals and evaluated performance of alternative actions for fixed, and progressively longer, periods of time [Nehmzow, 1992]. With the EPC interfacing directly to analog sensors an improved scheme has been developed. If we sum all sensors onto a neuron with fixed and equal weights we gain a measure of total sensory activity. Let us call this somatic tension as an analogy to biological signal aggregation on the soma. If we have an instinct violation and an alternative action is performed we can monitor this somatic tension to gauge the performance of this action. If tension decreases significantly we continue the action. If it increases significantly we choose an alternative action. If tension remains high and roughly the same, we are in a tight situation, for example say a corner. In this case we perform actions for progressively longer periods continuing to monitor somatic tension for a drop. 4.3 RESULTS AND DISCUSSION The instinct rule robot has been constructed and its performance is comparable with software-controlled predecessors. Unfortunately direct comparisons are not possible due to unavailability of the original exemplars and differing physical characteristics of the robots themselves. In developing the application several observations were made concerning the behaviour of the system that would not have come to light in a simulated environment. In any system including real mechanics and real analog signals, imperfections and noise are present. For example, in a real robot we cannot guarantee that a forward motion directive will result in perfect forward motion due to inherent asymmetries in the system. The instinct rule architecture does not assume a-priori knowledge such as this so behaviour is not affected adversely. This was tested by retarding one drive motor of the robot to give it a bias to one side. In early development, as the monitor was being tuned, the robot showed a tendency to oscillatory motion, thus exhibiting undesirable behaviour that satisfies its instincts. It could, for example, oscillate back and forth at a corner. In a simulated environment this continues indefinitely. However, with real mechanics and noisy analog sensors the robot breaks out of this undesirable behaviour. These observations strengthen the arguments for hardware development aimed at embedded systems. The robot application is but an example of the different, and often surprising conditions that pertain in a "real" system. If neural networks are to find applications in real-world, low-cost and analog-interface applications, these are the conditions we must deal with, and appropriate, analog hardware is the optimal medium for a solution. Competence Acquisition in an Autonomous Mobile Robot 1037 5 CONCLUSIONS This paper has described pulse stream neural networks that have been developed to a system level to aid development of applications. We have therefore defined areas of strengths of this technology along with suggestions of where this is best applied. The strengths of this system include: 1. Direct interfacing to analog signals. 2. The ability to fuse direct analog sensor data with digital sensor data processed elsewhere in the system. 3. Distributed processing. Several EPC's may be embedded in a system to allow multiple networks and/or multi layer networks. 4. The EPC represents a flexible system level development environment. It is easily reconfigured for new applications or improved chip technology. 5. The EPC requires very little computational overhead from the host system and can operate independently if needed. A demonstration application of an instinct rule robot has been presented highlighting the use of neural networks as an interface between real-world analog signals and digital control. In conclusion we believe that the immediate future of neural analog VLSI is in small applications based systems that interface directly to the real-world. We see this as the primary niche area where analog VLSI neural networks will replace conventional digital systems. Acknow ledgements Thanks are due to Ulrich Nehmzow, University of Manchester, for discussions and information on the instinct-rule controller and the loan of his original robot - Alder. References [Caudell, 1990] Caudell, M. and Butler, C. (1990). Naturally Intelligent Systems. MIT Press, Cambridge, Ma. [Jackson, 1994] Jackson, G., Hamilton, A., and Murray, A. F. (1994). Pulse stream VLSI neural systems: into robotics. In Proceedings ISCAS'94, volume 6, pages 375-378. IEEE Press. [Maren, 1990] Maren, A., Harston, C., and Pap, R. (1990). Handbook of Neural Computing Applications. Academic Press, San Diego, Ca. [Murray,1992] Murray, A. F., Baxter, D. J., Churcher, S., Hamilton, A., Reekie, H. M., and Tarassenko, L. (1992). The Edinburgh pulse stream implementation of a learning-oriented network (EPSILON) chip. In Neural Information Processing Systems (NIPS) Conference. [Nehmzow, 1992] Nehmzow, U. (1992). Experiments in Competence Acquisition for Autonomous Mobile Robots. PhD thesis, University of Edinburgh. [Widrow, 1988] Widrow, B. (1988). DARPA Neural Network Study. AFCEA International Press.
|
1995
|
72
|
1,120
|
Gaussian Processes for Regression Christopher K. I. Williams Neural Computing Research Group Aston University Birmingham B4 7ET, UK c.k.i.williams~aston.ac.uk Carl Edward Rasmussen Department of Computer ,Science University of Toronto Toronto, ONT, M5S lA4, Canada carl~cs.toronto.edu Abstract The Bayesian analysis of neural networks is difficult because a simple prior over weights implies a complex prior distribution over functions. In this paper we investigate the use of Gaussian process priors over functions, which permit the predictive Bayesian analysis for fixed values of hyperparameters to be carried out exactly using matrix operations. Two methods, using optimization and averaging (via Hybrid Monte Carlo) over hyperparameters have been tested on a number of challenging problems and have produced excellent results. 1 INTRODUCTION In the Bayesian approach to neural networks a prior distribution over the weights induces a prior distribution over functions. This prior is combined with a noise model, which specifies the probability of observing the targets t given function values y, to yield a posterior over functions which can then be used for predictions. For neural networks the prior over functions has a complex form which means that implementations must either make approximations (e.g. MacKay, 1992) or use Monte Carlo approaches to evaluating integrals (Neal, 1993). As Neal (1995) has argued, there is no reason to believe that, for real-world problems, neural network models should be limited to nets containing only a "small" number of hidden units. He has shown that it is sensible to consider a limit where the number of hidden units in a net tends to infinity, and that good predictions can be obtained from such models using the Bayesian machinery. He has also shown that a large class of neural network models will converge to a Gaussian process prior over functions in the limit of an infinite number of hidden units. In this paper we use Gaussian processes specified parametrically for regression problems. The advantage of the Gaussian process formulation is that the combination of Gaussian Processes for Regression 515 the prior and noise models can be carried out exactly using matrix operations. We also show how the hyperparameters which control the form of the Gaussian process can be estimated from the data, using either a maximum likelihood or Bayesian approach, and that this leads to a form of "Automatic Relevance Determination" (Mackay 1993j Neal 1995). 2 PREDICTION WITH GAUSSIAN PROCESSES A stochastic process is a collection of random variables {Y (x) Ix EX} indexed by a set X. In our case X will be the input space with dimension d, the number of irlputs. The stochastic process is specified by giving the probability distribution for every finite subset of variables Y(x(1)), ... , Y(x(k)) in a consistent manner. A Gaussian process is a stochastic process which can be fully specified by its mean function J.1.(:x:) = E[Y(x)] and its covariance function C(X ,X/) = E[(Y(x) - J.1.(x))(Y(x/)J.1.( Xl))]; any finite set of points will have a joint multivariate Gaussian distribution. Below we consider Gaussian processes which have J.1.( x) == O. In section 2.1 we will show how to parameterise covariances using hyperparametersj for now we consider the form of the covariance C as given. The training data consists of n pairs of inputs and targets {( xCi) , t(i)) , i = 1 .. . n} . The input vector for a test case is denoted x (with no superscript). The inputs are d-dimensional Xl, . .. , Xd and the targets are scalar. The predictive distribution for a test case x is obtained from the n + 1 dimensional joint Gaussian distribution for the outputs of the n training cases and the test case, by conditioning on the observed targets in the training set. This procedure is illustrated in Figure 1, for the case where there is one training point and one test point. In general, the predictive distribution is Gaussian with mean and variance kT (x)K- 1t C(x,x) - kT(x)K- 1k(x), (1) (2) where k(x) = (C(x, x(1)), ... , C(x, x(n))f , K is the covariance matrix for the training cases Kij = C(x(i), x(j)), and t = (t(l), ... , t(n))T . The matrix inversion step in equations (1) and (2) implies that the algorithm has O( n3 ) time complexity (if standard methods of matrix inversion are employed); for a few hundred data points this is certainly feasible on workstation computers, although for larger problems some iterative methods or approximations may be needed. 2.1 PARAMETERIZING THE COVARIANCE FUNCTION There are many choices of covariance functions which may be reasonable. Formally, we are required to specify functions which will generate a non-negative definite covariance matrix for any set of points (x(1 ), ... , x(k )). From a modelling point of view we wish to specify covariances so that points with nearby inputs will give rise to similar predictions. We find that the following covariance function works well: d Vo exp{ -t L WI(x~i) x~j))2} 1=1 d +ao + a1 Lx~i)x~j) + V18(i ,j), 1=1 (3) 516 / / y / / y1 c. K. I. WILLIAMS, C. E. RASMUSSEN y p(y) Figure 1: An illustration of prediction using a Gaussian process. There is one training case (x(1), t(1)) and one test case for which we wish to predict y. The ellipse in the lefthand plot is the one standard deviation contour plot of the joint distribution of Yl and y. The dotted line represents an observation Yl = t(1). In the right-hand plot we see the distribution of the output for the test case, obtained by conditioning on the observed target. The y axes have the same scale in both plots. where (} = log(vo, V1, W1, . . . , Wd, ao, ad plays the role of hyperparameters1. We define the hyperparameters to be the log of the variables in equation (4) since these are positive scale-parameters. The covariance function is made up of three parts; the first term, a linear regression term (involving ao and aI) and a noise term V1b(i, j). The first term expresses the idea that cases with nearby inputs will have highly correlated outputs; the WI parameters allow a different distance measure for each input dimension. For irrelevant inputs, the corresponding WI will become small, and the model will ignore that input. This is closely related to the Automatic Relevance Determination (ARD) idea of MacKay and Neal (MacKay, 1993; Neal 1995). The Vo variable gives the overall scale of the local correlations. This covariance function is valid for all input dimensionalities as compared to splines, where the integrated squared mth derivative is only a valid regularizer for 2m > d (see Wahba, 1990). ao and a1 are variables controlling the scale the of bias and linear contributions to the covariance. The last term accounts for the noise on the data; VI is the variance of the noise. Given a covariance function, the log likelihood of the training data is given by 1= - ~ logdet I< ~tT I<-lt - !!.log27r. (4) 222 In section 3 we will discuss how the hyperparameters III C can be adapted, in response to the training data. 2.2 RELATIONSHIP TO PREVIOUS WORK The Gaussian process view provides a unifying framework for many regression methods. ARMA models used in time series analysis and spline smoothing (e.g. Wahba, 1990 and earlier references therein) correspond to Gaussian process prediction with 1 We call () the hyperparameters as they correspond closely to hyperparameters in neural networks; in effect the weights have been integrated out exactly. Gaussian Processes for Regression 517 a particular choice of covariance function2 . Gaussian processes have also been used in the geostatistics field (e.g. Cressie, 1993), and are known there as "kriging", but this literature has concentrated on the case where the input space is two or three dimensional, rather than considering more general input spaces. This work is similar to Regularization Networks (Poggio and Girosi, 1990; Girosi, Jones and Poggio, 1995), except that their derivation uses a smoothness functional rather than the equivalent covariance function. Poggio et al suggested that the hyperparameters be set by cross-validation. The main contributions of this paper are to emphasize that a maximum likelihood solution for 8 is possible, to recognize the connections to ARD and to use the Hybrid Monte Carlo method in the Bayesian treatment (see section 3). 3 TRAINING A GAUSSIAN PROCESS The partial derivative of the log likelihood of the training data I with respect to all the hyperparameters can be computed using matrix operations, and takes time O( n 3 ) . In this section we present two methods which can be used to adapt the hyperparameters using these derivatives. 3.1 MAXIMUM LIKELIHOOD In a maximum likelihood framework, we adjust the hyperparameters so as to maximize that likelihood of the training data. We initialize the hyperparameters to random values (in a reasonable range) and then use an iterative method, for example conjugate gradient, to search for optimal values of the hyperparameters. Since there are only a small number of hyperparameters (d + 4) a relatively small number of iterations are usually sufficient for convergence. However, we have found that this approach is sometimes susceptible to local minima, so it is advisable to try a number of random starting positions in hyperparameter space. 3.2 INTEGRATION VIA HYBRID MONTE CARLO According to the Bayesian formalism, we should start with a prior distribution P( 8) over the hyperparameters which is modified using the training data D to produce a posterior distribution P(8ID). To make predictions we then integrate over the posterior; for example, the predicted mean y( x) for test input x is given by y(x) = J Y8(x)P(8I D)d8 (5) where Y8( x) is the predicted mean (as given by equation 1) for a particular value of 8. It is not feasible to do this integration analytically, but the Markov chain Monte Carlo method of Hybrid Monte Carlo (HMC) (Duane et ai, 1987) seems promising for this application. We assign broad Gaussians priors to the hyperparameters, and use Hybrid Monte Carlo to give us samples from the posterior. HMC works by creating a fictitious dynamical system in which the hyperparameters are regarded as position variables, and augmenting these with momentum variables p. The purpose of the dynamical system is to give the hyperparameters "inertia" so that random-walk behaviour in 8-space can be avoided. The total energy, H, of the system is the sum of the kinetic energy, J{, (a function of the momenta) and the potential energy, E. The potential energy is defined such that p(8ID) ex: exp(-E). We sample from the joint distribution for 8 and p given by p(8,p) ex: exp(-E2Technically splines require generalized covariance functions. 518 C. K. I. WILUAMS, C. E. RASMUSSEN I<); the marginal of this distribution for 8 is the required posterior. A sample of hyperparameters from the posterior can therefore be obtained by simply ignoring the momenta. Sampling from the joint distribution is achieved by two steps: (i) finding new points in phase space with near-identical energies H by simulating the dynamical system using a discretised approximation to Hamiltonian dynamics, and (ii) changing the energy H by doing Gibbs sampling for the momentum variables. Hamiltonian Dynamics Hamilton's first order differential equations for H are approximated by a discrete step (specifically using the leapfrog method). The derivatives of the likelihood (equation 4) enter through the derivative of the potential energy. This proposed state is then accepted or rejected using the Metropolis rule depending on the final energy H* (which is not necessarily equal to the initial energy H because of the discretization). The same step size c is used for all hyperparameters, and should be as large as possible while keeping the rejection rate low. Gibbs Sampling for Momentum Variables The momentum variables are updated using a modified version of Gibbs sampling, thereby allowing the energy H to change. A "persistence" of 0.95 is used; the new value of the momentum is a weighted sum of the previous value (with weight 0.95) and the value obtained by Gibbs sampling (weight (1 - 0.952)1/2). With this form of persistence, the momenta change approximately twenty times more slowly, thus increasing the "inertia" of the hyperparameters, so as to further help in avoiding random walks. Larger values of the persistence will further increase the inertia, but reduce the rate of exploration of H . Practical Details The priors over hyperparameters are set to be Gaussian with a mean of -3 and a standard deviation of 3. In all our simulations a step size c = 0.05 produced a very low rejection rate « 1 %). The hyperparameters corresponding to V1 and to the WI ' S were initialised to -2 and the rest to O. To apply the method we first rescale the inputs and outputs so that they have mean of zero and a variance of one on the training set. The sampling procedure is run for the desired amount of time, saving the values of the hyperparameters 200 times during the last two-thirds of the run. The first third of the run is discarded; this "burn-in" is intended to give the hyperparameters time to come close to their equilibrium distribution. The predictive distribution is then a mixture of 200 Gaussians. For a squared error loss, we use the mean of this distribution as a point estimate. The width of the predictive distribution tells us the uncertainty of the prediction. 4 EXPERIMENTAL RESULTS We report the results of prediction with Gaussian process on (i) a modified version of MacKay's robot arm problem and (ii) five real-world data sets. 4.1 THE ROBOT ARM PROBLEM We consider a version of MacKay's robot arm problem introduced by Neal (1995). The standard robot arm problem is concerned with the mappings Y1 = r1 cos Xl + r2 COS(X1 + X2) Y2 = r1 sin Xl + r2 sin(x1 + X2) (6) Gaussian Processes for Regression 519 Method No. of inputs sum squared test error Gaussian process 2 1.126 Gaussian process 6 1.138 MacKay 2 1.146 Neal 2 1.094 Neal 6 1.098 Table 1: Results on the robot arm task. The bottom three lines of data were obtained from Neal (1995) . The MacKay result is the test error for the net with highest "evidence". The data was generated by picking Xl uniformly from [-1.932, -0.453] and [0.453, 1.932] and picking X2 uniformly from [0.534, 3.142]. Neal added four further inputs, two of which were copies of Xl and X2 corrupted by additive Gaussian noise of standard deviation 0.02, and two further irrelevant Gaussian-noise inputs with zero mean and unit variance. Independent zero-mean Gaussian noise of variance 0.0025 was then added to the outputs YI and Y2 . We used the same datasets as Neal and MacKay, with 200 examples in the training set and 200 in the test set. The theory described in section 2 deals only with the prediction of a scalar quantity Y , so predictors were constructed for the two outputs separately, although a joint prediction is possible within the Gaussian process framework (see co-kriging, §3.2.3 in Cressie, 1993). Two experiments were conducted, the first using only the two "true" inputs, and the second one using all six inputs. In this section we report results using maximum likelihood training; similar results were obtained with HMC. The log( v),s and loge w )'s were all initialized to values chosen uniformly from [-3.0, 0.0], and were adapted separately for the prediction of YI and Y2 (in these early experiments the linear regression terms in the covariance function involving aa and al were not present) . The conjugate gradient search algorithm was allowed to run for 100 iterations, by which time the likelihood was changing very slowly. Results are reported for the run which gave the highest likelihood of the training data, although in fact all runs performed very similarly. The results are shown in Table 1 and are encouraging, as they indicate that the Gaussian process approach is giving very similar performance to two well-respected techniques. All of the methods obtain a level of performance which is quite close to the theoretical minimum error level of 1.0 . ...Jt is interesting to look at the values of the w's obtained after the optimization; for the Y2 task the values were 0.243,0.237,0.0639,7.0 x 10-4 , 2.32 x 10- 6 ,1.70 x 10- 6 , and Va and VI were 7.5278 and 0.0022 respectively. The w values show nicely that the first two inputs are the most important, followed by the corrupted inputs and then the irrelevant inputs. During training the irrelevant inputs are detected quite quickly, but the w's for the corrupted inputs shrink more slowly, implying that the input noise has relatively little effect on the likelihood. 4.2 FIVE REAL-WORLD PROBLEMS Gaussian Processes as described above were compared to several other regression algorithms on five real-world data sets in (Rasmussen, 1996; in this volume). The data sets had between 80 and 256 training examples, and the input dimension ranged from 6 to 16. The length of the HMC sampling for the Gaussian processes was from 7.5 minutes for the smallest training set size up to 1 hour for the largest ones on a R4400 machine. The results rank the methods in the order (lowest error first) a full-blown Bayesian treatment of neural networks using HMC, Gaussian 520 C. K. I. WILLIAMS, C. E. RASMUSSEN processes, ensembles of neural networks trained using cross validation and weight decay, the Evidence framework for neural networks (MacKay, 1992), and MARS. We are currently working on assessing the statistical significance of this ordering. 5 DISCUSSION We have presented the method of regression with Gaussian processes, and shown that it performs well on a suite of real-world problems. We have also conducted some experiments on the approximation of neural nets (with a finite number of hidden units) by Gaussian processes, although space limitations do not allow these to be described here. Some other directions currently under investigation include (i) the use of Gaussian processes for classification problems by softmaxing the outputs of k regression surfaces (for a k-class classification problem), (ii) using non-stationary covariance functions, so that C(x, Xl) f:- C(lx - XII) and (iii) using a covariance function containing a sum of two or more terms of the form given in line 1 of equation 3. We hope to make our code for Gaussian process prediction publically available in the near future. Check http://www.cs.utoronto.ca/neuron/delve/delve.html for details. Acknowledgements We thank Radford Neal for many useful discussions, David MacKay for generously providing the robot arm data used in this paper, and Chris Bishop, Peter Dayan, Radford Neal and Huaiyu Zhu for comments on earlier drafts. CW was partially supported by EPSRC grant GRjJ75425. References Cressie, N. A. C. (1993). Statistics for Spatial Data. Wiley. Duane, S., Kennedy, A. D., Pendleton, B. J., and Roweth, D. (1987). Hybrid Monte Carlo. Physics Letters B, 195:216-222. Girosi, F., Jones, M., and Poggio, T. (1995). Regularization Theory and Neural Networks Architectures. Neural Computation, 7(2):219-269. MacKay, D. J. C. (1992). A Practical Bayesian Framework for Backpropagation Networks. Neural Computation, 4(3):448-472. MacKay, D. J. C. (1993). Bayesian Methods for Backpropagation Networks. In van Hemmen, J. L., Domany, E., and Schulten, K., editors, Models of Neural Networks II. Springer. Neal, R. M. (1993). Bayesian Learning via Stochastic Dynamics. In Hanson, S. J., Cowan, J. D., and Giles, C. L., editors, Neural Information Processing Systems, Vol. 5, pages 475-482. Morgan Kaufmann, San Mateo, CA. Neal, R. M. (1995). Bayesian Learning for Neural Networks. PhD thesis, Dept. of Computer Science, University of Toronto. Poggio, T. and Girosi, F. (1990). Networks for approximation and learning. Proceedings of IEEE, 78:1481-1497. Rasmussen, C. E. (1996). A Practical Monte Carlo Implementation of Bayesian Learning. In Touretzky, D. S., Mozer, M. C., and Hasselmo, M. E., editors, Advances in Neural Information Processing Systems 8. MIT Press. Wahba, G. (1990). Spline Models for Observational Data. Society for Industrial and Applied Mathematics. CBMS-NSF Regional Conference series in applied mathematics.
|
1995
|
73
|
1,121
|
Reorganisation of Somatosensory Cortex after Tactile Training Rasmus S. Petersen John G. Taylor Centre for Neural Networks, King's College London Strand, London WC2R 2LS, UK Abstract Topographic maps in primary areas of mammalian cerebral cortex reorganise as a result of behavioural training. The nature of this reorganisation seems consistent with the behaviour of competitive neural networks, as has been demonstrated in the past by computer simulation. We model tactile training on the hand representation in primate somatosensory cortex, using the Neural Field Theory of Amari and his colleagues. Expressions for changes in both receptive field size and magnification factor are derived, which are consistent with owl monkey experiments and make a prediction which goes beyond them. 1. INTRODUCTION The primary cortical areas of mammals are now known to be plastic throughout life; reviewed recently by Kaas(1995). The problem of how and why the underlying learning processes work is an exciting one, for which neural network modelling appears well suited. In this contribution, we model the long-term effects of tactile training (Jenkins et ai, 1990) on the functional organisation of monkey primary somatosensory cortex, by perturbing a topographic net (Takeuchi and Amari, 1979). 1.1 ADAPTATION IN ADULT SOMATOSENSORY CORTEX Light touch activates skin receptors which in primates are mapped, largely topographically, in area 3b. In a series of papers, Merzenich and colleagues describe how area 3b becomes reorganised following peripheral nerve damage (Merzenich et ai, 1983a; 1983b) or digit amputation (Merzenich et ai, 1984). The underlying learning processes may also explain the phenomenon of phantom limb "telescoping" (Haber, 1955). Recent advances in brain scanning are beginning to make them observable even in the human brain (Mogilner et ai, 1993). 1.2 ADAPTATION ASSOCIATED WITH TACTILE TRAINING Jenkins et al trained owl monkeys to maintain contact with a rotating disk. The apparatus was arranged so that success eventually involved touching the disk with only the digit tips. Hence these regions received selective stimulation. Some time after training had been completed electro-physiological recordings were made from area 3b. These revealed an increase in Magnification Factor (MF) for the stimulated skin and a decrease in Reorganization of Somatosensory Cortex after Tactile Training 83 the size of Receptive Fields (RFs) for that region. The net territory gained for light touch of the digit tips came from area 3a and/or the face region of area 3b, but details of any changes in these representations were not reported. 2. THEORETICAL FRAMEWORK 2.1 PREVIOUS WORK Takeuchi and Amari(1979), Ritter and Schulten(1986), Pearson et al(1987) and Grajski and Merzenich( 1990) have all modelled amputationldenervation by computer simulation of competitive neural networks with various Hebbian weight dynamics. Grajski and Merzenich(1990) also modelled the data of Jenkins et al. We build on this research within the Neural Field Theory framework (Amari, 1977; Takeuchi and Amari, 1979; Amari, 1980) of the Neural Activity Model of Willshaw and von der Malsburg(1976). 2.2 NEURAL ACTIVITY MODEL Consider a "cortical" network of simple, laterally connected neurons. Neurons sum inputs linearly and output a sigmoidal function of this sum. The lateral connections are excitatory at short distances and inhibitory at longer ones. Such a network is competitive: the steady state consists of blobs of activity centred around those neurons locally receiving the greatest afferent input (Amari, 1977). The range of the competition is limited by the range of the lateral inhibition. Suppose now that the afferent synapses adapt in a Hebbian manner to stimuli that are localised in the sensory array; the lateral ones are fixed. Willshaw and von der Malsburg(1976) showed by computer simulation that this network is able to form a topographic map of the sensory array. Takeuchi and Amari( 1979) amended the WillshawMalsburg model slightly: neurons possess an adaptive firing threshold in order to prevent synaptic weight explosion, rather than the more usual mechanism of weight normalisation. They proved that a topographic mapping is stable under certain conditions. 2.3 TAKEUCHI-AMARI THEORY Consider a one-dimensional model. The membrane dynamics are: au(~y,t) = -u(x,y,t)+ f s(x,y' ,t)a(y- y')dy'so(x,t)ao + f w(x-x')f[u(x' ,y,t)]dx'-h (1) Here u(x,y,t) is the membrane potential at time I for point x when a stimulus centred at y is being presented; h is a positive resting potential; w(z) is the lateral inhibitory weight between two points in the neural field separated by a distance z - positive for small Izl and negative for larger Izl; s(x,y,t) is the excitatory synaptic weight from y to x at time I and sr/X,I) is an inhibitory weight from a tonically active inhibitory input aD to x at time t - it is the adaptive firing threshold. f[u] is a binary threshold function that maps positive membrane potentials to 1 and non-positive ones to O. Idealised, point-like stimuli are assumed, which "spread out" somewhat on the sensory surface or subcortically. The spreading process is assumed to be independent of y and is described in the same coordinates. It is represented by the function a(y-y'), which describes the effect of a point input at y spreading to the point y'. This is a decreasing, positive, symmetric function of Iy-y'l. With this type of input, the steady-state activity of the network is a single blob, localised around the neuron with maximum afferent input. 84 R. S. PETERSEN, J. O. TAYLOR The afferent synaptic weights adapt in a leaky Hebbian manner but with a time constant much larger than that of the membrane dynamics (1). Effectively this means that learning occurs on the steady state of the membrane dynamics. The following averaged weight dynamics can be justified (Takeuchi and Amari, 1979; Geman 1979): as( x, y, t) ( J) ( [A )] at =-s x,y,t)+b p(y' a Y-Y')f u(x,y' dy' aso(~y,t) =-so(x,y,t)+b' aoJ p(y')f[u(x,y')]dy' (2) where r1(x,y') is the steady-state of the membrane dynamics at x given a stimulus at y' and p(y') is the probability of a stimulus at y '; b, b' are constants. Empirically, the "classical" Receptive Field (RF) of a neuron is defined as the region of the input field within which localised stimulation causes change in its activity. This concept can be modelled in neural field theory as: the RF of a neuron at x is the portion of the input field within which a stimulus evokes a positive membrane potential (inhibitory RFs are not considered). If the neural field is a continuous map of the sensory surface then the RF of a neuron is fully described by its two borders rdx), rix), defined formally: i = 1,2 (3) which are illustrated in figure 1. Let RF size and RF position be denoted respectively by the functions rex) and m(x), which represent experimentally measurable quantities. In terms of the border functions they can be expressed: r(x) = r2 (x) - r1 (x) m(x) = -} (rl {x} + r2 (x)) y ~--------------------------- x (4) Figure 1: RF boundaries as a function of position in the neural field, for a topographically ordered network. Only the region in-between rdx) and rix) has positive steadystate membrane potential r1(x,y). rdx) and rix) are defined by the condition r1(x,r;(x))=O i=J,2. for Using (1), (2) and the definition (3), Takeuchi and Amari(1979) derived dynamical equations for the change in RF borders due to learning. In the case of uniform stimulus probability, they found solutions for the steady-state RF border functions. With periodic boundary conditions, the basic solution is a linear map with constant RF size: Reorganization of Somatosensory Cortex after Tactile Training r(x) = ro = const m(x) = px ++ro uni ( ) rl x = px r~tni (x) = px+ ro 85 (5) This means that both RF size and activity blob size are uniform across the network and that RF position m(x) is a linear function of network location. (The value of p is determined by boundary conditions; ro is then determined from the joint equilibrium of (I), (2». The inverse of the RF position function, denoted by m-l(y), is the centre of the cortical active region caused by a stimulus centred at y. The change in m-l(y) over a unit interval in the input field is, by empirical definition, the cortical magnification factor (MF). Here we model MF as the rate of change of m-l(y). The MF for the system described by (5) is: d _I () -I -m y =p dy (6) 3. ANALYSIS OF TACTILE TRAINING 3.1 TRAINING MODEL AND ASSUMPTIONS Jenkins et aI's training sessions caused an increase in the relative frequency of stimulation to the finger tips, and hence a decrease in relative frequency of stimulation elsewhere. Over a long time, we can express this fact as a localised change in stimulus probability (figure 2). (This is not sufficient to cause cortical reorganisation - Recanzone et al( 1992) showed that attention to the stimulation is vital. We consider only attended stimulation in this model). To account for such data it is clearly necessary to analyse non-uniform stimulus probabilities, which demands extending the results of Takeuchi and Amari. Unfortunately, it seems to be hard to obtain general results. However, a perturbation analysis around the uniform probability solution (5) is possible. To proceed in this way, we must be able to assume that the change in the stimulus probability density function away from uniformity is small. This reasoning is expressed by the following equation: p(y) = Po + E p(y) (7) where pry) is the new stimulus probability in terms of the uniform one and a perturbation due to training: E is a small constant. The effect of the perturbation is to ease the weight dynamics (2) away from the solution (5) to a new steady-state. Our goal is to discover the effect of this on the RF border functions, and hence for RF size and MF. p(y) o y Figure 2: The type of change in stimulus probability density that we assume to model the effects of behavioural training. 86 R. S. PETERSEN, J. G. TAYLOR 3.2 PERTURBATION ANALYSIS 3.2.1 General Case For a small enough perturbation, the effect on the RF borders and on the activity blob size ought also to be small. We consider effects to first order in E, seeking new solutions of the form: i = 1,2 ,{x} = r; {x} - ~ {x} m{x} = +(~ (X}+'2 (x}) (8) where the superscript peT denotes the new, perturbed equilibrium and uni denotes the unperturbed, uniform probability equilibrium. Using (1) and (2) in (3) for the post-training RF borders, expanding to first order in E, a pair of difference equations may be obtained for the changes in RF borders. It is convenient to define the following terms: ro rt '(x) At (x) = J p(y+ px)k(y)dy-b' a~ J p(y)dy o r,"no (x) o r;-n' (x) A2 {x} = J p(y + px + TO )k(y)dy - b' a~ J p(y)dy k(y) = b J a(y - y' )a(y' )dy' (9) B = b' a~p() -k(ro)po > 0 C= w(p-tTo)p-t <0 where the signs of Band C arise due to stability conditions (Amari, 1977; Takeuchi and Amari, 1979). In terms of RF size and RF position (4), the general result is: B~2 ,(X} = ~(~ + I)At (x) - M2 (x) BC~2m{X) = (B- C -+ C~)(~+ I}At (x) + (C- B++(C -2B)~)A2 (x) (10) where ~ is the difference operator: ~ f{ x) = f( x + p - t To) - f( x) (11 ) 3.2.2 Particular Case The second order difference equations (l0) are rather opaque. This is partly due to coupling in y caused by the auto-correlation function key): (10) simplifies considerably if very narrow stimuli are assumed - a(y)=O(y) (see also Amari, 1980). For periodic boundary conditions: (12) where: Reorganization of Somatosensory Cortex after Tactile Training 87 m -I P(W (y) = m -I pre (y) + Em -I (y) = p-l(y_+ro)+Em-l(y) (13) and we have used the crude approximation: d _() 1 ( 1 _I ) dx m x "" t;: ~m x - 2" P ro (14) which demands smoothness on the scale of 10. However, for perturbations like that sketched in figure 2, this is sufficient to tell us about the constant regions of MF. (We would not expect to be able to model the data in the transition region in any case, as its form is too dependent upon fine detail of the model). Our results (12) show that the change in RF size of a neuron is simply minus the total change in stimulus probability over its RF. Hence RF size decreases where p(y) increases and vice versa. Conversely, the change in MF at a given stimulus location is roughly the local average change in stimulus probability there. Note that changes in RF size correlate inversely with changes in MF. Figure 3 is a sketch of these results for the perturbation of figure 2. MF RF o \ I I o L.J y Figure 3: Results of perturbation analysis for how behavioural training (figure 2) changes RF size and MF respectively, in the case where stimulus width can be neglected. For MF - due to the approximation (14) - predictions do not apply near the transitions. 4. DISCUSSION Equations (12) are the results of our model for RF size and MF after area 3b has fully adapted to the behavioural task, in the case where stimulus width can be neglected. They appear to be fully consistent with the data of Jenkins et al described above: RF size decreases in the region of cortex selective for the stimulated body part and the MF for this body part increases. Our analysis also makes a specific prediction that goes beyond Jenkins et aI's data, directly due to the inverse relationship between changes in RF size and those in MF. Within the regions that surrender territory to the entrained finger tips (sometimes the face region), for which MF decreases, RF sizes should increase. Surprisingly perhaps, these changes in RF size are not due to adaptation of the afferent weights s(x,y). The changes are rather due to the adaptive threshold term six). This point will be discussed more fully elsewhere. A limitation of our analysis is the assumption that the change in stimulus probability is in some sense small. Such an approximation may be reasonable for behavioural training but seems less so as regards important experimental protocols like amputation or denervation. Evidently a more general analysis would be highly desirable. 88 R. S. PETERSEN,J. O. TAYLOR 5. CONCLUSION We have analysed a system with three interacting features: lateral inhibitory interactions; Hebbian adaptivity of afferent synapses and an adaptive firing threshold. Our results indicate that such a system can account for the data of Jenkins et aI, concerning the response of adult somatosensory cortex to the changing environmental demands imposed by tactile training. The analysis also brings out a prediction of the model, that may be testable. Acknowledgements RSP is very grateful for a travel stipend from the NIPS Foundation and for a Nick Hughes bursary from the School of Physical Sciences and Engineering, King's College London, that enabled him to participate in the conference. References Amari S. (1977) BioI. Cybern. 2777-87 Amari S. (1980) Bull. Math. Biology 42339-364 Geman S. (1979) SIAM 1. App. Math. 36 86-105 Grajski K.A., Merzenich M.M. (1990) in Neural Information Processing Systems 2 Touretzky D.S. (Ed) 52-59 HaberW.B. (1955)1. Psychol. 40115-123 Jenkins W.M., Merzenich M.M., Ochs M.T., Allard T., Gufc-Robles E. (1990) 1. Neurophysiol. 63 82-104 Kaas J.H. (1995) in The Cognitive Neurosciences Gazzaniga M.S. (Ed ic) 51-71 Merzenich M.M., Kaas J.H., Wall J.T., Nelson R.J., Sur M., Felleman DJ. (1983a) Neuroscience 8 35-55 Merzenich M.M., Kaas J.H., Wall J.T., Sur M., Nelson R.I., Felleman DJ. (1983b) Neuroscience 10639-665 Merzenich M.M., Nelson R.I., Stryker M.P., Cynader M.S., Schoppmann A., Zook J.M. (1984) 1. Compo Neural. 224591-605 Mogilner A., Grossman A.T., Ribrary V., Joliot M., Vol mann J., Rapaport D., Beasley R., L1inas R. (1993) Proc. Natl. Acad. Sci. USA 90 3593-3597 Pearson J.e., Finkel L.H., Edelman G.M. (1987) 1. Neurosci. 124209-4223 Recanzone G.H., Merzenich M.M., Jenkins W.M., Grajski K.A., Dinse H.R. (1992) 1. Neurophysiol. 67 1031-1056 Ritter H., Schulten K. (1986) BioI. Cybern. 5499-106 Takeuchi A., Amari S. (1979) BioI. Cybern. 35 63-72 Willshaw DJ., von der Malsburg e. (1976) Proc. R. Soc. Lond. B194 203-243
|
1995
|
74
|
1,122
|
Prediction of Beta Sheets in Proteins Anders Krogh The Sanger Centre Hinxton, Carobs CBIO IRQ, UK. Email: krogh@sanger.ac. uk S~ren Kamaric Riis Electronics Institute, Building 349 Technical University of Denmark 2800 Lyngby, Denmark Email: riis@ei.dtu.dk Abstract Most current methods for prediction of protein secondary structure use a small window of the protein sequence to predict the structure of the central amino acid. We describe a new method for prediction of the non-local structure called ,8-sheet, which consists of two or more ,8-strands that are connected by hydrogen bonds. Since,8strands are often widely separated in the protein chain, a network with two windows is introduced. After training on a set of proteins the network predicts the sheets well, but there are many false positives. By using a global energy function the ,8-sheet prediction is combined with a local prediction of the three secondary structures a-helix, ,8-strand and coil. The energy function is minimized using simulated annealing to give a final prediction. 1 INTRODUCTION Proteins are long sequences of amino acids. There are 20 different amino acids with varying chemical properties, e. g. , some are hydrophobic (dislikes water) and some are hydrophilic [1]. It is convenient to represent each amino acid by a letter and the sequence of amino acids in a protein (the primary structure) can be written as a string with a typical length of 100 to 500 letters. A protein chain folds back on itself, and the resulting 3D structure (the tertiary structure) is highly correlated to the function of the protein. The prediction of the 3D structure from the primary structure is one of the long-standing unsolved problems in molecular biology. As an important step on the way a lot of work has been devoted to predicting the local conformation of the protein chain, which is called the secondary structure. Neural network methods are currently the most successful for predicting secondary structure. The approach was pioneered by Qian and Sejnowski [2] and Bohr et al. [3], but later extended in various ways, see e.g. [4] for an overview. In most of this work, only the two regular secondary structure elements a-helix and ,8-strand are being distinguished, and everything else is labeled coil. Thus, the methods based 918 A. KROGH, S. K. RIIS H-\ H-\ t !" o=c ,,=c {-~ {-H \ ' /=0 HH-\ f fa o=c /o=c ~-o ~-: Figure 1: Left: Anti-parallel,B-sheet. The vertical lines correspond to the backbone of the protein. An amino acid consists of N-Ca-C and a side chain on the Ca that is not shown (the 20 amino acids are distinguished by different side chains). In the anti-parallel sheet the directions of the strands alternate, which is here indicated quite explicitly by showing the middle strand up-side down. The H-bonds between the strands are shown by 11111111. A sheet has two or more strands, here the antiparallel sheet is shown with three strands. Right: Parallel ,B-sheet consisting of two strands. on a local window of amino acids give a three-state prediction of the secondary structure of the central amino acid in the window. Current predictions of secondary structure based on single sequences as input have accuracies of about 65-66%. It is widely believed that this accuracy is close to the limit of what can be done from a local window (using only single sequences as input) [5], because interactions between amino acids far apart in the protein chain are important to the structure. A good example of such non-local interactions are the ,B-sheets consisting of two or more ,B-strands interconnected by H-bonds, see fig. 1. Often the ,B-strands in a sheet are widely separated in the sequence, implying that only part of the available sequence information about a ,B-sheet can be contained in a window of, say, 13 amino acids. This is one of the reasons why the accuracy of ,B-strand predictions are generally lower than the accuracy of a-helix predictions. The aim of this work is to improve prediction of secondary structures by combining local predictions of a-helix, ,B-strand and coil with a non-local method predicting ,B-sheets. Other work along the same directions include [6] in which ,B-sheet predictions are done by linear methods and [7] where a so-called density network is applied to the problem. 2 A NEURAL NETWORK WITH TWO WINDOWS We aim at capturing correlations in the ,B-sheets by using a neural network with two windows, see fig. 2. While window 1 is centered around amino acid number i (ai), window 2 slides along the rest of the chain. When the amino acids centered in each of the two windows sit opposite each other in a ,B-sheet the target output is 1, and otherwise O. After the whole protein has been traversed by window 2, window 1 is moved to the next position (i + 1) and the procedure is repeated. If the protein is L amino acids long this procedure yields an output value for each of the L(L -1)/2 Prediction of Beta Sheets in Proteins Figure 2: Neural network for predicting ,B-sheets. The network employs weight sharing to improve the encoding of the amino acids and to reduce the number of adjustable parameters. 919 pairs of amino acids. We display the output in a L x L gray-scale image as shown in fig. 3. We assume symmetry of sheets, i.e., if the two windows are interchanged, the output does not change. This symmetry is ensured (approximately) during training by presenting all inputs in both directions. Each window of the network sees K amino acids. An amino acid is represented by a vector of20 binary numbers all being zero, except one, which is 1. That is, the amino acid A is represented by the vector 1,0,0, ... ,0 and so on. This coding ensures that the input representations are un correlated , but it is a very inefficient coding, since 20 amino acids could in principle be represented by only 5 bit. Therefore, we use weight sharing [8] to learn a better encoding [4]. The 20 input units corresponding to one window position are fully connected to three hidden units. The 3 x (20 + 1) weights to these units are shared by all window positions, i.e., the activation of the 3 hidden units is a new learned encoding of the amino acids, so instead of being represented by 20 binary values they are represented by 3 real values. Of course the number of units for this encoding can be varied, but initial experiments showed that 3 was optimal [4]. The two windows of the network are made the same way with the same number of inputs etc .. The first layer of hidden units in the two windows are fully connected to a hidden layer which is fully connected to the output unit, see fig. 2. Furthermore, two structurally identical networks are used: one for parallel and one for anti-parallel ,B-sheets. The basis for the training set in this study is the set of 126 non-homologous protein chains used in [9], but chains forming ,B-sheets with other chains are excluded. This leaves us with 85 proteins in our data set. For a protein of length L only a very small fraction of the L(L - 1)/2 pairs are positive examples of ,B-sheet pairs. Therefore it is very important to balance the positive and negative examples to avoid the situation where the network always predicts no ,B-sheet. Furthermore, there are several types of negative examples with quite different occurrences: 1) two amino acids of which none belong to a ,B-sheet; 2) one in a ,B-sheet and one which is not in a ,B-sheet; 3) two sitting in ,B-sheets, but not opposite to each other. The balancing was done in the following way. For each positive example selected at random a negative example from each of the three categories were selected at random. If the network does not have a second layer of hidden units, it turns out that the result is no better than a network with only one input window, i.e., the network cannot capture correlations between the two windows. Initial experiments indicated that about 10 units in the second hidden layer and two identical input windows of size K = 9 gave the best results. In fig. 3(left) the prediction of anti-parallel sheets is shown for the protein identified as 1acx in the Brookhaven Protein Data Bank 920 120 100 .. 80 :g '" o .!: ~ 60 / 40 ". 20 A. KROGH, S. K. RIIS Figure 3: Left: The prediction of anti-parallel ,8-sheets in the protein laex. In the upper triangle the correct structure is shown by a black square for each ,8-sheet pair. The lower triangle shows the prediction by the two-window network. For any pair of amino acids the network output is a number between zero (white) and one (black), and it is displayed by a linear gray-scale. The diagonal shows the prediction of a-helices. Right: The same display for parallel ,8-sheets in the protein 4fxn. Notice that the correct structure are lines parallel to the diagonal, whereas they are perpendicular for anti-parallel sheets. For both cases the network was trained on a training set that did not contain the protein for which the result is shown. [10]. First of all, one notices the checker board structure of the prediction of ,8sheets. This is related to the structure of ,8-sheets. Many sheets are hydrophobic on one side and hydrophilic on the other. The side chains of the amino acids in a strand alternates between the two sides of the sheet, and this gives rise to the periodicity responsible for the pattern. Another network was trained on parallel ,8-sheets. These are rare compared to the anti-parallel ones, so the amount of training data is limited. In fig. 3(right) the result is shown for protein 4fxn. This prediction seems better than the one obtained for anti-parallel sheets, although false positive predictions still occurs at some positions with strands that do not pair. Strands that bind in parallel ,8-sheets are generally more widely separated in the sequence than strands in anti-parallel sheets. Therefore, one can imagine that the strands in parallel sheets have to be more correlated to find each other in the folding process, which would explain the better prediction accuracy. The results shown in fig. 3 are fairly representative. The network misses some of the sheets, but false positives present a more severe problem. By calculating correlation coefficients we can show that the network doe!> capture some correlations, but they seem to be weak. Based on these results, we hypothesize that the formation of ,8sheets is only weakly dependent on correlations between corresponding ,8-strands. This is quite surprising. However weak these correlations are, we believe they can still improve the accuracy of the three state secondary structure prediction. In order to combine local methods with the non-local ,8-sheet prediction, we introduce a global energy function as described below. Prediction of Beta Sheets in Proteins 921 3 A GLOBAL ENERGY FUNCTION We use a newly developed local neural network method based on one input window [4] to give an initial prediction of the three possible structures. The output from this network is constrained by soft max [11], and can thus be interpreted as the probabilities for each of the three structures. That is, for amino acid ai, it yields three numbers Pi,n, n = 1,2 or 3 indicating the probability of a-helix (Pi,l) , (3sheet (pi,2), or coil (pi,3). Define Si,n = 1 if amino acid i is assigned structure n and Si,n = 0 otherwise. Also define hi,n = 10gPi,n. We now construct the 'energy function' (1) i n where weights Un are introduced for later usage. Assuming the probabilities Pi,n are independent for any two amino acids in a sequence, this is the negative log likelihood of the assigned secondary structure represented by s, provided that Un = 1. As it stands, alone, it is a fairly trivial energy function, because the minimum is the assignment which corresponds to the prediction with the maximum Pi,n at each position i-the assignment of secondary structure that one would probably use anyway. For amino acids ai and aj the logarithm of the output of the (3-sheet network described previously is called qfj for parallel (3-sheets and qfj for anti-parallel sheets. We interpret these numbers as the gain in energy if a (3-sheet pair is formed. (As more terms are added to the energy, the interpretation as a log-likelihood function is gradually fading.) If the two amino acids form a pair in a parallel (3-sheet, we set the variable T~ equal to 1, and otherwise to 0, and similarly with Tii for antiparallel sheets. Thus the Tii and T~ are sparse binary matrices. Now the total energy of the (3-sheets can be expressed as Hf3(s, Ta, TP) = ~[CaqfjTij + CpqfjT~], (2) 'J where Ca and Cp determine the weights of the two terms in the function. Since an amino acid can only be in one structure, the dynamic T and S variables are constrained: Only Tii or T~ can be 1 for the same (i, j), and if any of them is 1 the amino acids involved must be in a (3-sheet, so Si,2 = Sj,2 = 1. Also, Si ,2 can only be 1 if there exists a j with either Iii or T~ equal to 1. Because of these constraints we have indicated an S dependence of H f3. The last term in our energy function introduces correlations between neighboring amino acids. The above assumption that the secondary structure of the amino acids are independent is of course a bad assumption, and we try to repair it with a term Hn(s) = L: L: Jnm Si,n Si+l,m, i nm (3) that introduces nearest neighbor interactions in the chain. A negative J11, for instance, means that a following a is favored, and e.g., a positive h2 discourages a (3 following an a. Now the total energy is (4) Since (3-sheets are introduced in two ways, through hi ,2 and qij, we need the weights Un in (1) to be different from 1. The total energy function (4) has some resemblance with a so-called Potts glass in an external field [12]. The crucial difference is that the couplings between the 922 A. KROGH, S. K. RIIS 'spins' Si are dependent on the dynamic variables T. Another analogy of the energy function is to image analysis, where couplings like the T's are sometimes used as edge elements. 3.1 PARAMETER ESTIMATION The energy function contains a number of parameters, Un, Ca , Cp and Jnm . These parameters were estimated by a method inspired by Boltzmann learning [13]. In the Boltzmann machine the estimation of the weights can be formulated as a minimization of the difference between the free energy of the 'clamped' system and that of the 'free-running' system [14]. If we think of our energy function as a free energy (at zero temperature), it corresponds to minimizing the difference between the energy of the correct protein structure and the minimum energy, where p is the total number of proteins in the training set. Here the correct structure of protein J-l is called S(J-l) , Ta(J-l), TP(p), whereas s(J-l), Ta(J-l) , TP(J-l) represents the structure that minimizes the energy Htotal. By definition the second term of C is less than the first, so C is bounded from below by zero. The cost function C is minimized by gradient descent in the parameters. This is in principle straightforward, because all the parameters appear linearly in Htotal. However, a problem with this approach is that C is minimal when all the parameters are set to zero, because then the energy is zero. It is cured by constraining some of the parameters in Htotal. We chose the constraint l:n Un = 1. This may not be the perfect solution from a theoretical point of view, but it works well. Another problem with this approach is that one has to find the minimum of the energy Htotal in the dynamic variables in each iteration of the gradient descent procedure. To globally minimize the function by simulated annealing each time would be very costly in terms of computer time. Instead of using the (global) minimum of the energy for each protein, we use the energy obtained by minimizing the energy from the correct structure. This minimization is done by a greedy algorithm in the following way. In each iteration the change in s, Ta, TP which results in the largest decrease in Htotal is carried out. This is repeated until any change will increase Htotal. This algorithm works towards a local stability of the protein structures in the training set. We believe it is not only an efficient way of doing it, but also a very sensible way. In fact, the method may well be applicable in other models, such as Boltzmann machines. 3.2 STRUCTURE PREDICTION BY SIMULATED ANNEALING After estimation of the parameters on which the energy function Htotal depends, we can proceed to predict the structure of new proteins. This was done using simulated annealing and the EBSA package [15]. The total procedure for prediction is, 1. A neural net predicts a-helix, ,8-strand or coil. The logarithm of these predictions give all the hi,n for that protein. 2. The two-window neural networks predict the ,8-sheets. The result is the qfj from one network and the qfj from the other. 3. A random configuration of S, Ta, TP variables is generated from which the simulated annealing minimization of Htotal was started. During annealing, all constraints on s, Ta, TP variables are strictly enforced. Prediction of Beta Sheets in Proteins 923 4. The final minimum configuration s is the prediction of the secondary structure. The ,B-sheets are predicted by t a and tv. Using the above scheme, an average secondary structure accuracy of 66.5% is obtained by seven-fold cross validation. This should be compared to 66.3% obtained by the local neural network based method [4] on the same data set. Although these preliminary results do not represent a significant improvement, we consider them very encouraging for future work. Because the method not only predicts the secondary structure, but also which strands actually binds to form ,B-sheets, even a modest result may be an important step on the way to full 3D predictions. 4 CONCLUSION In this paper we introduced several novel ideas which may be applicable in other contexts than prediction of protein structure. Firstly, we described a neural network with two input windows that was used for predicting the non-local structure called ,B-sheets. Secondly, we combined local predictions of a-helix, ,B-strand and coil with the ,B-sheet prediction by minimization of a global energy function. Thirdly, we showed how the adjustable parameters in the energy function could be estimated by a method similar to Boltzmann learning. We found that correlations between ,B-strands in ,B-sheets are surprisingly weak. Using the energy function to combine predictions improves performance a little. Although we have not solved the protein folding problem, we consider the results very encouraging for future work. This will include attempts to improve the performance of the two-window network as well as experimenting with the energy function, and maybe add more terms to incorporate new constraints. Acknowledgments: We would like to thank Tim Hubbard, Richard Durbin and Benny Lautrup for interesting comments on this work and Peter Salamon and Richard Frost for assisting with simulated annealing. This work was supported by a grant from the Novo Nordisk Foundation. References [1] C. Branden and J. Tooze, Introduction to Protein Structure (Garland Publishing, Inc., New York, 1991). [2] N. Qian and T. Sejnowski, Journal of Molecular Biology 202, 865 (1988). [3] H. Bohr et al., FEBS Letters 241, 223 (1988). [4] S. Riis and A. Krogh, Nordita Preprint 95/34 S, submitted to J. Compo BioI. [5] B. Rost, C. Sander, and R. Schneider, J Mol. BioI. 235, 13 (1994). [6] T. Hubbard, in Proc. of the 27th HICSS, edited by R. Lathrop (IEEE Computer Soc. Press, 1994), pp. 336-354. [7] D. J. C. MacKay, in Maximum Entropy and Bayesian Methods, Cambridge 1994, edited by J. Skilling and S. Sibisi (Kluwer, Dordrecht, 1995). [8] Y. Le Cun et al., Neural Computation 1, 541 (1989). [9] B. Rost and C. Sander, Proteins 19, 55 (1994). [10] F. Bernstein et al., J Mol. BioI. 112,535 (1977). [11] J. Bridle, in Neural Information Processing Systems 2, edited by D. Touretzky (Morgan Kaufmann, San Mateo, CA, 1990), pp. 211-217. [12] K. Fisher and J. Hertz, Spin glasses (Cambridge University Press, 1991). [13] D. Ackley, G. Hinton, and T. Sejnowski, Cognitive Science 9, 147 (1985). [14] J. Hertz, A. Krogh, and R. Palmer, Introduction to the Theory of Neural Computation (Addison-Wesley, Redwood City, 1991). [15] R. Frost, SDSC EBSA, C Library Documentation, version 2.1. SDSC Techreport.
|
1995
|
75
|
1,123
|
Active Learning in Multilayer Perceptrons Kenji Fukumizu Information and Communication R&D Center, Ricoh Co., Ltd. 3-2-3, Shin-yokohama, Yokohama, 222 Japan E-mail: fuku@ic.rdc.ricoh.co.jp Abstract We propose an active learning method with hidden-unit reduction. which is devised specially for multilayer perceptrons (MLP). First, we review our active learning method, and point out that many Fisher-information-based methods applied to MLP have a critical problem: the information matrix may be singular. To solve this problem, we derive the singularity condition of an information matrix, and propose an active learning technique that is applicable to MLP. Its effectiveness is verified through experiments. 1 INTRODUCTION When one trains a learning machine using a set of data given by the true system, its ability can be improved if one selects the training data actively. In this paper, we consider the problem of active learning in multilayer perceptrons (MLP). First, we review our method of active learning (Fukumizu el al., 1994), in which we prepare a probability distribution and obtain training data as samples from the distribution. This methodology leads us to an information-matrix-based criterion similar to other existing ones (Fedorov, 1972; Pukelsheim, 1993). Active learning techniques have been recently used with neural networks (MacKay, 1992; Cohn, 1994). Our method, however, as well as many other ones has a crucial problem: the required inverse of an information matrix may not exist (White, 1989). We propose an active learning technique which is applicable to three-layer perceptrons. Developing a theory on the singularity of a Fisher information matrix, we present an active learning algorithm which keeps the information matrix nonsingular. We demonstrate the effectiveness of the algorithm through experiments. 296 K. FUKUMIZU 2 STATISTICALLY OPTIMAL TRAINING DATA 2.1 A CRITERION OF OPTIMALITY We review the criterion of statistically optimal training data (Fukumizu et al., 1994). We consider the regression problem in which the target system maps a given input z to y according to y = I(z) + Z, where I( z) is a deterministic function from R L to R M , and Z is a random variable whose law is a normal distribution N(O,(12IM ), (IM is the unit M x M matrix). Our objective is to estimate the true function 1 as accurately as possible. Let {/( z; O)} be a parametric model for estimation. We use the maximum likelihood estimator (MLE) 0 for training data ((z(v), y(v»)}~=l' which minimizes the sum of squared errors in this case. In theoretical derivations, we assume that the target function 1 is included in the model and equal to 1(,; (0 ). We make a training example by choosing z(v) to try, observing the resulting output y(v), and pairing them. The problem of active learning is how to determine input data {z(v)} ~=l to minimize the estimation error after training. Our approach is a statistical one using a probability for training, r( z), and choosing {z(v) }:Y"=l as independent samples from r(z) to minimize the expectation of the MSE in the actual environment: In the above equation, Q is the environmental probability which gives input vectors to the true system in the actual environment, and E{(zlv"yIV')} means the expectation on training data. Eq.(I), therefore, shows the average error of the trained machine that is used as a substitute of the true function in the actual environment. 2.2 REVIEW OF AN ACTIVE LEARNING METHOD Using statistical a.~ymptotic theory, Eq. (1) is approximated a.~ follows: 2 EMSE = (12 + ~ Tr [I(Oo)J-1(Oo)] + O(N-3j2), (2) where the matrixes I and J are (Fisher) illformation matrixes defined by 1(0) = J I(z;O)dQ(z). J(O) = J I(z;O)r(z)dz. The essential part of Eq.(2) is Tr[I(Oo)J-1(Oo»), computed by the unavailable parameter 00 • We have proposed a practical algorithm in which we replace 00 with O. prepare a family of probability {r( z; 'lI) I 'U : paramater} to choose training samples, and optimize 'U and {) iteratively (Fllkumizll et al., 1994). Active Learning Algorithm 1. Select an initial training data set D[o] from r( z; 'lI[O])' and compute 0[0]' 2. k:= 1. 3. Compute the optimal v = V[k] to minimize Tr[I(O[k_l])J-1(O[k_l]»)' Active Learning in Multilayer Perceptrons 297 4. Choose ~ new training data from r(z;V[k]) and let D[k] be a union of D[k-l] and the new data. 5. Compute the MLE 9[k] based on the training data set D[k]. 6. k := k + 1 and go to 3. The above method utilizes a probability to generate training data. It has the advantage of making many data in one step compared to existing ones in which only one data is chosen in each step, though their criterions are similar to each other. 3 SINGULARITY OF AN INFORMATION MATRIX 3.1 A PROBLEM ON ACTIVE LEARNING IN MLP Hereafter, we focus on active learning in three-layer perceptrons with H hidden units, NH = {!(z, O)}. The map !(z; 0) is defined by H L h(z; 0) = L Wij s(L UjkXk + (j) + 7]i, (1~i~M), (3) j=1 k=1 where s(t) is the sigmoidal function: s(t) = 1/(1 + e-t ). Our active learning method as well as many other ones requires the inverse of an information matrix J. The information matrix of MLP, however, is not always invertible (White, 1989). Any statistical algorithms utilizing the inverse, then, cannot be applied directly to MLP (Hagiwara et al., 1993). Such problems do not arise in linear models, which almost always have a nonsingular information matrix. 3.2 SINGULARITY OF AN INFORMATION MATRIX OF MLP The following theorem shows that the information matrix of a three-layer perceptron is singular if and only if the network has redundant hidden units. We can deduce tha.t if the information matrix is singular, we can make it nonsingular by eliminating redundant hidden units without changing the input-output map. Theorem 1 Assume r(z) is continuous and positive at any z. Then. the Fisher information matrix J is singular if and only if at least one of the follo'wing three con(litions is satisfied: (1) u,j := (Ujl, ... , UjL)T = 0, for some j. (2) Wj:= (Wlj, ... ,WMj) = OT , for some j. (3) For difJerenth andh, (U,h,(jt) = (U,1,(h) or (U,h,(it) = -(U,h,(h)· The rough sketch of the proof is shown below. The complete proof will appear in a forthcoming pa.per ,(Fukumizu, 1996). Rough sketch of the proof. We know easily that an information matrix is singular if and ouly if {()fJ:~(J)}a are linearly dependent. The sufficiency can be proved easily. To show the necessity, we show that the derivatives are linearly independent if none of the three conditions is satisfied. Assume a linear relation: 298 K. FUKUMIZU We can show there exists a basis of R L , (Z(l), ... , Z(L», such that Uj . z(l) i- 0 for 'Vj, 'VI, and Uj! . z(l) + (h i- ±(u12 . z(l) + (h) for jl i- h,'VI. We replace z in eq.(4) by z(l)t (t E R). Let my) := Uj· z(l), Sjl) := {z E C I z = ((2n+ 1)1T/=1(j)/m~l), n E Z}, and D(l) := C - UjSY). The points in S~l) are the singularities of s(m~l) z + (j). We define holomorphic functions on D(l) as q,~l)(z) ._ 'Ef=l aijs(my> z + (j) + aiO + 'E~l 'E~=l,BjkWijS'(my) z + (j)x~l> z +'E~l,BjOWijS'(my)z+(j), (1 ~ i ~ M). From eq.( 4), we have q,~l) (t) = 0 for all t E R. Using standard arguments on isolated singularities of holomorphic functions, we know SY) are removable singularities of q,~l)(z), and finally obtain Wij 'E~=l,BjkX~I) = 0, Wij,BjO = 0, aij = 0, aiO = o. It is easy to see ,Bjk = O. This completes the proof. 3.3 REDUCTION PROCEDURE We introduce the following reduction procedure based on Theorem 1. Used during BP training, it eliminates redundant hidden units and keeps the information matrix nonsingular. The criterion of elimination is very important, because excessive elimination of hidden units degrades the approximation capacity. We propose an algorithm which does not increase the mean squared error on average. In the following, let Sj := s( itj . z + llj) and £( N) == A/ N for a positive number A. Reduction Procedure 1. If IIWjll2 J(Sj - s((j))2dQ < £(N), then eliminate the jth hidden unit, and lli -. lli + WijS((j) for all i. 2. If IIwjll2 J(sj)2dQ < €(N), then eliminate the jth hidden unit. 3. If IIwhll2 J(sh - sjJ2dQ < €(N) for different it and h, then eliminate the hth hidden unit and Wij! -. wih + Wijz for all i. 4. If IIwhll2 J(1 - sh - sjJ 2dQ < €(N) for different jl and h, ~hen eliminate the j2th hidden unit and wih -. Wij! - wih, ili -. ili + wih for all 'i. From Theorem 1, we know that Wj, itj, (ith' (h) - (it'};, (j!), or (ith, (h )+( it]:, (h) can be reduced to 0 if the information matrix is singular. Let 0 E NK denote the reduced parameter from iJ according to the above procedure. The above four conditions are, then, given by calculating J II/(x; 0) -/(x; iJ)WdQ. We briefly explain how the procedure keeps the information matrix nonsingular and does not increase EMSE in high probability. First, suppose detJ(Oo) = 0, then there exists Off E NK (K < H) such that f(x;Oo) = f(x;Off) and detJ(Of) i- 0 in N K. The elimination of hidden units up to K, of course, does not increase the EMSE. Therefore, we have only to consider the case in which detJ(Oo) i- 0 and hidden units are eliminated. Suppose J II/(z; Off) -/(z; Oo)1I2dQ > O(N- 1 ) for any reduced parameter Off from 00 • The probability of satisfying J II/(z;iJ) -/(z;O)WdQ < A/N is very small for Active Learning in Multilayer Perceptrons 299 a sufficiently small A. Thus, the elimination of hidden units occurs in very tiny probability. Next, suppose J 1I!(x; (Jff) - !(x; (Jo)1I 2dQ = O(N-l). Let 0 E NK be a reduced parameter made from 9 with the same procedure as we obtain (Jff from (Jo. We will show for a sufficiently small A, where OK is MLE computed in NK. We write (J = ((J(l),(J(2») in which (J(2) is changed to 0 in reduction, changing the coordinate system if necessary. The Taylor expansion and asymptotic theory give E [JII!(x; OK) - !(x; (Jo)1I2dQ] ~ JII!(x; (Jf)-!(x; (Jo)11 2dQ+ ~ Tr[In((Jf)Jil1((Jf)), 2 E [JII!(x; 9) - !(x; O)WdQ] ~ JII!(x; (Jf)-!(x; (Jo)1I 2dQ+ ;, Tr[h2 ((Jf)J2;l ((Jo)], where Iii and Jii denote the local information matrixes w.r.t. (J(i) ('i = 1,2). Thus, E [JII!(x; 0) - !(x; (Jo)1I 2dQ] - E [JII!(x; OK) - !(x; (Jo)1I 2dQ] 2 ~ -E [JII!(x;o) - !(X;O)1I 2dQ] +;' Tr[h2((Jf)J;1((Jo)) 2 - ;, Tr[Ill((Jf)Jil1((Jf)] + E [JII!(x; 0) - !(x; (Jo)1I 2dQ] . Since the sum of the last two terms is positive, the 1.h.s is positive if E[f II!( x; OK)_ !(x; 0)1I 2dQ) < BIN for a sufficiently small B. Although we cannot know the value of this expectation, we can make the probability of holding this enequality very high by taking a small A. 4 ACTIVE LEARNING WITH REDUCTION PROCEDURE The reduction procedure keeps the information matrix nonsingular and makes the active learning algorithm applicable to MLP even with surplus hidden units. Active Learning with Hidden Unit Reduction 1. Select initial training data set Do from r( x; V[O]). and compute 0[0]' 2. k:= 1, and do REDUCTION PROCEDURE. 3. Compute the optimal v = 1'[k] to minimize Tr[I(9[k_l])J-l (9[k-l] )). using the steepest descent method. 4. Choose Nk new training data from r( x; V[k]) and let D[k] be a union of D[k-l] and the new data. 5. Compute the MLE 9[kbbased on the training data D[k] using BP with REDUCTION PROCE URE. 6. k:= k + 1 and go to 3. The BP with reduction procedure is applicable not only to active learning, but to a variety of statistical techniques that require the inverse of an information matrix. We do not discuss it in this paper. however. 300 • • -- Active Learning • Active Learning [Av·Sd,Av+Sd] - .. - . Passive Learning + Passive Learning [Av·Sd,Av+Sd] ~ + + + + • • + • • • 200 400 600 800 100> The Number of Training nata K. FUKUMIZU 0.00001 4 -- Learning Curve ..•.. It of hidden units O.IXXXlOOI 0 100 200 300 400 soo 600 700 800 900 100> The Number of Training nata Figure 1: Active/Passive Learning: f(x) = s(x) 5 EXPERIMENTS We demonstrate the effect of the proposed active learning algorithm through experiments. First we use a three-layer model with 1 input unit, 3 hidden units, and 1 output unit. The true function f is a MLP network with 1 hidden unit. The information matrix is singular at 0o, then. The environmental probability, Q, is a normal distribution N(O,4). We evaluate the generalization error in the actual environment using the following mean squared error of the function values: ! 1If(:l!; 0) - f(:l!)11 2dQ. We set the deviation in the true system II = 0.01. As a family of distributions for training {r(:l!;v)}, a mixture model of 4 normal distributions is used. In each step of active learning, 100 new samples are added. A network is trained using online BP, presented with all training data 10000 times in each step, and operated the reduction procedure once a 100 cycles between 5000th and 10000th cycle. We try 30 trainings changing the seed of random numbers. In comparison, we train a network passively based on training samples given by the probability Q. Fig.1 shows the averaged learning curves of active/passive learning and the number of hidden units in a typical learning curve. The advantage of the proposed active learning algorithm is clear. We can find that the algorithm has expected effects on a simple, ideal approximation problem. Second, we apply the algorithm to a problem in which the true function is not included in the MLP model. We use MLP with 4 input units, 7 hidden units, and 1 output unit. The true function is given by f(:l!) = erf(xt), where erf(t) is the error function. The graph of the error function resembles that of the sigmoidal function, while they never coincide by any affine transforms. We set Q = N(0,25 X 14). We train a network actively/passively based on 10 data sets, and evaluate MSE's of function values. Other conditions are the same as those of the first experiment. Fig.2 shows the averaged learning curves and the number of hidden units in a typical learning curve. We find tha.t the active learning algorithm reduces the errors though the theoretical condition is not perfectly satisfied in this case. It suggests the robustness of our active learning algorithm. Active Learning in Multilayer Perceptrons -- Active Learning O.IXXlI - .. - . Passive Learning 200 400 600 800 IIXXl The Number ofTraining nata -- Learning Curve ..•.. # of hidden units L .... .... . ... ......... : 301 8 ;; 7 It :s c :; ~ 6 e. :r is: ~ :s .. 5; r-~~~r-~-r~--r-~-+4 100 200 300 400 500 600 700 800 900 IIXXl The Number of Training nata ~ Figure 2: Active/Passive Learning: f(z) = erf(xI) 6 CONCLUSION We review statistical active learning methods and point out a problem in their application to MLP: the required inverse of an information matrix does not exist if the network has redundant hidden units. We characterize the singularity condition of an information matrix and propose an active learning algorithm which is applicable to MLP with any number of hidden units. The effectiveness of the algorithm is verified through computer simulations, even when the theoretical assumptions are not perfectly satisfied. References D. A. Cohn. (1994) Neural network exploration using optimal experiment design. In J. Cowan et al. (ed.), A d'vances in Neural Information Processing SYHtems 6, 679-686. San Mateo, CA: Morgan Kaufmann. V. V. Fedorov. (1972) Theory of Optimal Experiments. NY: Academic Press. K. Fukumizu. (1996) A Regularity Condition of the Information Matrix of a Multilayer Percept ron Network. Neural Networks, to appear. K. Fukumizu, & S. Watanabe. (1994) Error Estimation and Learning Data Arrangement for Neural Networks. Proc. IEEE Int. Conf. Neural Networks :777-780. K. Hagiwara, N. Toda, & S. Usui. (1993) On the problem of applying AIC to determine the structure of a layered feed-forward neural network. Proc. 1993 Int. Joint ConI. Neural Networks :2263-2266. D. MacKay. (1992) Information-based objective functions for active data selection, Ne'ural Computation 4(4):305-318. F. Pukelsheim. (1993) Optimal Design of Experiments. NY: John Wiley & Sons. H. White. (1989) Learning in artificial neural networks: A statistical perspective Neural Computation 1 ( 4 ):425-464.
|
1995
|
76
|
1,124
|
A model of transparent motion and non-transparent motion aftereffects Alexander Grunewald* Max-Planck Institut fur biologische Kybernetik Spemannstrafie 38 D-72076 Tubingen, Germany Abstract A model of human motion perception is presented. The model contains two stages of direction selective units. The first stage contains broadly tuned units, while the second stage contains units that are narrowly tuned. The model accounts for the motion aftereffect through adapting units at the first stage and inhibitory interactions at the second stage. The model explains how two populations of dots moving in slightly different directions are perceived as a single population moving in the direction of the vector sum, and how two populations moving in strongly different directions are perceived as transparent motion. The model also explains why the motion aftereffect in both cases appears as non-transparent motion. 1 INTRODUCTION Transparent motion can be studied using displays which contain two populations of moving dots. The dots within each population have the same direction of motion, but directions can differ between the two populations. When the two directions are very similar, subjects report seeing dots moving in the average direction (Williams & Sekuler, 1984). However, when the difference between the two directions gets large, subjects perceive two overlapping sheets of moving dots. This percept is called transparent motion. The occurrence of transparent motion cannot be explained by direction averaging, since that would result in a single direction of perceived motion. Rather than just being a quirk of the human visual system, transparent motion is an important issue in motion processing. For example, when a robot is moving its • Present address: Caltech, Mail Code 216-76, Pasadena, CA 91125. 838 A. GRUNEWALD motion leads to a velocity field. The ability to detect transparent motion within that velocity field enables the robot to detect other moving objects at the same time that the velocity field can be used to estimate the heading direction of the robot. Without the ability to code mUltiple directions of motion at the same location, i.e. without the provision for transparent motion, this capacity is not available. Traditional algorithms have failed to properly process transparent motion, mainly because they assigned a unique velocity signal to each location, instead of allowing the possibility for multiple motion signals at a single location. Consequently, the study of transparent motion has recently enjoyed widespread interest. STIMULUS PERCEPT Test Figure 1: Two populations of dots moving in different directions during an adaptation phase are perceived as transparent motion. Subsequent viewing of randomly moving dots during a test phase leads to an illusory percept of unidirectional motion, the motion aftereffect (MAE). Stimulus and percept in both phases are shown. After prolonged exposure to an adaptation display containing dots moving in one direction, randomly moving dots in a test display appear to be moving in the opposite direction (Hiris & Blake, 1992; Wohlgemuth, 1911). This illusory percept of motion is called the motion aftereffect (MAE). Traditionally this is explained by assuming that pairs of oppositely tuned direction selective units together code the presence of motion. When both are equally active, no motion is seen. Visual motion leads to stronger activation of one unit, and thus an imbalance in the activity of the two units. Consequently, motion is perceived. Activation of that unit causes it to fatigue, which means its response weakens. After motion offset, the previously active unit sends out a reduced signal compared to its partner due to adaptation. Thus adaptation generates an imbalance between the two units, and therefore illusory motion, the MAE, is perceived. This is the ratio model (Sutherland, 1961). Recent psychophysical results show that after prolonged exposure to transparent motion, observers perceive a MAE of a single direction of motion, pointing in the vector average of the adaptation directions (Mather, 1980; Verstraten, Fredericksen, & van de Grind, 1994). Thus adaptation to transparent motion leads to a non-transparent MAE. This is illustrated in Figure 1. This result cannot be accounted for by the ratio model, since the non-transparent MAE does not point in the direction opposite to either of the adaptation directions. Instead, this result suggests that direction selective units of all directions interact and thus contribute to the MAE. This explanation is called the distribution-shift model (Mather, 1980). However, thus far it has only been vaguely defined, and no demonstration has been given that shows how this mechanism might work. A Model of Transparent Motion and Non-transparent Motion Aftereffects 839 This study develops a model of human motion perception based on elements from both the ratio and the distribution-shift models for the MAE. The model is also applicable to the situation where two directions of motion are present. When the directions differ slightly, only a single direction is perceived. When the directions differ a lot, transparent motion is perceived. Both cases lead to a unitary MAE. 2 OUTLINE OF THE MODEL The model consists of two stages. Both stages contain units that are direction selective. The architecture of the model is shown in Figure 2. ~----~--~,~r---~---'---\ -, Stage 2 CD080CD 86) +----+~--+~--~----~--~--~--~~ Figure 2: The model contains two stages of direction selective units. Units at stage 1 excite units of like direction selectivity at stage 2, and inhibit units of opposite directions. At stage 2 recurrent inhibition sharpens directional motion responses. The grey level indicates the strength of interaction between units. Strong influence is indicated by black arrows, weak influence is indicated by light grey arrows. Units in stage 1 are broadly tuned motion detectors. In the present study the precise mechanism of motion detection is not central, and hence it has not been modeled. It is assumed that the bandwidth of motion detectors at this stage is about 30 degrees (Raymond, 1993; Williams, Tweten, & Sekuler, 1991). In the absence of any visual motion, all units are active at a baseline level; this is equivalent to neuronal noise. Whenever motion of a particular direction is present in the input, the activity of the corresponding unit (Vi) is activated maximally (Vi = 9), and units of similar direction selectivity are weakly activated (Vi = 3). The activities of all other units decrease to zero. Associated with each unit i at stage 1 is a weight Wi that denotes the adaptational state of unit i to fire a unit at stage 2. During prolonged exposure to motion these weights adapt, and their strength decreases. The equation governing the strength of the weights is given below: dWi = R(1- w·) - V·W · dt ~ ~~, where R = 0.5 denotes the rate of recovery to the baseline weight. When Wi = 1 the corresponding unit is not adapted. The further Wi is reduced from 1, the more 840 A. GRUNEWALD the corresponding unit is adapted. The products ViWi are transmitted to stage 2. Each unit of stage 1 excites units coding similar directions at stage 2, and inhibits units coding opposite directions of motion. The excitatory and inhibitory effects between units at stages 1 and 2 are caused by kernels, shown in Figure 3. Feedforward kernels Feedback kernels 1 1 I 1excitatory excitatory 0.8 ------0.8 ---~--- inhibitory inhibitory 0.6 0 . 6 0.4 0.4 r0.2 0.2 f---------0 -0 ------+ --------180 0 180 -180 o 180 Figure 3: Kernels used in the model. Left: excitatory and inhibitory kernels between stages 1 and 2; right: excitatory and inhibitory feedback kernels within stage 2. Activities at stage 2 are highly tuned for the direction of motion. The broad activation of motion signals at stage 1 is directionally sharpened at stage 2 through the interactions between recurrent excitation and inhibition. Each unit in stage 2 excites itself, and interacts with other units at stage 2 through recurrent inhibition. This inhibition is maximal for close directions, and falls off as the directions become more dissimilar. The kernels mediating excitatory and inhibitory interactions within stage 2 are shown in Figure 3. Through these inhibitory interactions the directional tuning of units at stage 2 is sharpened; through the excitatory feedback it is ensured that one unit will be maximally active. Activities of units at stage 2 are given by Mi = max4(mi' 0), where the behavior of mi is governed by: F/ and Fi- denote the result of convolving the products of the activities at stage 1 and the corresponding adaptation level, VjWj , with excitatory and inhibitory feedforward kernels respectively. Similarly, Bt and Bj denote the convolution of the activities M j at stage 2 with the feedback kernels. 3 SIMULATIONS OF PSYCHOPHYSICAL RESULTS In the simulations there were 24 units at each stage. The model was simulated dynamically by integrating the differential equations using a fourth order RungeKutta method with stepsize H = 0.01 time units. The spacing of units in direction space was 15 degrees at both stages. Spatial interactions were not modeled. In the simulations shown, a motion stimulus is present until t = 3. Then the motion stimulus ceases. Activity at stage 2 after t = 3 corresponds to a MAE. A Model of Transparent Motion and Non-transparent Motion Aftereffects 841 3.1 UNIDIRECTIONAL MOTION When adapting to a single direction of motion, the model correctly generates a motion signal for that particular direction of motion. After offset of the motion input, the unit coding the opposite direction of motion is activated, as in the MAE. A simulation of this is shown in Figure 4. Stage 1 Stage 2 act act 360 360 Figure 4: Simulation of single motion input and resulting MAE. Motion input is presented until t = 3. During adaptation the motion stimulus excites the corresponding units at stage 1, which in turn activate units at stage 2. Due to recurrent inhibition only one unit at stage 2 remains active (Grossberg, 1973), and thus a very sharp motion signal is registered at stage 2. During adaptation the weights associated with the units that receive a motion input decrease. After motion offset, all units receive the same baseline input. Since the weights of the previously active units are decreased, the corresponding cells at stage 2 receive less feedforward excitation. At the same time, the previously active units receive strong feedforward inhibition, since they receive inhibition from units tuned to very different directions of motion and whose weights did not decay during adaptation. Similarly, the units coding the opposite direction of motion as those previously active receive more excitation and less inhibition. Through recurrent inhibition the unit at stage 2 coding the opposite direction to that which was active during adaptation is activated after motion offset: this activity corresponds to the MAE. Thus the MAE is primarily an effect of disinhibition. 3.2 TRANSPARENT MOTION: SIMILAR DIRECTIONS Two populations of dots moving in different, but very similar, directions lead to bimodal activation at stage 1. Since the feedforward excitatory kernel is broadly tuned, and since the directions of motion are similar, the ensuing distribution of activities at stage 2 is unimodal, peaking halfway between the two directions of motion. This corresponds to the vector average of the directions of motion of the two populations of dots. A simulation of this is shown in Figure 5. During adaptation the units at stage 1 corresponding to the input adapt. As before this means that after motion offset the previously active units receive less excitatory input and more inhibitory input. As during adaptation this signal is unimodal. Also, the unit at stage 2 coding the opposite direction to that of the stimulus receives 842 act Stage 1 60 120 180 direction 240 Stage 2 60 120 180 direction 240 A. GRUNEWALD Figure 5: Simulation of two close directions of motion. Stage 2 of the network model registers unitary motion and a unitary MAE. less inhibition and more excitation. Through the recurrent activities within stage 2, that unit gets maximally activated. A unimodal MAE results. 3.3 TRANSPARENT MOTION: DIFFERENT DIRECTIONS When the directions of the two populations of dots in a transparent motion display are sufficiently distinct, the distribution of activities at stage 2 is no longer unimodal, but bimodal. Thus, recurrent inhibition leads to activation of two units at stage 2. They correspond to the two stimulus directions. A simulation is shown in Figure 6. Stage 1 act Stage 2 60 120 180 direction 240 Figure 6: Simulation of two distinct directions of motion. Stage 2 of the model registers transparent motion during adaptation, but the MAE is unidirectional. Feedforward inhibition is tuned much broader than feedforward excitation, and as a consequence the inhibitory signal during adaptation is unimodal, peaking at the unit of stage 2 coding the opposite direction of the average of the two previously active directions. Therefore that unit receives the least amount of inhibition after motion offset. It receives the same activity from stage 1 as units coding nearby directions, since the corresponding weights at stage 1 did not adapt. Due to recurrent activities at stage 2 that unit becomes active: non-transparent motion is registered. A Model of Transparent Motion and Non-transparent Motion Aftereffects 843 4 DISCUSSION Recently Snowden, Treue, Erickson, and Andersen (1991) have studied the effect of transparent motion stimuli on neurons in areas VI and MT of macaque monkey. They simultaneously presented two populations of dots, one of which was moving in the preferred direction of the neuron under study, and the other population was moving in a different direction. They found that neurons in VI were barely affected by the second population of dots. Neurons in MT, on the other hand, were inhibited when the direction of the second population differed from the preferred direction, and inhibition was maximal when the second population was moving opposite to the preferred direction. These results support key mechanisms of the model. At stage 1 there is no interaction between opposing directions of motion. The feedforward inhibition between stages 1 and 2 is maximal between opposite directions. Thus activities of units at stage 1 parallel neural activities recorded at VI, and activities of units at stage 2 parallels those neural activities recorded in area MT. Acknowledgments This research was carried out under HFSP grant SF-354/94. Reference Grossberg, S. (1973). Contour enhancement, short term memory, and constancies in reverberating neural networks. Studies in Applied Mathematics, LII, 213-257. Hiris, E., & Blake, R. (1992). Another perspective in the visual motion aftereffect. Proceedings of the National Academy of Sciences USA, 89, 9025-9028. Mather, G. (1980). The movement aftereffect and a distribution-shift model for coding the direction of visual movement. Perception, 9, 379-392. Raymond, J. E. (1993). Movement direction analysers: independence and bandwidth. Vision Research, 33(5/6), 767-775. Snowden, R. J., Treue, S., Erickson, R. G., & Andersen, R. A. (1991). The response of area MT and VI neurons to transparent motion. Journal of Neuroscience, 11 (9), 2768-2785. Sutherland, N. S. (1961). Figural after-effects and apparent size. Quarterly Journal of Experimental Psychology, 13, 222-228. Verstraten, F. A. J., Fredericksen, R. E., & van de Grind, W. A. (1994). Movement aftereffect of bi-vectorial transparent motion. Vision Research, 34, 349-358. Williams, D., Tweten, S., & Sekuler, R. (1991). Using metamers to explore motion perception. Vision Research, 31 (2), 275-286. Williams, D. W., & Sekuler, R. (1984). Coherent global motion percept from stochastic local motions. Vision Research, 24 (1), 55-62. Wohlgemuth, A. (1911). On the aftereffect of seen movement. British Journal of Psychology (Monograph Supplement), 1, 1-117.
|
1995
|
77
|
1,125
|
Improved Gaussian Mixture Density Estimates Using Bayesian Penalty Terms and Network Averaging Dirk Ormoneit Institut fur Informatik (H2) Technische Universitat Munchen 80290 Munchen, Germany ormoneit@inJormatik.tu-muenchen.de Abstract Volker Tresp Siemens AG Central Research 81730 Munchen, Germany Volker. Tresp@zJe.siemens.de We compare two regularization methods which can be used to improve the generalization capabilities of Gaussian mixture density estimates. The first method uses a Bayesian prior on the parameter space. We derive EM (Expectation Maximization) update rules which maximize the a posterior parameter probability. In the second approach we apply ensemble averaging to density estimation. This includes Breiman's "bagging" , which recently has been found to produce impressive results for classification networks. 1 Introduction Gaussian mixture models have recently attracted wide attention in the neural network community. Important examples of their application include the training of radial basis function classifiers, learning from patterns with missing features, and active learning. The appeal of Gaussian mixtures is based to a high degree on the applicability of the EM (Expectation Maximization) learning algorithm, which may be implemented as a fast neural network learning rule ([Now91], [Orm93]). Severe problems arise, however, due to singularities and local maxima in the log-likelihood function. Particularly in high-dimensional spaces these problems frequently cause the computed density estimates to possess only relatively limited generalization capabilities in terms of predicting the densities of new data points. As shown in this paper, considerably better generalization can be achieved using regularization. Improved Gaussian Mixture Density Estimates Using Bayesian Penalty Terms 543 We will compare two regularization methods. The first one uses a Bayesian prior on the parameters. By using conjugate priors we can derive EM learning rules for finding the MAP (maximum a posteriori probability) parameter estimate. The second approach consists of averaging the outputs of ensembles of Gaussian mixture density estimators trained on identical or resampled data sets. The latter is a form of "bagging" which was introduced by Breiman ([Bre94]) and which has recently been found to produce impressive results for classification networks. By using the regularized density estimators in a Bayes classifier ([THA93], [HT94], [KL95]), we demonstrate that both methods lead to density estimates which are superior to the unregularized Gaussian mixture estimate. 2 Gaussian Mixtures and the EM Algorithm Consider the lroblem of estimating the probability density of a continuous random vector x E 'R based on a set x* = {xk 11 S k S m} of iid. realizations of x. As a density model we choose the class of Gaussian mixtures p(xle) = L:7=1 Kip(xli, pi, Ei ), where the restrictions Ki ~ 0 and L:7=1 Kj = 1 apply. e denotes the parameter vector (Ki' Iti, Ei)i=1. The p(xli, Pi, Ei ) are multivariate normal densities: p( xli, Pi , Ei) = (271")- 41Ei 1- 1/ 2 exp [-1/2(x - Pi)tEi 1 (x - Iti)] . The Gaussian mixture model is well suited to approximate a wide class of continuous probability densities. Based on the model and given the data x*, we may formulate the log-likelihood as lee) = log [rrm p(xkle)] = ",m log "'~ Kip(xkli, Pi, Ei). k=l .L...".k=1 .L...".J=l Maximum likelihood parameter estimates e may efficiently be computed with the EM (Expectation Maximization) algorithm ([DLR77]). It consists of the iterative application of the following two steps: 1. In the E-step, based on the current parameter estimates, the posterior probability that unit i is responsible for the generation of pattern xk is estimated as (1) 2. In the M-step, we obtain new parameter estimates (denoted by the prime): ~m hk k , wk-l i X Pi = ~m hi wl=l i , 1 L m k K · = h· J m k=1 J (2) (3) ~.' _ L:~1 hf(xk - pD(xk - pDt L.JJ m I L:l=l hi (4) Note that K~ is a scalar, whereas p~ denotes a d-dimensional vector and E/ is a d x d matrix. It is well known that training neural networks as predictors using the maximum likelihood parameter estimate leads to overfitting. The problem of overfitting is even more severe in density estimation due to singularities in the log-likelihood function. Obviously, the model likelihood becomes infinite in a trivial way if we concentrate all the probability mass on one or several samples of the training set. 544 D. ORMONEIT, V. TRESP In a Gaussian mixture this is just the case if the center of a unit coincides with one of the data points and E approaches the zero matrix. Figure 1 compares the true and the estimated probability density in a toy problem. As may be seen, the contraction of the Gaussians results in (possibly infinitely) high peaks in the Gaussian mixture density estimate. A simple way to achieve numerical stability is to artificially enforce a lower bound on the diagonal elements of E. This is a very rude way of regularization, however, and usually results in low generalization capabilities. The problem becomes even more severe in high-dimensional spaces. To yield reasonable approximations, we will apply two methods of regularization, which will be discussed in the following two sections. Figure 1: True density (left) and unregularized density estimation (right). 3 Bayesian Regularization In this section we propose a Bayesian prior distribution on the Gaussian mixture parameters, which leads to a numerically stable version of the EM algorithm. We first select a family of prior distributions on the parameters which is conjugate*. Selecting a conjugate prior has a number of advantages. In particular, we obtain analytic solutions for the posterior density and the predictive density. In our case, the posterior density is a complex mixture of densitiest . It is possible, however, to derive EM-update rules to obtain the MAP parameter estimates. A conjugate prior of a single multivariate normal density is a product of a normal density N(JLilft,1]-lEi) and a Wishart density Wi(E;lla,,8) ([Bun94]). A proper conjugate prior for the the mixture weightings '" = ("'1, ... , "'n) is a Dirichlet density D("'hV. Consequently, the prior of the overall Gaussian mixture is the product D(",lr) il7=1 N(JLilil, 71-1Ei)Wi(E;1Ia, ,8). Our goal is to find the MAP parameter estimate, that is parameters which assume the maximum of the log-posterior Ip(S) 2:=~=1 log 2:=;=1 "'iP(X k Ii, JLi, Ei ) + log D("'lr) + 2:=;=1 [logN(JLilft, 71-1Ei) + log Wi(E;lla, ,8)]. As in the unregularized case, we may use the EM-algorithm to find a local maximum • A family F of probability distributions on 0 is said to be conjugate if, for every 1r E F, the posterior 1r(0Ix) also belongs to F ([Rob94]). tThe posterior distribution can be written as a sum of nm simple terms. tThose densities are defined as follows (b and c are normalizing constants): D(1I:17) N(Il.lp,1,-IE.) W i(Ei l la,,8) bIIn 11:7,-1, with 11:, ~ 0 and ",n 11:. = 1 .=1 ~.=l (21r)-i 11,-IE;I-l/2 exp [-~(Il' - Mt Ei1(1l' - M] = cIEillo-Cd+l)/2 exp [-tr(,8Ei 1)] • Improved Gaussian Mixture Density Estimates Using Bayesian Penalty Terms 545 of Ip(8). The E-step is identical to (1). The M-step becomes "m hk + 1 "m hk k + A , L.."k-l i ri (5) ,L.."k=l i x '1J1. "'i = "n J1.i = "m hi m + L.."i=l ri - n L..,,1=1 i + 11 (6) E~ = 2:;-1 hf(xk - J1.D(xk - J1.Dt + 11(J1.i - jJ.)(J1.i - jJ.)t + 2f3 I 2:~1 h~ + 20: - d (7) As typical for conjugate priors, prior knowledge corresponds to a set of artificial training data which is also reflected in the EM-update equations. In our experiments, we focus on a prior on the variances which is implemented by f3 =F 0, where o denotes the d x d zero matrix. All other parameters we set to "neutral" values: ri=l'v'i : l::;i::;n, 0:= (d+I)/2, 11=0, f3=iJld ld is the d x d unity matrix. The choice of 0: introdu~es a bias which favors large variances§. The effect of various values of the scalar f3 on the density estimate is illustrated in figure 2. Note that if iJ is chosen too small, overfitting still occurs. If it is chosen to large, on the other hand, the model is too constraint to recognize the underlying structure. Figure 2: Regularized density estimates (left: iJ = 0.05, right: 'iJ = 0.1). Typically, the optimal value for iJ is not known a priori. The simplest procedure consists of using that iJ which leads to the best performance on a validation set, analogous to the determination of the optimal weight decay parameter in neural network training. Alternatively, iJ might be determined according to appropriate Bayesian methods ([Mac9I]). Either way, only few additional computations are required for this method if compared with standard EM. 4 Averaging Gaussian Mixtures In this section we discuss the averaging of several Gaussian mixtures to yield improved probability density estimation. The averaging over neural network ensembles has been applied previously to regression and classification tasks ([PC93]). There are several different variants on the simple averaging idea. First, one may train all networks on the complete set of training data. The only source of disagreement between the individual predictions consists in different local solutions found by the likelihood maximization procedure due to different starting points. Disagreement is essential to yield an improvement by averaging, however, so that this proceeding only seems advantageous in cases where the relation between training data and weights is extremely non-deterministic in the sense that in training, §If A is distributed according to Wi(AIO', (3), then E[A- 1 ] = (0' - (d + 1)/2)-1 {3. In our case A is B;-I, so that E[Bi] -+ 00 • {3 for 0' -+ (d + 1)/2. 546 D. ORMONEIT, V. TRESP different solutions are found from different random starting points. A straightforward way to increase the disagreement is to train each network on a resampled version of the original data set. If we resample the data without replacement, the size of each training set is reduced, in our experiments to 70% of the original. The averaging of neural network predictions based on resampling with replacement has recently been proposed under the notation "bagging" by Breiman ([Bre94]), who has achieved dramatic.ally improved results in several classification tasks. He also notes, however, that an actual improvement of the prediction can only result if the estimation procedure is relatively unstable. As discussed, this is particularly the case for Gaussian mixture training. We therefore expect bagging to be well suited for our task. 5 Experiments and Results To assess the practical advantage resulting from regularization, we used the density estimates to construct classifiers and compared the resulting prediction accuracies using a toy problem and a real-world problem. The reason is that the generalization error of density estimates in terms of the likelihood based on the test data is rather unintuitive whereas performance on a classification problem provides a good impression of the degree of improvement. Assume we have a set of N labeled data z* = {(xk, lk)lk = 1, ... , N}, where lk E Y = {I, ... , C} denotes the class label of each input xk . A classifier of new inputs x is yielded by choosing the class I with the maximum posterior class-probability p(llx). The posterior probabilities may be derived from the class-conditional data likelihood p(xll) via Bayes theorem: p(llx) = p(xll)p(l)/p(x) ex p(xll)p(l). The resulting partitions ofthe input space are optimal for the true p(llx). A viable way to approximate the posterior p(llx) is to estimate p(xll) and p(l) from the sample data. 5.1 Toy Problem In the toy classification problem the task is to discriminate the two classes of circulatory arranged data shown in figure 3. We generated 200 data points for each class and subdivided them into two sets of 100 data points. The first was used for training, the second to test the generalization performance. As a network architecture we chose a Gaussian mixture with 20 units. Table 1 summarizes the results, beginning with the unregularized Gaussian mixture which is followed by the averaging and the Bayesian penalty approaches. The three rows for averaging correspond to the results yielded without applying resampling (local max.), with resampling withFigure 3: Toy Classification Task. Improved Gaussian Mixture Density Estimates Using Bayesian Penalty Terms 547 out replacement (70% subsets), and with resampling with replacement (bagging). The performances on training and test set are measured in terms of the model loglikelihood. Larger values indicate a better performance. We report separate results for dass A and B, since the densities of both were estimated separately. The final column shows the prediction accuracy in terms of the percentage of correctly classified data in the test set. We report the average results from 20 experiments. The numbers in brackets denote the standard deviations u of the results. Multiplying u with T19;95%/v'20 = 0.4680 yields 95% confidence intervals. The best result in each category is underlined. Algorithm Log-Likelihood Training Test Accuracy A B A B I unreg. -120.8 (13.3) -120.4 (10.8) -224.9 (32.6) -241.9 (34.1) 80.6'70 (2.8) Averaging: local max. -115.6 (6.0) -112.6 (6.6) -200.9 (13.9) -209.1 (16.3) 81.8% (3.1) 70% subset -106.8 (5.8) -105.1 (6.7) -188.8 (9.5) -196.4 (11.3) 83.2% (2.9) bagging -83.8 (4.9) -83.1 (7.1) -194.2 (7.3) -200.1 (11.3) 82.6% (3.4) Penalty: [3 = 0.01 -149.3 (18.5) -146.5 (5.9) -186.2 (13.9) -182.9 (11.6) 83.1% (2.9) [3 = 0.02 -156.0 (16.5) -153.0 (4.8) -177.1 (11.8) -174.9 (7.0) 84.4% (6.3) [3 = 0.05 -173.9 (24.3) -167.0 (15.8) -182.0 (20.1) -173.9 (14.3) 81.5% (5.9) [3 = 0.1 -183.0 (21.9) -181.9 (21.1) -184.6 (21.0) -182.5 (21.1) 78.5% (5.1) Table 1: Performances in the toy classification problem . As expected, all regularization methods outperform the maximum likelihood approach in terms of correct classification. The performance of the Bayesian regularization is hereby very sensitive to the appropriate choice of the regularization parameter (3. Optimality of (3 with respect to the density prediction and oytimality with respect to prediction accuracy on the test set roughly coincide (for (3 = 0.02). A veraging is inferior to the Bayesian approach if an optimal {3 is chosen. 5.2 BUPA Liver Disorder Classification As a second task we applied our methods to a real-world decision problem from the medical environment. The problem is to detect liver disorders which might arise from excessive alcohol consumption. Available information consists of five blood tests as well as a measure of the patients' daily alcohol consumption. We subdivided the 345 available samples into a training set of 200 and a test set of 145 samples. Due to the relatively few data we did not try to determine the optimal regularization parameter using a validation process and will report results on the test set for different parameter values. Algorithm unregularized Bayesian penalty ({3 = 0.05) Bayesian penalty «(3 = 0.10) Bayesian penal ty (3 = 0.20 averaging local maxima averaging (70 % subset) averaging (bagging) Accuracy 64.8 % 65.5 % 66.9 % 61.4 % 65.5 0 72.4 % 71.0 % Table 2: Performances in the liver disorder classification problem. 548 D. ORMONEIT. V. TRESP The results of our experiments are shown in table 2. Again, both regularization methods led to an improvement in prediction accuracy. In contrast to the toy problem, the averaged predictor was superior to the Bayesian approach here. Note that the resampling led to an improvement of more than five percent points compared to unresampled averaging. 6 Conclusion We proposed a Bayesian and an averaging approach to regularize Gaussian mixture density estimates. In comparison with the maximum likelihood solution both approaches led to considerably improved results as demonstrated using a toy problem and a real-world classification task. Interestingly, none of the methods outperformed the other in both tasks. This might be explained with the fact that Gaussian mixture density estimates are particularly unstable in high-dimensional spaces with relatively few data. The benefit of averaging might thus be greater in this case. A veraging proved to be particularly effective if applied in connection with resampIing of the training data, which agrees with results in regression and classification tasks. If compared to Bayesian regularization, averaging is computationally expensive. On the other hand, Baysian approaches typically require the determination of hyper parameters (in our case 13), which is not the case for averaging approaches. References [Bre94] L. Breiman. Bagging predictors. Technical report, UC Berkeley, 1994. [Bun94] W. Buntine. Operations for learning with graphical models. Journal of Artificial Intelligence Research, 2:159-225, 1994. [DLR77] A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum likelihood from incomplete data via the EM algorithm. J. Royal Statistical Society B, 1977. [HT94] T. Hastie and R. Tibshirani. Discriminant analysis by gaussian mixtures. Technical report, AT&T Bell Labs and University of Toronto, 1994. [KL95] N. Kambhatla and T. K. Leen. Classifying with gaussian mixtures and clusters. In Advances in Neural Information Processing Systems 7. Morgan Kaufman, 1995. [Mac91] D. MacKay. Bayesian Modelling and Neural Networks. PhD thesis, California Institute of Technology, Pasadena, 1991. [Now91] S. J. Nowlan. Soft Competitive Adaption: Neural Network Learning Algorithms based on Fitting Statistical Mixtures. PhD thesis, School of Computer Science, Carnegie Mellon University, Pittsburgh, 1991. [Orm93] D. Ormoneit. Estimation of probability densities using neural networks. Master's thesis, Technische Universitiit Munchen, 1993. [PC93] M. P. Perrone and L. N. Cooper. When networks disagree: Ensemble methods for hybrid Neural networks. In Neural Networks for Speech and Image Processing. Chapman Hall, 1993. [Rob94] C. P. Robert. The Bayesian Choice. Springer-Verlag, 1994. [THA93] V. Tresp, J. Hollatz, and S. Ahmad. Network structuring and training using rule-based knowledge. In Advances in Neural Information Processing Systems 5. Morgan Kaufman, 1993.
|
1995
|
78
|
1,126
|
A Practical Monte Carlo Implementation of Bayesian Learning Carl Edward Rasmussen Department of Computer Science University of Toronto Toronto, Ontario, M5S 1A4, Canada carl@cs.toronto.edu Abstract A practical method for Bayesian training of feed-forward neural networks using sophisticated Monte Carlo methods is presented and evaluated. In reasonably small amounts of computer time this approach outperforms other state-of-the-art methods on 5 datalimited tasks from real world domains. 1 INTRODUCTION Bayesian learning uses a prior on model parameters, combines this with information from a training set, and then integrates over the resulting posterior to make predictions. With this approach, we can use large networks without fear of overfitting, allowing us to capture more structure in the data, thus improving prediction accuracy and eliminating the tedious search (often performed using cross validation) for the model complexity that optimises the bias/variance tradeoff. In this approach the size of the model is limited only by computational considerations. The application of Bayesian learning to neural networks has been pioneered by MacKay (1992), who uses a Gaussian approximation to the posterior weight distribution. However, the Gaussian approximation is poor because of multiple modes in the posterior. Even locally around a mode the accuracy of the Gaussian approximation is questionable, especially when the model is large compared to the amount of training data. Here I present and test a Monte Carlo method (Neal, 1995) which avoids the Gaussian approximation. The implementation is complicated, but the user is not required to have extensive knowledge about the algorithm. Thus, the implementation represents a practical tool for learning in neural nets. A Practical Monte Carlo Implementation of Bayesian Learning 599 1.1 THE PREDICTION TASK The training data consists of n examples in the form of inputs x = {x(i)} and corresponding outputs y = {y(i)} where i = 1 ... n. For simplicity we consider only real-valued scalar outputs. The network is parametrised by weights w, and hyperparameters h that control the distributions for weights, playing a role similar to that of conventional weight decay. Weights and hyperparameters are collectively termed 0, and the network function is written as F/I (x), although the function value is only indirectly dependent on the hyperparameters (through the weights). Bayes' rule gives the posterior distribution for the parameters in terms of the likelihood, p(ylx, 0), and prior, p(O): (Olx ) = p(O)p(ylx, O) p ,y p(ylx) To minimize the expected squared error on an unseen test case with input x(n+l), we use the mean prediction (1) 2 MONTE CARLO SAMPLING The following implementation is due to Neal (1995). The network weights are updated using the hybrid Monte Carlo method (Duane et al. 1987). This method combines the Metropolis algorithm with dynamical simulation. This helps to avoid the random walk behavior of simple forms of Metropolis, which is essential if we wish to explore weight space efficiently. The hyperparameters are updated using Gibbs sampling. 2.1 NETWORK SPECIFICATION The networks used here are always of the same form: a single linear output unit, a single hidden layer of tanh units and a task dependent number of input units. All layers are fully connected in a feed forward manner (including direct connections from input to output). The output and hidden units have biases. The network priors are specified in a hierarchical manner in terms of hyperparameters; weights of different kinds are divided into groups, each group having it's own prior. The output-bias is given a zero-mean Gaussian prior with a std. dev. of u = 1000, so it is effectively unconstrained. The hidden-biases are given a two layer prior: the bias b is given a zero-mean Gaussian prior b '" N(O, ( 2 ); the value of u is specified in terms of precision r = u- 2 , which is given a Gamma prior with mean p = 400 (corresponding to u = 0.05) and shape parameter a = 0.5; the Gamma density is given by p(r) '" Gamma(p, a) ex: r Ol / 2- 1 exp( -ra/2p). Note that this type of prior introduces a dependency between the biases for different hidden units through the common r. The prior for the hidden-to-output weights is identical to the prior for the hidden-biases, except that the variance of these weights under the prior is scaled down by the square root of the number of hidden units, such that the network output magnitude becomes independent of the number of hidden units. The noise variance is also given a Gamma prior with these parameters. 600 C. E. RASMUSSEN The input-to-hidden weights are given a three layer prior: again each weight is given a zero-mean Gaussian prior w rv N(O, (12); the corresponding precision for the weights out of input unit i is given a Gamma prior with a mean J.l and a shape parameter a1 = 0.5: Ti rv Gamma(J.l, a1). The mean J.l is determined on the top level by a Gamma distribution with mean and shape parameter ao = 1: J.li rv Gamma(400,ao). The direct input-to-output connections are also given this prior. The above-mentioned 3 layer prior incorporates the idea of Automatic Relevance Determination (ARD), due to MacKay and Neal, and discussed in Neal (1995) . The hyperparameters, Ti, associated with individual inputs can adapt according to the relevance of the input; for an unimportant input, Ti can grow very large (governed by the top level prior), thus forcing (1i and the associated weights to vanish. 2.2 MONTE CARLO SPECIFICATION Sampling from the posterior weight distribution is performed by iteratively updating the values of the network weights and hyperparameters. Each iteration involves two components: weight updates and hyperparameter updates. A cursory description of these steps follows. 2.2.1 Weight Updates Weight updates are done using the hybrid Monte Carlo method. A fictitious dynamical system is generated by interpreting weights as positions, and augmenting the weights w with momentum variables p. The purpose of the dynamical system is to give the weights "inertia" so that slow random walk behaviour can be avoided during exploration of weight space. The total energy, H, of the system is the sum of the kinetic energy, I<, (a function of the momenta) and the potential energy, E. The potential energy is defined such that p(w) ex exp( -E). We sample from the joint distribution for wand p given by p(w,p) ex exp(-E - I<), under which the marginal distribution for w is given by the posterior. A sample of weights from the posterior can therefore be obtained by simply ignoring the momenta. Sampling from the joint distribution is achieved by two steps: 1) finding new points in phase space with near-identical energies H by simulating the dynamical system using a discretised approximation to Hamiltonian dynamics, and 2) changing the energy H by doing Gibbs sampling for the momentum variables. Hamiltonian Dynamics. Hamilton's first order differential equations for Hare approximated by a series of discrete first order steps (specifically by the leapfrog method). The first derivatives of the network error function enter through the derivative of the potential energy, and are computed using backpropagation. In the original version of the hybrid Monte Carlo method the final position is then accepted or rejected depending on the final energy H'" (which is not necessarily equal to the initial energy H because of the discretisation). Here we use a modified version that uses an average over a window of states instead. The step size of the discrete dynamics should be as large as possible while keeping the rejection rate low. The step sizes are set individually using several heuristic approximations, and scaled by an overall parameter c. We use L = 200 iterations, a window size of 20 and a step size of c = 0.2 for all simulations. Gibbs Sampling for Momentum Variables. The momentum variables are updated using a modified version of Gibbs sampling, allowing the energy H to change. A "persistence" of 0.95 is used; the new value of the momentum is a weighted sum of the previous value (weight 0.95) and the value obtained by Gibbs sampling (weight (1 - 0.952)1/2). With this form of persistence, the momenta A Practical Monte Carlo Implementation of Bayesian Learning 601 changes approx. 20 times more slowly, thus increasing the "inertia" of the weights, so as to further help in avoiding random walks. Larger values of the persistence will further increase the weight inertia, but reduce the rate of exploration of H. The advantage of increasing the weight inertia in this way rather than by increasing L is that the hyperparameters are updated at shorter intervals, allowing them to adapt to the rapidly changing weights. 2.2.2 Hyperparameter Updates The hyperparameters are updated using Gibbs sampling. The conditional distributions for the hyperparameters given the weights are of the Gamma form, for which efficient generators exist, except for the top-level hyperparameter in the case of the 3 layer priors used for the weights from the inputs; in this case the conditional distribution is more complicated and a form of rejection sampling is employed. 2.3 NETWORK TRAINING AND PREDICTION The network training consists of two levels of initialisation before sampling for networks used for prediction. At the first level of initialisation the hyperparameters (variance of the Gaussians) are kept constant at 1, allowing the weights to grow during 1000 leapfrog iterations. Neglecting this phase can cause the network to get caught for a long time in a state where weights and hyperparameters are both very small. The scheme described above is then invoked and run for as long as desired, eventually producing networks from the posterior distribution. The initial 1/3 of these nets are discarded, since the algorithm may need time to reach regions of high posterior probability. Networks sampled during the remainder of the run are saved for making predictions. The predictions are made using an average of the networks sampled from the posterior as an approximation to the integral in eq. (1). Since the output unit is linear the final prediction can be seen as coming from a huge (fully connected) ensemble net with appropriately scaled output weights. All the results reported here were for ensemble nets with 4000 hidden units. The size of the individual nets is given by the rule that we want at least as many network parameters as we have training examples (with a lower limit of 4 hidden units). We hope thereby to be well out of the underfitting region. Using even larger nets would probably not gain us much (in the face of the limited training data) and is avoided for computational reasons. All runs used the parameter values given above. The only check that is necessary is that the rejection rate stays low, say below 5%; if not, the step size should be lowered. In all runs reported here, c = 0.2 was adequate. The parameters concerning the Monte Carlo method and the network priors were all selected based on intuition and on experience with toy problems. Thus no parameters need to be set by the user. 3 TESTS The performance of the algorithm was evaluated by comparing it to other state-ofthe-art methods on 5 real-world regression tasks. All 5 data sets have previously been studied using a 10-way cross-validation scheme (Quinlan 1993). The tasks in these domains is to predict price or performance of an object from various discrete and real-valued attributes. For each domain the data is split into two sets of roughly equal size, one for training and one for testing. The training data is 602 C. E. RASMUSSEN further subdivided into full-, half-, quarter- and eighth-sized subsets, 15 subsets in total. Networks are trained on each of these partitions, and evaluated on the large common test set. On the small training sets, the average performance and one std. dev. error bars on this estimate are computed. 3.1 ALGORITHMS The Monte Carlo method was compared to four other algorithms. For the three neural network methods nets with a single hidden layer and direct input-output connections were used. The Monte Carlo method was run for 1 hour on each of the small training sets, and 2,4 and 8 hours respectively on the larger training sets. All simulations were done on a 200 MHz MIPS R4400 processor. The Gaussian Process method is described in a companion paper (Williams & Rasmussen 1996). The Evidence method (MacKay 1992) was used for a network with separate hyperparameters for the direct connections, the weights from individual inputs (ARD), hidden biases, and output biases. Nets were trained using a conjugate gradient method, allowing 10000 gradient evaluations (batch) before each of 6 updates of the hyperparameters. The network Hessian was computed analytically. The value of the evidence was computed without compensating for network symmetries, since this can lead to a vastly over-estimated evidence for big networks where the posterior Gaussians from different modes overlap. A large number of nets were trained for each task, with the number of hidden units computed from the results of previous nets by the following heuristics: The min and max number of hidden units in the 20% nets with the highest evidences were found. The new architecture is picked from a Gaussian (truncated at 0) with mean (max - min)/2 and std. dev. 2 + max - min, which is thought to give a reasonable trade-off between exploration and exploitation. This procedure is run for 1 hour of cpu time or until more than 1000 nets have been trained. The final predictions are made from an ensemble of the 20% (but a maximum of 100) nets with the highest evidence. An ensemble method using cross-validation to search over a 2-dimensional grid for the number of hidden units and the value of a single weight decay parameter has been included, as an attempt to have a thorough version of "common practise". The weight decay parameter takes on the values 0, 0.01, 0.04, 0.16, 0.64 and 2.56. Up to 6 sizes of nets are used, from 0 hidden units (a linear model) up to a number that gives as many weights as training examples. Networks are trained with a conjugent gradient method for 10000 epochs on each of these up to 36 networks, and performance was monitored on a validation set containing 1/3 of the examples, selected at random. This was repeated 5 times with different random validation sets, and the architecture and weight decay that did best on average was selected. The predictions are made from an ensemble of 10 nets with this architecture, trained on the full training set. This algorithm took several hours of cpu time for the largest training sets. The Multivariate Adaptive Regression Splines (MARS) method (Friedman 1991) was included as a non-neural network approach. It is possible to vary the maximum number of variables allowed to interact in the additive components of the model. It is common to allow either pairwise or full interactions. I do not have sufficient experience with MARS to make this choice. Therefore, I tried both options and reported for each partition on each domain the best performance based on the test error, so results as good as the ones reported here might not be obtainable in practise. All other parameters of MARS were left at their default values. MARS always required less than 1 minute of cpu time. A Practical Monte Carlo Implementation of Bayesian Learning 603 2 1.5 1 0.5 Auto price 0* + x o~------~----~----~--0.6 0.5 0.4 0.3 0.2 0.1 10 20 40 House t >«1>* + IS! 80 o~~----~------~----~-32 64 128 256 Servo 1 0.8 0.6 0.4 0.2 OtIS! X * o~~------~----~----~--11 22 44 88 Cpu 0.6 0.5 0.4 + 0.3 o IS! 0.2 X * 0.1 OL-~----~------~----~-13 0.25 0.2 0.15 0.1 0.05 26 52 Mpg 104 * Xo+ IS! OL-~----~----~----~-24 48 96 192 Geometric mean x Monte Carlo o Gaussian Evidence + Backprop * MARS IS! Gaussian Process 0.283 0.364 0.339 0.371 0.304 Figure 1: Squared error on test cases for the five algorithms applied to the five problems. Errors are normalized with respect to the variance on the test cases. The x-axis gives the number of training examples; four different set sizes were used on each domain. The error bars give one std. dev. for the distribution of the mean over training sets. No error bar is given for the largest size, for which only a single training set was available. Some of the large error bars are cut of at the top. MARS was unable to run on the smallest partitions from the Auto price and the servo domains; in these cases the means of the four other methods were used in the reported geometric mean for MARS. 604 C. E. RASMUSSEN Table 1: Data Sets domain # training cases # test cases # binary inputs # real inputs Auto Price 80 79 0 16 Cpu 104 105 0 6 House 256 250 1 12 Mpg 192 200 6 3 Servo 88 79 10 2 3.2 PERFORMANCE The test results are presented in fig. 1. On the servo domain the Monte Carlo method is uniformly better than all other methods, although the difference should probably not always be considered statistically significant. The Monte Carlo method generally does well for the smallest training sets. Note that no single method does well on all these tasks. The Monte Carlo method is never vastly out-performed by the other methods. The geometric mean of the performances over all 5 domains for the the 4 different training set sizes is computed. Assuming a Gaussian distribution of prediction errors, the log of the error variance can (apart from normalising constants) be interpreted as the amount of information unexplained by the models. Thus, the log of the geometric means in fig. 1 give the average information unexplained by the models. According to this measure the Monte Carlo method does best, closely followed by the Gaussian Process method. Note that MARS is the worst, even though the decision between pairwise and full interactions were made on the basis of the test errors. 4 CONCLUSIONS I have outlined a black-box Monte Carlo implementation of Bayesian learning in neural networks, and shown that it has an excellent performance. These results suggest that Monte Carlo based Bayesian methods are serious competitors for practical prediction tasks on data limited domains. Acknowledgements I am grateful to Radford Neal for his generosity with insight and software. This research was funded by a grant to G. Hinton from the Institute for Robotics and Intelligent Systems. References S. Duane, A. D. Kennedy, B. J. Pendleton & D. Roweth (1987) "Hybrid Monte Carlo", Physics Letters B, vol. 195, pp. 216-222. J. H. Friedman (1991) "Multivariate adaptive regression splines" (with discussion), Annals of Statistics, 19,1-141 (March). Source: http://lib.stat.cmu.edu/general/mars3.5. D. J. C. MacKay (1992) "A practical Bayesian framework for backpropagation networks", Neural Computation, vol. 4, pp. 448- 472. R. M. Neal (1995) Bayesian Learning for Neural Networks, PhD thesis, Dept. of Computer Science, University of Toronto, ftp: pub/radford/thesis. ps. Z from ftp. cs . toronto. edu. J. R. Quinlan (1993) "Combining instance-based and model-based learning", Proc. ML '93 (ed P.E. Utgoff), San Mateo: Morgan Kaufmann. C. K. I. Williams & C. E. Rasmussen (1996). "Regression with Gaussian processes", NIPS 8, editors D. Touretzky, M. Mozer and M. Hesselmo. (this volume).
|
1995
|
79
|
1,127
|
On the Computational Power of Noisy Spiking Neurons Wolfgang Maass Institute for Theoretical Computer Science, Technische Universitaet Graz Klosterwiesgasse 32/2, A-8010 Graz, Austria, e-mail: maass@igi.tu-graz.ac.at Abstract It has remained unknown whether one can in principle carry out reliable digital computations with networks of biologically realistic models for neurons. This article presents rigorous constructions for simulating in real-time arbitrary given boolean circuits and finite automata with arbitrarily high reliability by networks of noisy spiking neurons. In addition we show that with the help of "shunting inhibition" even networks of very unreliable spiking neurons can simulate in real-time any McCulloch-Pitts neuron (or "threshold gate"), and therefore any multilayer perceptron (or "threshold circuit") in a reliable manner. These constructions provide a possible explanation for the fact that biological neural systems can carry out quite complex computations within 100 msec. It turns out that the assumption that these constructions require about the shape of the EPSP's and the behaviour of the noise are surprisingly weak. 1 Introduction We consider networks that consist of a finite set V of neurons, a set E ~ V x V of synapses, a weightwu,v ~ 0 and a response junctioncu,v : R+ -+ R for each synapse 212 W.MAASS (u,v) E E (where R+ := {x E R: x ~ O}), and a threshold/unction Sv : R+ --t R+ for each neuron v E V. If F u ~ R + is the set of firing times of a neuron u, then the potential at the trigger zone of neuron v at time t is given by Pv(t) := L L wu,v' u : (u, v) E EsE Fu : s < t eu,v(t - s). The threshold function Sv(t - t') quantifies the "reluctance" of v to fire again at time t, if its last previous firing was at time t'. We assume that Sv(O) E (0,00), Sv(x) = 00 for x E (0, 'TreJ] (for some constant 'TreJ > 0, the "absolute refractory period"), and sup{Sv(x) : X ~ 'T} < 00 for any'T > 'TreJ. In a deterministic model for a spiking neuron (Maass, 1995a, 1996) one can assume that a neuron v fires exactly at those time points t when Pv(t) reaches (from below) the value Sv(t - t'). We consider in this article a biologically more realistic model, where as in (Gerstner, van Hemmen, 1994) the size of the difference Pv(t)-Sv(t-t') just governs the probability that neuron v fires. The choice of the exact firing times is left up to some unknown stochastic processes, and it may for example occur that v does not fire in a time intervall during which Pv (t) - Sv(t - t') > 0, or that v fires "spontaneously" at a time t when Pv(t) -Sv(t-t') < O. We assume that (apart from their communication via potential changes) the stochastic processes for different neurons v are independent. It turns out that the assumptions that one has to make about this stochastic firing mechanism in order to prove our results are surprisingly weak. We assume that there exist two arbitrary functions L, U : R X R+ ----1 [0,1] so that L(~, i) provides a lower bound (and U(~, i) provides an upper bound) for the probability that neuron v fires during a time intervall of length e with the property that Pv(t)-Sv(t-t') ~ ~ (respectively Pv(t)-Sv(t-t') ~ ~) for all tEl up to the next firing of v (t' denotes the last firing time of v be/ore I). We just assume about these functions Land U that they are non-decreasing in each of their two arguments (for any fixed value of the other argument), that lim U(~, i) = ° for any fixed ~~-oo i > 0, and that lim L(~, e) > 0 for allY fixed e ~ R/6 (where R is the assumed ~~OO length of the rising segment of an EPSP, see below). The neurons are allowed to be "arbitrarily noisy" in the sense that the difference lim L(~, i) lim U(~, i) ~~OO ~~-oo can be arbitrarily small. Hence our constructions also apply to neurons that exhibit persistent firing failures, and they also allow for synapses that fail with a rather high probability. Furthermore a detailed analysis of our constructions shows that we can relax the somewhat dubious assumption that the noise-distributions for different neurons are independent. Thus we are also able to deal with "systematic noise" in the distribution of firing times of neurons in a pool (e.g. caused by changes in the biochemical environment that simultaneously affect many neurons in a pool). It turns out that it suffices to assume only the following rather weak properties of the other functions involved in our model: 1) Each response function CU , I ) : R+ ----1 R is either excitatory or inhibitory (and for the sake of biological realism one may assume that each neuron u induces only one type of response). All excitatory response functions eu,v(x) have the value On the Computational Power of Noisy Spiking Neurons 213 o for x E [O,~u,v), and the value eE(X ~u ,v) for x ~ ~u,v, where ~u,v ~ 0 is the delay for this synapse between neurons u and v, and e E is the common shape of all excitatory response functions ("EPSP's))). Corresponding assumptions are made about the inhibitory response functions ("IPSP's))), whose common shape is described by some function eI : R+ -+ {x E R : x ~ O}. 2) eE is continuous, eE(O) = 0, eE(X) = 0 for all sufficiently large x, and there exists some parameter R > 0 such that eE is non-decreasing in [0, R], and some parameter p > 0 such that eE(X + R/6) ~ p + eE (x) for all x E [O,2R/3]. 3) _eI satisfies the same conditions as eE . 4) There exists a source BN- of negative "background noise", that contributes to the potential Pv(t) of each neuron v an additive term that deviates for an arbitrarily long time interval by an arbitrarily small percentage from its average value w; ~ 0 (which we can choose). One can delete this assumption if one assumes that the firing threshold of neurons can be shifted by some other mechanism. In section 3 we will assume in addition the availability of a corresponding positive background noise BN+ with average value wt ~ O. In a biological neuron tI one can interpret BN- and BN+ as the combined effect of a continuous bombardment with a very large number of IPSP's (EPSP's) from randomly firing neurons that arrive at remote synapses on the dendritic tree of v. We assume that we can choose the values of delays ~u , v and weights Wu,v, wt ,w; . We refer to all assumptions specified in this section as our "weak assumptions" about noisy spiking neurons. It is easy to see that the most frequently studied concrete model for noisy spiking neurons, the spike response model (Gerstner and van Hemmen, 1994) satisfies these weak assumptions, and is hence a special case. However not even for the more concrete spike response model (or any other model for noisy spiking neurons) there exist any rigorous results about computations in these models. In fact, one may view this article as being the first that provides results about the computational complexity of neural networks for a neuron model that is acceptable to many neurobiologistis as being reasonably realistic. In this article we only address the problem of reliable digital computing with noisy spiking neurons. For details of the proofs we refer to the forthcoming journal-version of this extended abstract. For results about analog computations with noisy spiking neurons we refer to Maass, 1995b. 2 Simulation of Boolean Circuits and Finite Automata with Noisy Spiking Neurons Theorem 1: For any deterministic finite automaton D one can construct a network N(D) consisting of any type of noisy spiking neurons that satisfy our weak assumptions, so that N(D) can simulate computations of D of any given length with arbitrarily high probability of correctness. 214 W.MAASS Idea of the proof: Since the behaviour of a single noisy spiking neuron is completely unreliable, we use instead pools A, B, ... of neurons as the basic building blocks in our construction, where all neurons v in the same pool receive approximately the same "input potential" Pv(t). The intricacies of our stochastic neuron model allow us only to employ a "weak coding" of bits, where a "1" is represented by a pool A during a time interval I, if at least PI ·IAI neurons in A fire (at least once) during I (where PI > 0 is a suitable constant), and "0" is represented if at most Po ·IAI firings of neurons occur in A during I, where Po with 0 < Po < PI is another constant (that can be chosen arbitrarily small in our construction). The described coding scheme is weak since it provides no useful upper bound (e.g. 1.5·Pl ·IAI) on the number of neurons that fire during I if A represents a "1" (nor on the number of firings of a single neuron in A). It also does not impose constraints on the exact timing of firings in A within I. However a "0" can be represented more precisely in our model, by choosing po sufficiently small. The proof of Theorem 1 shows that this weak coding of bits suffices for reliable digital computations. The idea of these simulations is to introduce artificial negations into the computation, which allow us to exploit that "0" has a more precise representation than "1". It is apparently impossible to simulate an AND-gate in a straightforward fashion for a weak coding of bits, but one can simulate a NOR-gate in a reliable manner. • Corollary 2: Any boolean function can be computed by a sufficiently large network of noisy spiking neurons (that satisfy our weak assumptions) with arbitrarily high probability of correctness. 3 Fast Simulation of Threshold Circuits via Shunting Inhibition For biologically realistic parameters, each computation step in the previously constructed network takes around 25 msec (see point b) in section 4}. However it is well-known that biological neural systems can carry out complex computations within just 100 msec (Churchland, Sejnowski, 1992). A closer inspection of the preceding construction shows, that one can simulate with the same speed also OR- and NOR-gates with a much larger fan-in than just 2. However wellknown results from theoretical computer science (see the results about the complexity class ACo in the survey article by Johnson in (van Leeuwen, 1990)) imply that for any fixed number of layers the computational power of circuits with gates for OR, NOR, AND, NOT remains very weak, even if one allows any polynomial size fan-in for such gates. In contrast to that, the construction in this section will show that by using a biologically more realistic model for a noisy spiking neuron, one can in principle simulate within 100 msec 3 or more layers of a boolean circuit that employs substantially more powerful boolean gates: threshold gates (Le. "Mc Culloch-Pitts neurons", also called "perceptrons"). The use of these gates provides a giant leap in computational On the Computational Power of Noisy Spiking Neurons 215 power for boolean circuits with a small number of layers: In spite of many years of intensive research, one has not been able to exhibit a single concrete computational problem in the complexity classes P or NP that can be shown to be not computable by a polynomial size threshold circuit with 3 layers (for threshold circuits with integer weights of unbounded size the same holds already for just 2 layers). In the neuron model that we have employed so far in this article, we have assumed (as it is common in the spike response model) that the potential Pv(t) at the trigger zone of neuron v depends linearly on all the terms Wu,v . cu,v(t - s). There exists however ample biological evidence that this assumption is not appropriate for certain types of synapses. An example are synapses that carry out shunting inhibition (see. e.g. (Abeles, 1991) and (Shepherd, 1990)). When a synapse of this type (located on the dendritic tree of a neuron v) is activated, it basically erases (through a short circuit mechanism) for a short time all EPSP's that pass the location of this synapse on their way to the trigger zone of v. However in contrast to those IPSP's that occur linearly in the formula for Pv(t) , the activation of such synapse for shunting inhibition has no impact on those EPSP's that travel to the trigger ZOne of v through another part of its dendritic tree. We model shunting inhibition in our framework as follows. We write r for the subset of all neurons 'Y in V that can "veto" other synapses (u, v) via shunting inhibition (we assume that the neurons in r have no other role apart from that). We allow in our formal model that certain 'Y in r are assigned as label to certain synapses (u, v) that have an excitatory response function cu,v. If'Y is a label of (u, v), then this models the situation that 'Y can intercept EPSP's from u on their way to the trigger zone of v via shunting inhibition. We then define Pv(t) = L (L wtt,tJ . Ett,v(t - s) . II s...,(t)) , u E V : (u, v) E EsE Ftt : s < t 'Y is label of (u, v) where we assume that S...,(t) E [0,1] is arbitrarily close to 0 for a short time interval after neuron 'Y has fired, and else equal to 1. The firing mechanism for neurons 'Y E r is defined like for all other neurons. Theorem 3: One can simulate any threshold circuit T by a sufficiently large network N(T) of noisy spiking neurons with shunting inhibition (with arbitrarily high probability of correctness). The computation time of N(T) does not depend on the number of gates in each layer, and is proportional to the number of layers in the threshold circuit T. Idea of the proof of Theorem 3: It is already impossible to simulate in a straightforward manner an AND-gate with weak coding of bits. The same difficulties arise in an even more drastic way if one wants to simulate a threshold gate with large fan-in. The left part of Figure 1 indicates that with the help of shunting inhibition one can transform via an intermediate pool of neurons Bl the bit that is weakly encoded by 216 W.MAASS Al into a contribution to Pv(t) for neurons v E C that is throughout a time interval J arbitrarily close to 0 if Al encodes a "0", and arbitrarily close to some constant P* > 0 if Al encodes a "I" (we will call this a "strong coding" of a bit). Obviously it is rather easy to realize a threshold gate if one can make use of such strong coding of bits. 8 r--------------------I I , , ® ~ 11 : E E , , , , : , E ) IAII I )IB11 SI ,- , lE :) ) '------+ C I )~-4[!] I E ) : -----+ , : H' ~ : : I ----------------------~ Figure 1: Realization of a threshold gate G via shunting inhibition (SI). The task of the module in Figure 1 is to simulate with noisy spiking neurons a n given boolean threshold gate G that outputs 1 if L: Q:iXi ~ e, and 0 else. For i=I simplicity Figure 1 shows only the pool Al whose firing activity encodes (in weak coding) the first input bit Xl. The other input bits are represented (in weak coding) simultaneously in pools A:l> ... , An parallel to AI. If Xl = 0, then the firing activity in pool Al is low, hence the shunting inhibition from pool Bl intercepts those EPSP's that are sent from BN+ to each neuron v in pool C. More precisely, we assume that each pool Bi associated with a different input bit Xi carries out shunting inhibition on a different subtree of the dendritic tree of such neurOn v (where each such subtree receives EPSP's from BN+). If Xl = 1, the higher firing activity in pool Al inhibits the neurons in BI for some time period. Hence during the relevant time interval BN+ contributes an almost constant positive summand to the potential Pv(t) of neurons v in C. By choosing wt and w; appropriately, one can achieve that during this time interval the potential Pv(t) of neurons v in 11 C is arbitrarily much positive if L: Q:iXi ~ e, and arbitrarily much negative if i=1 n L: Q:iXi < e. Hence the activity level of C encodes the output bit of the threshold i=l gate G (in weak coding). The purpose of the subsequent pools D and F is to synchronize (with the help of "double-negation") the output of this module via a pacemaker or synfire chain PM. In this way one can achieve that all input "bits" to another module that simulates a threshold gate On the next layer of circuit T arrive simultaneously. • On the Computational Power of Noisy Spiking Neurons 217 4 ConcI usion Our constructions throw new light on various experimental data, and on our attempts to understand neural computation and coding: a) If One would record all firing times of a few arbitrarily chosen neurons in our networks during many repetitions of the same computation, one is likely to see that each run yields quite different seemingly random firing sequences, where however a few firing patterns will occur more frequently than could be explained by mere chance. This is consistent with the experimental results reported in (Abeles, 1991), and one should also note that the synfire chains of (Abeles, 1991) have many features in common with the here constructed networks. b) If one plugs in biologically realistic values (see (Shepherd, 1990), (Churchland, Sejnowski, 1992)) for the length of transmission delays (around 5 msec) and the duration of EPSP's and IPSP's (around 15 msec for fast PSP's), then the computation time of our modules for NOR- and threshold gates comes out to be not more than 25 msec. Hence in principle a multi-layer perceptron with up to 4 layers can be simulated within 100 msec. c) Our constructions provide new hypotheses about the computational roles of regular and shunting inh'ibition, that go far beyond their usually assumed roles. d) We provide new hypotheses regarding the computational role of randomly firing neurons, and of EPSP's and IPSP's that arrive through synapses at distal parts of biological neurons (see the use of BN+ and BN- in our constructions). References: M. Abeles. (1991) Corticonics: Neural Circuits of the Cerebral Cortex. Cambridge University Press. P. S. Churchland, T. J. Sejnowski. (1992) The Computational Brain. MIT-Press. W. Gerstner, J. L. van Hemmen. (1994) How to describe neuronal activity: spikes, rates, or assemblies? Advances in Neural Information Processing Systems, vol. 6, Morgan Kaufmann: 463-470. W. Maass. (1995a) On the computational complexity of networks of spiking neuronS (extended abstract). Advances in Neural Information Processing Systems, vol. 7 (Proceedings of NIPS '94), MIT-Press, 183-190. W. Maass. (1995b) An efficient implementation of sigmoidal neural nets in temporal coding with noisy spiking neurons. IGI-Report 422 der Technischen Universitiit Graz, submitted for publication. W. Maass. (1996) Lower bounds for the computational power of networks of spiking neurons. N eu.ral Computation 8: 1, to appear. G. M. Shepherd. (1990) The Synaptic Organization of the Brain. Oxford University Press. J. van Leeuwen, ed. (1990) Handbook of Theoretical Computer Science, vol. A: Algorithms and Complexity. MIT-Press.
|
1995
|
8
|
1,128
|
A Smoothing Regularizer for Recurrent Neural Networks Lizhong Wu and John Moody Oregon Graduate Institute, Computer Science Dept., Portland, OR 97291-1000 Abstract We derive a smoothing regularizer for recurrent network models by requiring robustness in prediction performance to perturbations of the training data. The regularizer can be viewed as a generalization of the first order Tikhonov stabilizer to dynamic models. The closed-form expression of the regularizer covers both time-lagged and simultaneous recurrent nets, with feedforward nets and onelayer linear nets as special cases. We have successfully tested this regularizer in a number of case studies and found that it performs better than standard quadratic weight decay. 1 Introd uction One technique for preventing a neural network from overfitting noisy data is to add a regularizer to the error function being minimized. Regularizers typically smooth the fit to noisy data. Well-established techniques include ridge regression, see (Hoerl & Kennard 1970), and more generally spline smoothing functions or Tikhonov stabilizers that penalize the mth-order squared derivatives of the function being fit, as in (Tikhonov & Arsenin 1977), (Eubank 1988), (Hastie & Tibshirani 1990) and (Wahba 1990). Thes( -ilethods have recently been extended to networks of radial basis functions (Girosi, Jones & Poggio 1995), and several heuristic approaches have been developed for sigmoidal neural networks, for example, quadratic weight decay (Plaut, Nowlan & Hinton 1986), weight elimination (Scalettar & Zee 1988),(Chauvin 1990),(Weigend, Rumelhart & Huberman 1990) and soft weight sharing (Nowlan & Hinton 1992).1 All previous studies on regularization have concentrated on feedforward neural networks. To our knowledge, recurrent learning with regularization has not been reported before. ITwo additional papers related to ours, but dealing only with feed forward networks, came to our attention or were written after our work was completed. These are (Bishop 1995) and (Leen 1995). Also, Moody & Rognvaldsson (1995) have recently proposed several new classes of smoothing regularizers for feed forward nets. A Smoothing Regularizer for Recurrent Neural Networks 459 In Section 2 of this paper, we develop a smoothing regularizer for general dynamic models which is derived by considering perturbations of the training data. We present a closed-form expression for our regularizer for two layer feedforward and recurrent neural networks, with standard weight decay being a special case. In Section 3, we evaluate our regularizer's performance on predicting the U.S. Index of Industrial Production. The advantage of our regularizer is demonstrated by comparing to standard weight decay in both feedforward and recurrent modeling. Finally, we conclude our paper in Section 4. 2 Smoothing Regularization 2.1 Prediction Error for Perturbed Data Sets Consider a training data set {P: Z(t),X(t)}, where the targets Z(t) are assumed to be generated by an unknown dynamical system F*(I(t)) and an unobserved noise process: Z(t) = F*(I(t» + E*(t) with I(t) = {X(s), s = 1,2,···, t} . (1) Here, I(t) is, the information set containing both current and past inputs X(s), and the E*(t) are independent random noise variables with zero mean and variance (F*2. Consider next a dynamic network model Z(t) = F(~, I(t)) to be trained on data set P, where ~ represents a set of network parameters, and F( ) is a network transfer function which is assumed to be nonlinear and dynamic. We assume that F( ) has good approximation capabilities, such that F(~p,I(t)) ~ F*(I(t)) for learnable parameters ~ p. Our goal is to derive a smoothing regularizer for a network trained on the actual data set P that in effect optimizes the expected network performance (prediction risk) on perturbed test data sets of form {Q : Z(t),X(t)}. The elements of Q are related to the elements of P via small random perturbations Ez(t) and Ez(t), so that Z(t) = Z(t) + Ez(t) , (2) X(t) = X(t) + Ez(t) . (3) The Ez(t) and Ez(t) have zero mean and variances (Fz2 and (Fz2 respectively. The training and test errors for the data sets P and Q are N Dp = ~ L [Z(t) - F(~p,I(t))]2 (4) t=l N DQ = ~ L[Z(t) - F(~p,i(t)W , t=l (5) where ~ p denotes the network parameters obtained by training on data set P, and l(t) = {X(s),s = 1,2,··· ,t} is the perturbed information set of Q. With this notation, our goal is to minimize the expected value of DQ, while training on D p. Consider the prediction error for the perturbed data point at time t: d(t) = [Z(t) F(~p,i(t)W . (6) With Eqn (2), we obtain d(t) = [Z(t) + Ez(t) F(~p,I(t)) + F(~p,I(t)) F(~p,i(t)W, [Z(t) F(~p,I(t)W + [F(~p,I(t)) F(~p,l(t)W + [Ez(t)]2 +2[Z(t) F(~p,I(t))JIF(~p,I(t)) F(~p,i(t))] +2Ez(t)lZ(t) F(~p,l(t))]. (7) 460 L. WU. 1. MOODY Assuming that C:z(t) is uncorrelated with [Z(t) F(~p,i(t»] and averaging over the exemplars of data sets P and Q, Eqn(7) becomes 1 N 1 N Dp+ NL[F(~p,I(t»-F(~p,i(t)W+ NL[c:z(t)]2 t=1 t=1 DQ = 2 N + N L[Z(t) F(~p,I(t»)][F(~p,I(t» F(~p,i(t»]. t=l (8) The third term, 2:::'1 [C:z (t)]2, in Eqn(8) is independent of the weights, so it can be neglected during the learning process. The fourth term in Eqn(8) is the crosscovariance between [Z~t) F(~p,I(t»] and [F(~p,I(t» F(~p,i(t»]. Using the inequality 2ab ~ a + b2 , we can see that minimizing the first term D p and the second term ~ 2:~I[F(~p,I(t» F(~p,i(t»]2 in Eqn (8) during training will automatically decrease the effect of the cross-covariance term. Therefore, we exclude the cross-covariance term from the training criterion. The above analysis shows that the expected test error DQ can be minimized by minimizing the objective function D: 1 N 1 N D = N L[Z(t) F(~, I(t»]2 + N L[F(~p, I(t» - F(~ p,i(t»]2. (9) t=l t=l In Eqn (9), the second term is the time average of the squared disturbance IIZ(t) - Z(t)1I2 of the trained network output due to the input perturbation lIi(t) - I(t)W. Minimizing this term demands that small changes in the input variables yield correspondingly small changes in the output. This is the standard smoothness prior, nanlely that if nothing else is known about the function to be approximated, a good option is to assume a high degree of smoothness. Without knowing the correct functional form of the dynamical system F- or using such prior assumptions, the data fitting problem is ill-posed. In (Wu & Moody 1996), we have shown that the second term in Eqn (9) is a dynamic generalization of the first order Tikhonov stabilizer. 2.2 Form of the Proposed Smoothing Regularizer Consider a general, two layer, nonlinear, dynamic network with recurrent connections on the internal layer 2 as described by Yet) = f (WY(t - T) + V X(t» ,Z(t) = UY(t) (10) where X(t), Yet) and Z(t) are respectively the network input vector, the hidden output vector and the network output; ~ = {U, V, W} is the output, input and recurrent connections of the network; f( ) is the vector-valued nonlinear transfer function of the hidden units; and T is a time delay in the feedback connections of hidden layer which is pre-defined by a user and will not be changed during learning. T can be zero, a fraction, or an integer, but we are interested in the cases with a small T.3 20ur derivation can easily be extended to other network structures. 3When the time delay T exceeds some critical value, a recurrent network becomes unstable and lies in oscillatory modes. See, for example, (Marcus & Westervelt 1989). A Smoothing Regularizer for Recurrent Neural Networks 461 When T = 1, our model is a recurrent network as described by (Elman 1990) and (Rumelhart, Hinton & Williams 1986) (see Figure 17 on page 355). When T is equal to some fraction smaller than one, the network evolves ~ times within each input time interval. When T decreases and approaches zero, our model is the same as the network studied by (Pineda 1989), and earlier, widely-studied additive networks. In (Pineda 1989), T was referred to as the network relaxation time scale. (Werbos 1992) distinguished the recurrent networks with zero T and non-zero T by calling them simultaneous recurrent networks and time-lagged recurrent networks respectively. We have found that minimizing the second term of Eqn(9) can be obtained by smoothing the output response to an input perturbation at every time step. This yields, see (Wu & Moody 1996): IIZ(t)-Z(t)W~p/(~p)IIX(t)-X(t)W for t=1,2, ... ,N. (11) We call PT 2 (~ p) the output sensitivity of the trained network ~ p to an input perturbation. PT 2 ( ~ p) is determined by the network parameters only and is independent of the time variable t. We obtain our new regularizer by training directly on the expected prediction error for perturbed data sets Q. Based on the analysis leading to Eqns (9) and (11), the training criterion thus becomes 1 N D = N 2:[Z(t) F(~,I(t)W + .\p/(~) . t=l (12) The coefficient .\ in Eqn(12) is a regularization parameter that measures the degree of input perturbation lIi(t) - I(t)W. The algebraic form for PT(~) as derived in (Wu & Moody 1996) is: P (~)- ,IIUIIIIVII {1(,IIWIl-l)} T 1 _ ,IIWII exp T ' (13) for time-lagged recurrent networks (T > 0). Here, 1111 denotes the Euclidean matrix norm. The factor, depends upon the maximal value of the first derivatives of the activation functions of the hidden units and is given by: , = m~ II/(oj(t)) I , (14) t ,] where j is the index of hidden units and OJ(t) is the input to the ph unit. In general, , ~ 1. 4 To insure stability and that the effects of small input perturbations are damped out, it is required, see (Wu & Moody 1996), that ,IIWII < 1 . (15) The regularizer Eqn(13) can be deduced for the simultaneous recurrent networks in the limit THO by: p(~) = P (~) = ,IIUIIIIVII (16) 0 1 - ,IIWII . If the network is feedforward, W = 0 and T = 0, Eqns (13) and (16) become p(~) = ,11U1I11V1l . (17) Moreover, if there is no hidden layer and the inputs are directly connected to the outputs via U, the network is an ordinary linear model, and we obtain p(~) = IIUII , (18) 4For instance, f'(x} = [1- f(x})f(x} if f(x) = l+!-z. Then, "'{ = max 1 f'(x}} 1= t. 462 L. WU, J. MOODY which is standard quadratic weight decay (Plaut et al. 1986) as is used in ridge regression (Hoerl & Kennard 1970). The regularizer (Eqn(17) for feedforward networks and Eqn (13) for recurrent networks) was obtained by requiring smoothness of the network output to perturbations of data. We therefore refer to it as a smoothing regularizer. Several approaches can be applied to estimate the regularization parameter..x, as in (Eubank 1988), (Hastie & Tibshirani 1990) and (Wahba 1990). We will not discuss this subject in this paper. In the next section, we evaluate the new regularizer for the task of predicting the U.S. Index of Industrial Production. Additional empirical tests can be found in (Wu & Moody 1996). 3 Predicting the U.S. Index of Industrial Production The Index of Industrial Production (IP) is one of the key measures of economic activity. It is computed and published monthly. Our task is to predict the onemonth rate of change of the index from January 1980 to December 1989 for models trained from January 1950 to December 1979. The exogenous inputs we have used include 8 time series such as the index of leading indicators, housing starts, the money supply M2, the S&P 500 Index. These 8 series are also recorded monthly. In previous studies by (Moody, Levin & Rehfuss 1993), with the same defined training and test data sets, the normalized prediction errors of the one month rate of change were 0.81 with the neuz neural network simulator, and 0.75 with the proj neural network simulator. We have simulated feedforward and recurrent neural network models. Both models consist of two layers. There are 9 input units in the recurrent model, which receive the 8 exogenous series and the previous month IP index change. We set the time-delayed length in the recurrent connections T = 1. The feedforward model is constructed with 36 input units, which receive 4 time-delayed versions of each input series. The time-delay lengths a,re 1, 3, 6 and 12, respectively. The activation functions of hidden units in both feedforward and recurrent models are tanh functions. The number of hidden units varies from 2 to 6. Each model has one linear output unit. We have divided the data from January 1950 to December 1979 into four nonoverlapping sub-sets. One sub-set consists of 70% of the original data and each of the other three subsets consists of 10% of the original data. The larger sub-set is used as training data and the three smaller sub-sets are used as validation data. These three validation data sets are respectively used for determination of early stopped training, selecting the regularization parameter and selecting the number of hidden units. We have formed 10 random training-validation partitions. For each trainingvalidation partition, three networks with different initial weight parameters are trained. Therefore, our prediction committee is formed by 30 networks. The committee error is the average of the errors of all committee members. All networks in the committee are trained simultaneously and stopped at the same time based on the committee error of a validation set. The value of the regularization parameter and the number of hidden units are determined by minimizing the committee error on separate validation sets. Table 1 compares the out-of-sample performance of recurrent networks and feedforA Smoothing Regularizer for Recurrent Neural Networks 463 Table 1: Nonnalized prediction errors for the one-month rate of return on the U.S. Index of Industrial Production (Jan. 1980 - Dec. 1989). Each result is based on 30 networks. Model Regularizer Mean ± Std Median Max Min Committee Recurrent Smoothing 0.646±0.008 0.647 0.657 0.632 0.639 Networks Weight Decay 0.734±0.018 0.737 0.767 0.704 0.734 Feedforward Smoothing 0.700±0.023 0.707 0.729 0.654 0.693 Networks Weight Decay 0.745±0.043 0.748 0.805 0.676 0.731 ward networks trained with our smoothing regularizer to that of networks trained with standard weight decay. The results are based on 30 networks. As shown, the smoothing regularizer again outperfonns standard weight decay with 95% confidence (in t-distribution hypothesis) in both cases of recurrent networks and feedforward networks. We also list the median, maximal and minimal prediction errors over 30 predictors. The last column gives the committee results, which are based on the simple average of 30 network predictions. We see that the median, maximal and minimal values and the committee results obtained with the smoothing regularizer are all smaller than those obtained with standard weight decay, in both recurrent and feedforward network models. 4 Concluding Remarks Regularization in learning can prevent a network from overtraining. Several techniques have been developed in recent years, but all these are specialized for feedforward networks. To our best knowledge, a regularizer for a recurrent network has not been reported previously. We have developed a smoothing regularizer for recurrent neural networks that captures the dependencies of input, output, and feedback weight values on each other. The regularizer covers both simultaneous and time-lagged recurrent networks, with feedforward networks and single layer, linear networks as special cases. Our smoothing regularizer for linear networks has the same fonn as standard weight decay. The regularizer developed depends on only the network parameters, and can easily be used. A more detailed description of this work appears in (Wu & Moody 1996). References Bishop, C. (1995), 'Training with noise is equivalent to Tikhonov regularization', Neural Computation 7(1), 108-116. Chauvin, Y. (1990), Dynamic behavior of constrained back-propagation networks, in D. Touretzky, ed., 'Advances in Neural Infonnation Processing Systems 2', Morgan Kaufmann Publishers, San Francisco, CA, pp. 642-649. Elman, J. (1990), 'Finding structure in time', Cognition Science 14, 179-211. Eubank, R. L. (1988), Spline Smoothing and Nonparametric Regression, Marcel Dekker, Inc. Girosi, F., Jones, M. & Poggio, T. (1995), 'Regularization theory and neural networks architectures', Neural Computation 7, 219-269. 464 L. WU, J. MOODY Hastie, T. J. & Tibshirani, R. J. (1990), Generalized Additive Models, Vol. 43 of Monographs on Statistics and Applied Probability, Chapman and Hall. Hoerl, A. & Kennard, R. (1970), 'Ridge regression: biased estimation for nonorthogonal problems', Technometrics 12, 55-67. Leen, T. (1995), 'From data distributions to regularization in invariant learning', Neural Computation 7(5), 974-98l. Marcus, C. & Westervelt, R. (1989), Dynamics of analog neural networks with time delay, in D. Touretzky, ed., 'Advances in Neural Information Processing Systems 1', Morgan Kaufmann Publishers, San Francisco, CA. Moody, J. & Rognvaldsson, T. (1995), Smoothing regularizers for feed-forward neural networks, Oregon Graduate Institute Computer Science Dept. Technical Report, submitted for publication, 1995. Moody, J., Levin, U. & Rehfuss, S. (1993), 'Predicting the U.S. index of industrial production', In proceedings of the 1993 Parallel Applications in Statistics and Economics Conference, Zeist, The Netherlands. Special issue of Neural Network World 3(6), 791-794. Nowlan, S. & Hinton, G. (1992), 'Simplifying neural networks by soft weightsharing', Neural Computation 4(4), 473-493. Pineda, F. (1989), 'Recurrent backpropagation and the dynamical approach to adaptive neural computation', Neural Computation 1(2), 161-172. Plaut, D., Nowlan, S. & Hinton, G. (1986), Experiments on learning by back propagation, Technical Report CMU-CS-86-126, Carnegie-Mellon University. Rumelhart, D., Hinton, G. & Williams, R. (1986), Learning internal representations by error propagation, in D. Rumelhart & J. McClelland, eds, 'Parallel Distributed Processing: Exploration in the microstructure of cognition', MIT Press, Cambridge, MA, chapter 8, pp. 319-362. Scalettar, R. & Zee, A. (1988), Emergence of grandmother memory in feed forward networks: learning with noise and forgetfulness, in D. Waltz & J. Feldman, eds, 'Connectionist Models and Their Implications: Readings from Cognitive Science', Ablex Pub. Corp. Tikhonov, A. N. & Arsenin, V. 1. (1977), Solutions of Ill-posed Problems, Winston; New York: distributed solely by Halsted Press. Scripta series in mathematics. Translation editor, Fritz John. Wahba, G. (1990), Spline models for observational data, CBMS-NSF Regional Conference Series in Applied Mathematics. Weigend, A., Rumelhart, D. & Huberman, B. (1990), Back-propagation, weightelimination and time series prediction, in T. Sejnowski, G. Hinton & D. Touretzky, eds, 'Proceedings of the connectionist models summer school', Morgan Kaufmann Publishers, San Mateo, CA, pp. 105-116. Werbos, P. (1992), Neurocontrol and supervised learning: An overview and evaluation, in D. White & D. Sofge, eds, 'Handbook of Intelligent Control', Van Nostrand Reinhold, New York. Wu, L. & Moody, J. (1996), 'A smoothing regularizer for feedforward and recurrent neural networks', Neural Computation 8(3), 463-491.
|
1995
|
80
|
1,129
|
REMAP: Recursive Estimation and Maximization of A Posteriori Probabilities - Application to Transition-Based Connectionist Speech Recognition Yochai Konig, Herve Bourlard~ and Nelson Morgan {konig, bourlard,morgan }@icsi.berkeley.edu International Computer Science Institute 1947 Center Street Berkeley, CA 94704, USA. Abstract In this paper, we introduce REMAP, an approach for the training and estimation of posterior probabilities using a recursive algorithm that is reminiscent of the EM-based Forward-Backward (Liporace 1982) algorithm for the estimation of sequence likelihoods. Although very general, the method is developed in the context of a statistical model for transition-based speech recognition using Artificial Neural Networks (ANN) to generate probabilities for Hidden Markov Models (HMMs). In the new approach, we use local conditional posterior probabilities of transitions to estimate global posterior probabilities of word sequences. Although we still use ANNs to estimate posterior probabilities, the network is trained with targets that are themselves estimates of local posterior probabilities. An initial experimental result shows a significant decrease in error-rate in comparison to a baseline system. 1 INTRODUCTION The ultimate goal in speech recognition is to determine the sequence of words that has been uttered. Classical pattern recognition theory shows that the best possible system (in the sense of minimum probability of error) is the one that chooses the word sequence with the maximum a posteriori probability (conditioned on the * Also affiliated with with Faculte Poly technique de Mons, Mons, Belgium REMAP: Recursive Estimation and Maximization of A Posteriori Probabilities 389 evidence). If word sequence i is represented by the statistical model M i , and the evidence (which, for the application reported here, is acoustical) is represented by a sequence X = {Xl, ... , X n , ... , X N }, then we wish to choose the sequence that corresponds to the largest P(MiIX). In (Bourlard & Morgan 1994), summarizing earlier work (such as (Bourlard & Wellekens 1989)), we showed that it was possible to compute the global a posteriori probability P(MIX) of a discriminant form of Hidden Markov Model (Discriminant HMM), M, given a sequence of acoustic vectors X. In Discriminant HMMs, the global a posteriori probability P(MIX) is computed as follows: if r represents all legal paths (state sequences ql, q2, ... , qN) in Mi, N being the length of the sequence, then P(Mi IX) = L P(Mi, ql, q2, ... , qNIX) r in which ~n represents the specific state hypothesized at time n, from the set Q = {ql, ... , q , qk, ... , qK} of all possible HMM states making up all possible models Mi. We can further decompose this into: P(Mi, ql, q2,···, qNIX) = P(ql, q2,···, qNIX)P(Milql, q2,···, qN, X) Under the assumptions stated in (Bourlard & Morgan 1994) we can compute N P(ql, q2,···, qNIX) = II p(qnlqn-l, xn) n=l The Discriminant HMM is thus described in terms of conditional transition probabilities p(q~lq~-l' xn), in which q~ stands for the specific state ql of Q hypothesized at time n and can be schematically represented as in Figure 1. P(IkIIIkI, x) p(/aell/ael, x) P(ltIlltI, x) P(/aelllkl, x) P(ltll/ael, x) Figure 1: An example Discriminant HMM for the word "cat". The variable X refers to a specific acoustic observation Xn at time n. Finally, given a state sequence we assume the following approximation: P(Milql, q2,···, qN, X) :::::::: P(Milql, q2,···, qN) We can estimate the right side of this last equation from a phonological model (in the case that a given state sequence can belong to two different models). All the required (local) conditional transition probabilities p(q~lq~-l> xn) can be estimated by the Multi-Layer Perceptron (MLP) shown in Figure 2. Recent work at lesl has provided us with further insight into the discriminant HMM, particularly in light of recent work on transition-based models (Konig & Morgan 1994j Morgan et al. 1994). This new perspective has motivated us to further develop the original Discriminant HMM theory. The new approach uses posterior probabilities at both local and global levels and is more discriminant in nature. In this paper, we introduce the Recursive Estimation-Maximization of A posteriori 390 Y. KONIG, H. BOURLARD, N. MORGAN P(CurrenCstlte I Acoustics, Prevlous_stlte) t ···· .. t 0.1 •• 0 Previous Stlte t t t t Acoustics Figure 2: An MLP that estimates local conditional transition probabilities. Probabilities (REMAP) training algorithm for hybrid HMM/MLP systems. The proposed algorithm models a probability distribution over all possible transitions (from all possible states and for all possible time frames n) rather than picking a single time point as a transition target. Furthermore, the algorithm incrementally increases the posterior probability of the correct model, while reducing the posterior probabilities of all other models. Thus, it brings the overall system closer to the optimal Bayes classifier. A wide range of discriminant approaches to speech recognition have been studied by researchers (Katagiri et al. 1991; Bengio et al. 1992; Bourlard et al. 1994). A significant difficulty that has remained in applying these approaches to continuous speech recognition has been the requirement to run computationally intensive algorithms on all of the rival sentences. Since this is not generally feasible, compromises must always be made in practice. For instance, estimates for all rival sentences can be derived from a list of the "N-best" utterance hypotheses, or by using a fully connected word model composed of all phonemes. 2 REMAP TRAINING OF THE DISCRIMINANT HMM 2.1 MOTIVATIONS The discriminant HMM/MLP theory as described above uses transition-based probabilities as the key building block for acoustic recognition. However, it is well known that estimating transitions accurately is a difficult problem (Glass 1988). Due to the inertia of the articulators, the boundaries between phones are blurred and overlapped in continuous speech. In our previous hybrid HMM/MLP system, targets were typically obtained by using a standard forced Viterbi alignment (segmentation). For a transition-based system as defined above, this procedure would thus yield rigid transition targets, which is not realistic. Another problem related to the Viterbi-based training of the MLP presented in Figure 2 and used in Discriminant HMMs, is the lack of coverage of the input space during training. Indeed, during training (based on hard transitions), the MLP only processes inputs consisting of "correct" pairs of acoustic vectors and correct previous state, while in recognition the net should generalize to all possible combinations of REMAP: Recursive Estimation and Maximization of A Posteriori Probabilities 391 acoustic vectors and previous states, since all possible models and transitions will be hypothesized for each acoustic input. For example, some hypothesized inputs may correspond to an impossible condition that has thus never been observed, such as the acoustics of the temporal center of a vowel in combination with a previous state that corresponds to a plosive. It is unfortunately possible that the interpolative capabilities of the network may not be sufficient to give these "impossible" pairs a sufficiently low probability during recognition. One possible solution to these problems is to use a full MAP algorithm to find transition probabilities at each frame for all possible transitions by a forward-backwardlike algorithm (Liporace 1982), taking all possible paths into account. 2.2 PROBLEM FORMULATION As described above, global maximum a posteriori training of HMMs should find the optimal parameter set e maximizing J II P(Mj IXj, e) (1) j=1 in which Mj represents the Markov model associated with each training utterance Xj, with j = 1, ... , J. Although in principle we could use a generalized back-propagation-like gradient procedure in e to maximize (1) (Bengio et al. 1992), an EM-like algorithm should have better convergence properties, and could preserve the statistical interpretation of the ANN outputs. In this case, training of the discriminant HMM by a global MAP criterion requires a solution to the following problem: given a trained MLP at iteration t providing a parameter set et and, consequently, estimates of P(q~lxn' q~-I' et ), how can we determine new MLP targets that: 1. will be smooth estimates of conditional transition probabilities q~-1 -+ q~, Vk,f E [1, K] and "In E [1, N], 2. when training the MLP for iteration t+ 1, will lead to new estimates of et+l and P(q~lxn' q~-I' et+1) that are guaranteed to incrementally increase the global posterior probability P(MiIX, e)? In (Bourlard et al. 1994), we prove that a re-estimate of MLP targets that guarantee convergence to a local maximum of (1) is given by1: (2) where we have estimated the left-hand side using a mapping from the previous state and the local acoustic data to the current state, thus making the estimator realizable by an MLP with a local acoustic window. 2 Thus, we will want to estimate 1 In most of the following, we consider only one particular training sequence X associated with one particular model M. It is, however, easy to see that all of our conclusions remain valid for the case of several training sequences Xj, j = 1, ... , J. A simple way to look at the problem is to consider all training sequences as a single training sequence obtained by concatenating all the X,'s with boundary conditions at every possible beginning and ending point. 2Note that, as done in our previous hybrid HMM/MLP systems, all conditional on Xn can be replaced by X;::!: = {x n - c , ., •. , X n , .•• , Xn+d} to take some acoustic context into account. 392 Y. KONIG, H. BOURLARD, N. MORGAN the transition probability conditioned on the local data (as MLP targets) by using the transition probability conditioned on all of the data. In (Bourlard et al. 1994), we further prove that alternating MLP target estimation (the "estimation" step) and MLP training (the" maximization" step) is guaranteed to incrementally increase (1) over t.3 The remaining problem is to find an efficient algorithm to express P(q~IX, q~-l' M) in terms of P(q~lxn, q~-l) so that the next iteration targets can be found. We have developed several approaches to this estimation, some of which are described in (Bourlard et al. 1994). Currently, we are implementing this with an efficient recursion that estimates the sum of all possible paths in a model, for every possible transition at each possible time. From these values we can compute the desired targets (2) for network training by P( t IX M k ) = P(M, q~, ~~_lIX) qn , , qn-l ~ . P(M J k IX) DJ ,qn, qn-l (3) 2.3 REMAP TRAINING ALGORITHM The general scheme of the REMAP training of hybrid HMM/MLP systems can be summarized as follow: 1. Start from some initial net providing P(q~lxn' q~-l' et ), t = 0, V possible (k,£)-pairs4. 2. Compute MLP targets P(q~IXj,q~_l,et,Mj) according to (3), V training sentences Xj associated with HMM Mj, V possible (k, £) state transition pairs in Mj and V X n, n = 1, ... , N in Xj (see next point). 3. For every Xn in the training database, train the MLP to minimize the relative entropy between the outputs and targets. See (Bourlard et ai, 1994) for more details. This provides us with a new set of parameters et , for t = t + 1. 4. Iterate from 2 until convergence. This procedure is thus composed of two steps: an Estimation (E) step, corresponding to step 2 above, and a Maximization (M) step, corresponding to step 3 above. In this regards, it is reminiscent of the Estimation-Maximization (EM) algorithm as discussed in (Dempster et al. 1977). However, in the standard EM algorithm, the M step involves the actual maximization of the likelihood function. In a related approach, usually referred to as Generalized EM (GEM) algorithm, the M step does not actually maximize the likelihood but simply increases it (by using, e.g., a gradient procedure). Similarly, REMAP increases the global posterior function during the M step (in the direction of targets that actually maximize that global function), rather than actually maximizing it. Recently, a similar approach was suggested for mapping input sequences to output sequences (Bengio & Frasconi 1995). 3Note here that one "iteration" does not stand for one iteration of the MLP training but for one estimation-maximization iteration for which a complete MLP training will be required. 4This can be done, for instance, by training up such a net from a hand-labeled database like TIMIT or from some initial forward-backward estimator of equivalent local probabilities (usually referred to as "gamma" probabilities in the Baum-Welch procedure). REMAP: Recursive Estimation and Maximization of A Posteriori Probabilities 393 System Error Rate DHMM, pre-REMAP 14.9% 1 REMAP iteration 13.6% 2 REMAP iterations 13.2% Table 1: Training and testing on continuous numbers, no syntax, no durational models. 3 EXPERIMENTS AND RESULTS For testing our theory we chose the Numbers'93 corpus. It is a continuous speech database collected by CSLU at the Oregon Graduate Institute. It consists of numbers spoken naturally over telephone lines on the public-switched network (Cole et al. 1994). The Numbers'93 database consists of 2167 speech files of spoken numbers produced by 1132 callers. We used 877 of these utterances for training and 657 for cross-validation and testing (200 for cross-validation) saving the remaining utterances for final testing purposes. There are 36 words in the vocabulary, namely zero, oh, 1, 2, 3, ... ,20, 30, 40, 50, ... ,100, 1000, a, and, dash, hyphen, and double. All our nets have 214 inputs: 153 inputs for the acoustic features, and 61 to represent the previous state (one unit for every possible previous state, one state per phoneme in our case). The acoustic features are combined from 9 frames with 17 features each (RASTA-PLP8 + delta features + delta log gain) computed with an analysis window of 25 ms computed every 12.5 ms (overlapping windows) and with a sampling rate of 8 Khz. The nets have 200 hidden units and 61 outputs. Our results are summarized in Table 1. The row entitled "DHMM, pre-REMAP" corresponds to a Discriminant HMM using the same training approach, with hard targets determined by the first system, and additional inputs to represent the previous state The improvement in the recognition rate as a result of REMAP iterations is significant at p < 0.05. However all the experiments were done using acoustic information alone. Using our (baseline) hybrid system under equal conditions, i.e., no duration information and no language information, we get 31.6% word error; adding the duration information back we get 12.4% word error. We are currently experimenting with enforcing minimum duration constraints in our framework. 4 CONCLUSIONS In summary: • We have a method for MAP training and estimation of sequences. • This can be used in a new form of hybrid HMM/MLP. Note that recurrent nets or TDNNs could also be used. As with standard HMM/MLP hybrids, the network is used to estimate local posterior probabilities (though in this case they are conditional transition probabilities, that is, state probabilities conditioned on the acoustic data and the previous state). However, in the case of REMAP these nets are trained with probabilistic targets that are themselves estimates of local posterior probabilities. • Initial experiments demonstrate a significant reduction in error rate for this process. 394 Y. KONIG, H. BOURLARD, N. MORGAN Acknowledgments We would like to thank Kristine Ma and Su-Lin Wu for their help with the Numbers'93 database. We also thank OGI, in particular to Ron Cole, for providing the database. We gratefully acknowledge the support of the Office of Naval Research, URI No. N00014-92-J-1617 (via UCB), the European Commission via ESPRIT project 20077 (SPRACH), and ICSI and FPMs in general for supporting this work. References BENGIO, Y., & P. FRASCONI. 1995. An input output HMM architecture. In Advances in Neural Information Processing Systems, ed. by G. Tesauro, D. Touretzky, & T. Leen, volume 7. Cambridge: MIT press. --, R. DE MORI, G. FLAMMIA, & R. KOMPE. 1992. Global optimization of a neural network-hidden Markov model hybrid. IEEE trans. on Neural Networks 3.252-258. BOURLARD, H., Y. KONIG, & N. MORGAN. 1994. REMAP: Recursive estimation and maximization of a posteriori probabilities, application to transition-based connectionist speech recognition. Technical Report TR-94-064, International Computer Science Institute, Berkeley, CA. --, & N. MORGAN. 1994. Connectionist Speech Recognition - A Hybrid Approach. Kluwer Academic Publishers. --, & C. J. WELLEKENS. 1989. Links between Markov models and multilayer perceptrons. In Advances in Neural Information Processing Systems 1, ed. by D.J. Touretzky, 502-510, San Mateo. Morgan Kaufmann. COLE, R.A., M. FANTY, & T. LANDER. 1994. Telephone speech corpus development at CSL U. In Proceedings Int 'I Conference on Spoken Language Processing, Yokohama, Japan. DEMPSTER, A. P., N. M. LAIRD, & D. B. RUBIN. 1977. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, Series B 34.1-38. GLASS, J. R., 1988. Finding Acoustic Regularities in Speech Applications to Phonetic Recognition. M.LT dissertation. KATAGIRI, S., C.H. LEE, & JUANG B.H. 1991. New discriminative training algorithms based on the generalized probabilistic decent method. In Proc. of the IEEE Workshop on Neural Netwroks for Signal Processing, ed. by RH. Juang, S.Y. Kung, & C.A. Kamm, 299-308. KONIG, Y., & N. MORGAN. 1994. Modeling dynamics in connectionist speech recognition - the time index model. In Proceedings Int'l Conference on Spoken Language Processing, 1523-1526, Yokohama, Japan. LIPORACE, L. A. 1982. Maximum likelihood estimation for multivariate observations of markov sources. IEEE Trans. on Information Theory IT-28.729-734. MORGAN, N., H. BOURLARD, S. GREENBERG, & H. HERMANSKY. 1994. Stochastic perceptual auditory-event-based models for speech recognition. In Proceedings Int'l Conference on Spoken Language Processing, 1943-1946, Yokohama, Japan.
|
1995
|
81
|
1,130
|
Harmony Networks Do Not Work Rene Gourley School of Computing Science Simon Fraser University Burnaby, B.C., V5A 1S6, Canada gourley@mprgate.mpr.ca Abstract Harmony networks have been proposed as a means by which connectionist models can perform symbolic computation. Indeed, proponents claim that a harmony network can be built that constructs parse trees for strings in a context free language. This paper shows that harmony networks do not work in the following sense: they construct many outputs that are not valid parse trees. In order to show that the notion of systematicity is compatible with connectionism, Paul Smolensky, Geraldine Legendre and Yoshiro Miyata (Smolensky, Legendre, and Miyata 1992; Smolen sky 1993; Smolen sky, Legendre, and Miyata 1994) proposed a mechanism, "Harmony Theory," by which connectionist models purportedly perform structure sensitive operations without implementing classical algorithms. Harmony theory describes a "harmony network" which, in the course of reaching a stable equilibrium, apparently computes parse trees that are valid according to the rules of a particular context-free grammar. Harmony networks consist of four major components which will be explained in detail in Section 1. The four components are, Tensor Representation: A means to interpret the activation vector of a connectionist system as a parse tree for a string in a context-free language. Harmony: A function that maps all possible parse trees to the non-positive integers so that a parse tree is valid if and only if its harmony is zero. Energy: A function that maps the set of activation vectors to the real numbers and which is minimized by certain connectionist networks!. Recursive Construction: A system for determining the weight matrix of a connectionist network so that if its activation vector is interpreted as a parse 1 Smolensky, Legendre and Miyata use the term "harmony" to refer to both energy and harmony. To distinguish between them, we will use the term that is often used to describe the Lyapunov function of dynamic systems, "energy" (see for example Golden 1986). 32 R. GOURLEY tree, then the network's energy is the negation of the harmony of that parse tree. Smolen sky et al. contend that, in the process of minimizing their energy values, harmony networks implicitly maximize the harmony of the parse tree represented by their activation vector. Thus, if the harmony network reaches a stable equilibrium where the energy is equal to zero, the parse tree that is represented by the activation vector must be a valid parse tree: When the lower-level description of the activation-spreading process satisfies certain mathematical properties, this process can be analyzed on a higher level as the construction of that structure including the given input structure which maximizes Harmony. (Smolensky 1993, p848, emphasis is original) Unfortunately, harmony networks do not work they do not always construct maximum-harmony parse trees. The problem is that the energy function is defined on the values of the activation vector. By contrast, the harmony function is defined on possible parse trees. Section 2 of this paper shows that these two domains are not equal, that is, there are some activation vectors that do not represent any parse tree. The recursive construction merely guarantees that the energy function passes through zero at the appropriate points; its minima are unrestricted. So, while it may be the case that the energy and harmony functions are negations of one another, it is not always the case that a local minimum of one is a local maximum of the other. More succinctly, the harmony network will find minima that are not even trees, let alone valid parse trees. The reason why harmony networks do not work is straightforward. Section 3 shows that the weight matrix must have only negative eigenvalues, for otherwise the network constructs structures which are not valid trees. Section 4 shows that if the weight matrix has only negative eigenvalues, then the energy function admits only a single zero the origin. Furthermore, we show that the origin cannot be interpreted as a valid parse tree. Thus, the stable points of a harmony network are not valid parse trees. 1 HARMONY NETWORKS 1.1 TENSOR REPRESENTATION Harmony theory makes use of tensor products (Smolensky 1990; Smolensky, Legendre, and Miyata 1992; Legendre, Miyata, and Smolensky 1991) to convolve symbols with their roles. The resulting products are then added to represent a labelled tree using the harmony network's activation vector. The particular tensor product used is very simple: (aI, a2,· · ·, an) <8> (bl , b2,.·., bm ) = (albl , alb2, ... , a}bm , a2bl, a2b2, ... , a2bm, .. . , anbm ) If two tensors of differing dimensions are to be added, then they are essentially concatenated. Binary trees are represented with this tensor product using the following recursive rules: 1. The tensor representation of a tree containing no vertices is O. Harmony Networks Do Not Work 33 Table 1: Rules for determining harmony and the weight matrix. Let G = (V, E, P, S) be a context-free grammar of the type suggested in section 1.2. The rules for determining the harmony of a tree labelled with V and E are shown in the second column. The rules for determining the system of equations for recursive construction are shown in the third column. (Smolensky, Legendre, and Miyata 1992; Smolensky 1993) Grammar Harmony Rule Energy Equation Element S For every node labelled Include (S+00r,)Wroot(S+00rr) = 2 S add -1 to H(T). in the system of equations xEE For every node labelled Include (x +60r,)Wroot (x +60r,) = 2 x add -1 to H(T). in the system of equations For every node labelled x add -2 or -3 to H(T) Include (x+60r,)Wroot(x+00r,) = 4 x E V\ depending on whether or 6 in the system of equations, depend{S} or not x appears on ing on whether or not x appears on the the left of a producleft of a production with two symbols tion with two symbols on the right. on the right. For every edge where Include in the system of equations, x yz x is the parent and y (x + 60 r,)Wroot (6 + y 0 r,) = -2 or x is the left child add 2. (0 + y 0 r,)Wroot(x + 60 r,) = -2 yE P Similarly, add 2 every (x + 60 r,)Wroot(O + z 0 r,) = -2 time z is the right child of x. (6 + z 0 r,)Wroot(x + 6® r,) = -2 2. If A is the root of a tree, and TL, TR are the tensor product representations of its left subtree and right subtree respectively, then A + TL 0 r, + TR 0 rr is the tensor representation of the whole tree. The vectors, r" and rr are called "role vectors" and indicate the roles of left child and right child. 1.2 HARMONY Harmony (Legendre, Miyata, and Smolensky 1990; Smolensky, Legendre, and Miyata 1992) describes a way to determine the well-formedness of a potential parse tree with respect to a particular context free grammar. Without loss of generality, we can assume that the right-hand side of each production has at most two symbols, and if a production has two symbols on the right, then it is the only production for the variable on its left side. For a given binary tree, T, we compute the harmony of T, H(T) by first adding the negative contributions of all the nodes according to their labels, and then adding the contributions of the edges (see first two columns of table 1). 34 R.GOURLEY 1.3 ENERGY Under certain conditions, some connectionist models are known to admit the following energy or Lyapunov function (see Legendre, Miyata, and Smolensky 1991): 1 E(a) = --atWa 2 Here, W is the weight matrix of the connectionist network, and a is its activation vector. Every non-equilibrium change in the activation vector results in a strict decrease in the network's energy. In effect, the connectionist network serves to minimize its energy as it moves towards equilibrium. 1.4 RECURSIVE CONSTRUCTION Smolensky, Legendre, and Miyata (1992) proposed that the recursive structure of their tensor representations together with the local nature of the harmony calculation could be used to construct the weight matrix for a network whose energy function is the negation of the harmony of the tree represented by the activation vector. First construct a matrix W root which satisfies a system of equations. The system of equations is found by including equations for every symbol and production in the grammar, as shown in column three of table 1. Gourley (1995) shows that if W is constructed from copies of W root according to a particular formula, and if aT is a tensor representation for a tree, T, then E(aT) = -H(T). 2 SOME ACTIVATIONS ARE NOT TREES As noted above, the reason why harmony networks do not work is that they seek minima in their state space which may not coincide with parse tree representations. One way to amelioarate this would be to make every possible activation vector represent some parse tree. If every activation vector represents some parse tree, then the rules that determine the weight matrix will ensure that the energy minima agree with the valid parse trees. Unfortunately, in that case, the system of equations used to determine W root has no solution. If every activation vector is to represent some parse tree, and the symbols of the grammar are two dimensional, then there are symbols represented by each vector, (Xl, xt), (Xl, X2), (X2' xt), and (X2' X2), where Xl 1= X2 . These symbols must satisfy the equations given in table 1 , and so, Xi{Wrootll + Wroot12 + Wroot~l + Wroot~~) XiWrootll + XIX2 W root12 + XIX2 W root:n + x~Wroot:n X~Wrootll + XIX2Wrootl~ + XIX2 W root:n + xiWroot~2 x~(Wrootll + Wroot12 + Wroot~l + Wrootn) Because hi E {2, 4, 6}, there must be a pair hi, hj which are equal. In that case, it can be shown using Gaussian elimination that there is no solution for Wrootll , Wrootl~' Wroot~l , Wroot~~. Similarly, if the symbols are represented by vectors of dimension three or greater, the same contradiction occurs. Thus there are some activation vectors that do not represent any tree valid or invalid. The question now becomes one of determining whether all of the harmony network's stable equilibria are valid parse trees. Harmony Networks Do Not Work 35 a b Figure 1: Energy functions of two-dimensional harmony networks. In each case, the points i and f respectively represent an initial and a final state of the network. In a, one eigenvector is positive and the other is negative; the hashed plane represents the plane E = 0 which intersects the energy function and the vertical axis at the origin. In b, one eigenvalue is negative while the other is zero; The heavy line represents the intersection of the surface with the plane E = 0 and it intersects the vertical axis at the origin. 3 NON-NEGATIVE EIGENVECTORS YIELD NON-TREES If any of the eigenvalues of the weight matrix, W, is positive, then it is easy to show that the harmony network will seek a stable equilibrium that does not represent a parse tree at all. Let A > 0 be a positive eigenvalue of W, and let e be an eigenvector, corresponding to A, that falls within the state space. Then, 1 1 E(e) = --etWe = --Aete < O. 2 2 Because the energy drops below zero, the harmony network would have to undergo an energy increase in order to find a zero-energy stable equilibrium. This cannot happen, and so, the network reaches an equilibrium with energy strictly less than zero. Figure la illustrates the energy function of a harmony network where one eigenvalue is positive. Because harmony is the negation of energy, in this figure all the valid parse trees rest on the hashed plane, and all the invalid parse trees are above it. As we can see, the harmony network with positive eigenvalues will certainly find stable equilibria which are not valid parse tree representations. Now, suppose W, the weight matrix, has a zero eigenvalue. If e is an eigenvector corresponding to that eigenvalue, then for every real a, aWe = O. Consequently, one of the following must be true: 1. ae is not a stable equilibrium. In that case, the energy function must drop below zero, yielding a sub-zero stable equilibrium a stable equilibrium that does not represent any tree. 2. ae is a stable equilibrium. Then for every a, ae must be a valid tree representation. Such a situation is represented in fig36 R. GOURLEY Figure 2: The energy function of a two-dimensional harmony network where both eigenvalues are negative. The vertical axis pierces the surface at the origin, and the points i and f respectively represent an initial and a final state of the network. ure Ib where the set of all points ae is represented by the heavy line. This implies that there is a symbol, (al, a2, . . . , an), such that Ckl(al , a2, .. . ,an),Ck2(al,a2, . . . ,an), .. . ,an2+l(al,a2, ... , an) are also all symbols. As before, this implies that Wroot must satisfy the equation, t hi hi E «al, ... , an) + 0 ® r,) Wroot«al, ... , an) + 0 0 r,) 2" ' {2 4 6} a · " , for i = 1 ... n2 + 1. Again using Gaussian elimination, it can be shown that there is no solution to this system of equations. In either case, the harmony network admits stable equilibria that do not represent any tree. Thus, the eigenvalues must all be negative. 4 NEGATIVE EIGENVECTORS YIELD NON-TREES If all the eigenvalues of the weight matrix are negative, then the energy function has a very special shape: it is a paraboloid centered on the origin and concave in the direction of positive energy. This is easily seen by considering the first and second derivatives of E: 8E(x) __ ~ W, .. x . 8 2 E(x) - -W, . . 8x; L..j '.1' 8x;8x; '.1 Clearly, all the first derivatives are zero at the origin, and so, it is a critical point. Now the origin is a strict minimum if all the roots of the following well-known equation are positive: 0= det = det I-W - All det 1- W - All is the characteristic polynomial of -W. If A is a root then it is an eigenvalue of - W, or equivalently, it is the negative of an eigenvalue of W . Because all of W's eigenvalues are negative, the origin is a strict minimum, and indeed it is the only minimum. Such a harmony network is illustrated in Figure 2. Hannony Networks Do Not Work 37 Thus the origin is the only stable point where the energy is zero, but it cannot represent a parse tree which is valid for the grammar. If it does, then S + TL 0 r, + TR (9 rr = (0, . . . ,0) where TL, TR are appropriate left and right subtree representations, and S is the start symbol of the grammar. Because each of the subtrees is multiplied by either r, or rr, they are not the same dimension as S, and are consequently concatenated instead of added. Therefore S = O. But then, Wroot must satisfy the equation (0 + 0 (9 r,)Wroot(O + 0 (9 r,) =-2 This is impossible, and so, the origin is not a valid tree representation. 5 CONCLUSION This paper has shown that in every case, a harmony network will reach stable equilibria that are not valid parse trees. This is not unexpected. Because the energy function is a very simple function, it would be more surprising if such a connectionist system could construct complicated structures such as parse trees for a context free grammar. Acknowledgements The author thanks Dr. Robert Hadley and Dr. Arvind Gupta, both of Simon Fraser University, for their invaluable comments on a draft of this paper. References Golden, R. (1986). The 'brain-state-in-a-box' neural model is a gradient descent algorithm. Journal of Mathematical Psychology 30, 73-80. Gourley, R. (1995). Tensor represenations and harmony theory: A critical analysis. Master's thesis, Simon Fraser University, Burnaby, Canada. In preparation. Legendre, G., Y. Miyata, and P. Smolensky (1990). Harmonic grammar - a formal multi-level connectionist theory of linguistic well-formedness: Theoretical foundations. In Proceedings of the Twelfth National Conference on Cognitive Science, Cambridge, MA, pp. 385- 395. Lawrence Erlbaum. Legendre, G., Y. Miyata, and P. Smolensky (1991). Distributedrecursive structure processing. In B. Mayoh (Ed.), Proceedings of the 1991 Scandinavian Conference on Artificial Intelligence, Amsterdam, pp. 47-53. lOS Press. Smolensky, P. (1990). Tensor product variable binding and the representation of symbolic structures in connectionist systems. Artificial Intelligence 46, 159-216. Smolensky, P. (1993). Harmonic grammars for formal languages. In S. Hanson, J. Cowan, and C. Giles (Eds.), Advances in Neural Information Processing Systems 5, pp. 847-854. San Mateo: Morgan Kauffman. Smolensky, P., G. Legendre, and Y. Miyata (1992). Principles for an integrated connectionist/symbolic theory of higher cognition. Technical Report CU-CS-60092, University of Colorado Computer Science Department. Smolensky, P., G. Legendre, and Y. Miyata (1994) . Integrating connectionist and symbolic computation for the theory of language. In V. Honavar and L. Uhr (Eds.), Artificial Intelligence and Neural Networks: Steps Toward Principled Integration, pp. 509-530. Boston: Academic Press.
|
1995
|
82
|
1,131
|
Learning Sparse Perceptrons Jeffrey C. Jackson Mathematics & Computer Science Dept. Duquesne University 600 Forbes Ave Pittsburgh, PA 15282 jackson@mathcs.duq.edu Abstract Mark W. Craven Computer Sciences Dept. University of Wisconsin-Madison 1210 West Dayton St. Madison, WI 53706 craven@cs.wisc.edu We introduce a new algorithm designed to learn sparse perceptrons over input representations which include high-order features. Our algorithm, which is based on a hypothesis-boosting method, is able to PAC-learn a relatively natural class of target concepts. Moreover, the algorithm appears to work well in practice: on a set of three problem domains, the algorithm produces classifiers that utilize small numbers of features yet exhibit good generalization performance. Perhaps most importantly, our algorithm generates concept descriptions that are easy for humans to understand. 1 Introd uction Multi-layer perceptron (MLP) learning is a powerful method for tasks such as concept classification. However, in many applications, such as those that may involve scientific discovery, it is crucial to be able to explain predictions. Multi-layer perceptrons are limited in this regard, since their representations are notoriously difficult for humans to understand. We present an approach to learning understandable, yet accurate, classifiers. Specifically, our algorithm constructs sparse perceptrons, i.e., single-layer perceptrons that have relatively few non-zero weights. Our algorithm for learning sparse perceptrons is based on a new hypothesis boosting algorithm (Freund & Schapire, 1995). Although our algorithm was initially developed from a learning-theoretic point of view and retains certain theoretical guarantees (it PAC-learns the class of sparse perceptrons), it also works well in practice. Our experiments in a number of real-world domains indicate that our algorithm produces perceptrons that are relatively comprehensible, and that exhibit generalization performance comparable to that of backprop-trained MLP's (Rumelhart et al., 1986) and better than decision trees learned using C4.5 (Quinlan, 1993). Learning Sparse Perceptrons 655 We contend that sparse perceptrons, unlike MLP's, are comprehensible because they have relatively few parameters, and each parameter describes a simple (Le. linear) relationship. As evidence that sparse perceptrons are comprehensible, consider that such linear functions are commonly used to express domain knowledge in fields such as medicine (Spackman, 1988) and molecular biology (Stormo, 1987). 2 Sparse Perceptrons A perceptron is a weighted threshold over the set of input features and over higherorder features consisting of functions operating on only a limited number of the input features. Informally, a sparse perceptron is any perceptron that has relatively few non-zero weights. For our later theoretical results we will need a more precise definition of sparseness which we develop now. Consider a Boolean function I : {O, 1 } n -t { -1, + 1 }. Let Ck be the set of all conjunctions of at most k of the inputs to I. Ck includes the "conjunction" of 0 inputs, which we take as the identically 1 function. All of the functions in Ck map to {-1,+1}, and every conjunction in Ck occurs in both a positive sense (+1 represents true) and a negated sense (-1 represents true). Then the function I is a k-perceptron if there is some integer s such that I(x) = sign(L::=1 hi(x)), where for all i, hi E Ck, and sign(y) is undefined if y = 0 and is y/lyl otherwise. Note that while we have not explicitly shown any weights in our definition of a k-perceptron I, integer weights are implicitly present in that we allow a particular hi E Ck to appear more than once in the sum defining I. In fact, it is often convenient to think of a k-perceptron as a simple linear discriminant function with integer weights defined over a feature space with O(nk) features, one feature for each element of Ck • We call a given collection of s conjunctions hi E Ck a k-perceptron representation of the corresponding function I, and we call s the size of the representation. We define the size of a given k-perceptron function I as the minimal size of any k-perceptron representation of I. An s-sparse k-perceptron is a k-perceptron I such that the size of I is at most s. We denote by PI: the set of Boolean functions over {O, 1}n which can be represented as k-perceptrons, and we define Pk = Un Pi:. The subclass of s-sparse k-perceptrons is denoted by Pk,/l" We are also interested in the class P~ of k-perceptrons with real-valued weights, at most r of which are non-zero. 3 The Learning Algorithm In this section we develop our learning algorithm and prove certain performance guarantees. Our algorithm is based on a recent "hypothesis boosting" algorithm that we describe after reviewing some basic learning-theory terminology. 3.1 PAC Learning and Hypothesis Boosting Following Valiant (1984), we say that a function class :F (such as Pk for fixed k) is (strongly) PAC-learnable if there is an algorithm A and a polynomial function PI such that for any positive f and 8, any I E :F (the target junction), and any probability distribution D over the domain of I, with probability at least 1 8, algorithm A(EX(f, D), f, 8) produces a function h (the hypothesis) such that Pr[PrD[/(x) I- hex)] > f] < 8. The outermost probability is over the random choices made by the EX oracle and any random choices made by A. Here EX(f, D) denotes an oracle that, when queried, chooses a vector of input values x with probability D and returns the pair (x,/(x)) to A. The learning algorithm A must run in time PI (n, s, c 1 , 8-1 ), where n is the length of the input vector to I and s is the size of 656 J. C. JACKSON, M. W. CRAVEN AdaBoost Input: training set S of m examples of function f, weak learning algorithm WL that is (~ - 'Y)-approximate, l' Algorithm: 1. T +-- ~ In(m) 2. for all xES, w(x) +-- l/m 3. for i = 1 to T do 4. for all XES, Di(X) +-- w(x)/ L:l=l w(x). 5. invoke WL on S and distribution Di, producing weak hypothesis hi 6. €i +-- L:z.h;(z);oI:/(z) Di(X) 7. (3i +-- €i/ (1 - €i) 8. for all XES, if h(x) = f(x) then w(x) +-- w(x) . (3i 9. enddo Output: h(x) == sign (L::=l -In((3i) . hi{x)) Figure 1: The AdaBoost algorithm. f; the algorithm is charged one unit of time for each call to EX. We sometimes call the function h output by A an €-approximator (or strong approximator) to f with respect to D. If F is PAC-learnable by an algorithm A that outputs only hypotheses in class 1£ then we say that F is PAC-learnable by 1£. If F is PAClearnable for € = 1/2 - 1/'P2(n, s), where'P2 is a polynomial function, then :F is weakly PA C-learnable, and the output hypothesis h in this case is called a weak approximator. Our algorithm for finding sparse perceptrons is, as indicated earlier, based on the notion of hypothesis boosting. The specific boosting algorithm we use (Figure 1) is a version of the recent AdaBoost algorithm (Freund & Schapire, 1995). In the next section we apply AdaBoost to "boost" a weak learning algorithm for Pk,8 into a strong learner for Pk,8' AdaBoost is given a set S of m examples of a function f : {O,1}n ---+ {-1, +1} and a weak learning algorithm WL which takes € = ! - l' for a given l' b must be bounded by an inverse polynomial in nand s). Adaf300st runs for T = In(m)/(2'Y2) stages. At each stage it creates a probability distribution Di over the training set and invokes WL to find a weak hypothesis hi with respect to Di (note that an example oracle EX(j, Di) can be simulated given Di and S). At the end of the T stages a final hypothesis h is output; this is just a weighted threshold over the weak hypotheses {hi I 1 ~ i ~ T}. If the weak learner succeeds in producing a (~-'Y)-approximator at each stage then AdaBoost's final hypothesis is guaranteed to be consistent with the training set (Freund & Schapire, 1995). 3.2 PAC-Learning Sparse k-Perceptrons We now show that sparse k-perceptrons are PAC learnable by real-weighted kperceptrons having relatively few nonzero weights. Specifically, ignoring log factors, Pk,8 is learnable by P~O(82) for any constant k. We first show that, given a training set for any f E Pk,8' we can efficiently find a consistent h E p~( 8 2)' This consistency algorithm is the basis of the algorithm we later apply to empirical learning problems. We then show how to turn the consistency algorithm into a PAC learning algorithm. Our proof is implicit in somewhat more general work by Freund (1993), although he did not actually present a learning algorithm for this class or analyze Learning Sparse Perceptrons 657 the sample size needed to ensure f-approximation, as we do. Following Freund, we begin our development with the following lemma (Goldmann et al., 1992): Lemma 1 (Goldmann Hastad Razhorov) For I: {0,1}n -+ {-1,+1} and H, any set 01 functions with the same domain and range, il I can be represented as I(x) = sign(L::=l hi(X», where hi E H, then lor any probability distribution D over {O, 1}n there is some hi such that PrD[f(x) ¥- hi(x)] ~ ~ 218 ' If we specialize this lemma by taking H = Ck (recall that Ck is the set of conjunctions of at most k input features of f) then this implies that for any I E Pk,8 and any probability distribution D over the input features of I there is some hi E Ck that weakly approximates I with respect to D. Therefore, given a training set S and distribution D that has nonzero weight only on instances in S, the following simple algorithm is a weak learning algorithm for Pk: exhaustively test each of the O(nk) possible conjunctions of at most k features until we find a conjunction that a - 218 )-approximates I with respect to D (we can efficiently compute the approximation of a conjunction hi by summing the values of D over those inputs where hi and I agree). Any such conjunction can be returned as the weak hypothesis. The above lemma proves that if I is a k-perceptron then this exhaustive search must succeed at finding such a hypothesis. Therefore, given a training set of m examples of any s-sparse k-perceptron I, AdaBoost run with the above weak learner will, after 2s2In(m) stages, produce a hypothesis consistent with the training set. Because each stage adds one weak hypothesis to the output hypothesis, the final hypothesis will be a real-weighted k-perceptron with at most 2s2In(m) nonzero weights. We can convert this consistency algorithm to a PAC learning algorithm as follows. First, given a finite set of functions F, it is straightforward to show the following (see, e.g., Haussler, 1988): Lemma 2 Let F be a finite set ollunctions over a domain X. For any function lover X, any probability distribution D over X, and any positive f and ~, given a set S ofm examples drawn consecutively from EX(f, D), where m ~ f-1(ln~-1 + In IFI), then Pr[3h E F I "Ix E S f(x) = h(x) & Prv[/(x) ¥- h(x)] > f] < ~, where the outer probability is over the random choices made by EX(f,D). The consistency algorithm above finds a consistent hypothesis in P~, where r = 2s2 In(m). Also, based on a result of Bruck (1990), it can be shown that In IP~I = o (r2 + kr log n). Therefore, ignoring log factors, a randomly-generated training set of size O(kS4 If) is sufficient to guarantee that, with high probability, our algorithm will produce an f-approximator for any s-sparse k-perceptron target. In other words, the following is a PAC algorithm for Pk,8: compute sufficiently large (but polynomial in the PAC parameters) m, draw m examples from EX(f, D) to create a training set, and run the consistency algorithm on this training set. So far we have shown that sparse k-perceptrons are learnable by sparse perceptron hypotheses (with potentially polynomially-many more weights). In practice, of course, we expect that many real-world classification tasks cannot be performed exactly by sparse perceptrons. In fact, it can be shown that for certain (reasonable) definitions of "noisy" sparse perceptrons (loosely, functions that are approximated reasonably well by sparse perceptrons), the class of noisy sparse k-perceptrons is still PAC-learnable. This claim is based on results of Aslam and Decatur (1993), who present a noise-tolerant boosting algorithm. In fact, several different boosting algorithms could be used to learn Pk,s (e.g., Freund, 1993). We have chosen to use AdaBoost because it seems to offer significant practical advantages, particularly in terms of efficiency. Also, our empirical results to date indicate that our algorithm 658 J. C. JACKSON, M. W. CRAVEN works very well on difficult (presumably "noisy") real-world problems. However, one potential advantage of basing the algorithm on one of these earlier boosters instead of AdaBoost is that the algorithm would then produce a perceptron with integer weights while still maintaining the sparseness guarantee of the AdaBoostbased algorithm. 3.3 Practical Considerations We turn now to the practical details of our algorithm, which is based on the consistency algorithm above. First, it should be noted that the theory developed above works over discrete input domains (Boolean or nominal-valued features). Thus, in this paper, we consider only tasks with discrete input features. Also, because the algorithm uses exhaustive search over all conjunctions of size k, learning time depends exponentially on the choice of k. In this study we to use k = 2 throughout, since this choice results in reasonable learning times. Another implementation concern involves deciding when the learning algorithm should terminate. The consistency algorithm uses the size of the target function in calculating the number of boosting stages. Of course, such size information is not available in real-world applications, and in fact, the target function may not be exactly representable as a sparse perceptron. In practice, we use cross validation to determine an appropriate termination point. To facilitate comprehensibility, we also limit the number of boosting stages to at most the number of weights that would occur in an ordinary perceptron for the task. For similar reasons, we also modify the criteria used to select the weak hypothesis at each stage so that simple features are preferred over conjunctive features. In particular, given distribution D at some stage j, for each hi E Ck we compute a correlation Ev[/ . hi]. We then mUltiply each high-order feature's correlation by i. The hi with the largest resulting correlation serves as the weak hypothesis for stage j. 4 Empirical Evaluation In our experiments, we are interested in assessing both the generalization ability and the complexity of the hypotheses produced by our algorithm. We compare our algorithm to ordinary perceptrons trained using backpropagation (Rumelhart et al., 1986), multi-layer perceptrons trained using backpropagation, and decision trees induced using the C4.5 system (Quinlan, 1993). We use C4.5 in our experiments as a representative of "symbolic" learning algorithms. Symbolic algorithms are widely believed to learn hypotheses that are more comprehensible than neural networks. Additionally, to test the hypothesis that the performance of our algorithm can be explained solely by its use of second-order features, we train ordinary perceptrons using feature sets that include all pairwise conjunctions, as well as the ordinary features. To test the hypothesis that the performance of our algorithm can be explained by its use of relatively few weights, we consider ordinary perceptrons which have been pruned using a variant of the Optimal Brain Damage (OBD) algorithm (Le Cun et al., 1989). In our version of OBD, we train a perceptron until the stopping criteria are met, prune the weight with the smallest salience, and then iterate the process. We use a validation set to decide when to stop pruning weights. For each training set, we use cross-validation to select the number of hidden units (5, 10, 20, 40 or 80) for the MLP's, and the pruning confidence level for the C4.5 trees. We use a validation set to decide when to stop training for the MLP's. We evaluate our algorithm using three real-world domains: the voting data set from the UC-Irvine database; a promoter data set which is a more complex superset of Learning Sparse Perceptrons 659 a e : es -se accuracy. T bl 1 11 t t perceptrons domain boosting C4.5 multi-layer ordinary 2nd-order pruned voting 91.5% 89.2% * 92.2% 90.8% 89.2% * 87.6% * promoter 92.7 84.4 * 90.6 90.0 * 88.7 * 88.2 * coding 72.9 62.6 * 71.6 * 70.7 * 69.8 * 70.3 * Table 2: Hypothesis complexity (# weights). perceptrons domain boosting multi-layer ordinary 2nd-order pruned voting 12 651 30 450 12 promoters 41 2267 228 25764 59 protein coding 52 4270 60 1740 37 U C-Irvine one; and a data set in which the task is to recognize protein-coding regions in DNA (Craven & Shavlik, 1993). We remove the physician-fee-freeze feature from the voting data set to make the problem more difficult. We conduct our experiments using a lO-fold cross validation methodology, except for in the protein-coding domain. Because of certain domain-specific characteristics of this data set, we use 4-fold cross-validation for our experiments with it. Table 1 reports test-set accuracy for each method on all three domains. We measure the statistical significance of accuracy differences using a paired, two-tailed t-test. The symbol '*' marks results in cases where another algorithm is less accurate than our boosting algorithm at the p ::; 0.05 level of significance. No other algorithm is significantly better than our boosting method in any of the domains. From these results we conclude that (1) our algorithm exhibits good generalization performance on number of interesting real-world problems, and (2) the generalization performance of our algorithm is not explained solely by its use of second-order features, nor is it solely explained by the sparseness of the perceptrons it produces. An interesting open question is whether perceptrons trained with both pruning and second-order features are able to match the accuracy of our algorithm; we plan to investigate this question in future work. Table 2 reports the average number of weights for all of the perceptrons. For all three problems, our algorithm produces perceptrons with fewer weights than the MLP's, the ordinary perceptrons, and the perceptrons with second-order features. The sizes of the OBD-pruned perceptrons and those produced by our algorithm are comparable for all three domains. Recall, however, that for all three tasks, the perceptrons learned by our algorithm had significantly better generalization performance than their similar-sized OBD-pruned counterparts. We contend that the sizes of the perceptrons produced by our algorithm are within the bounds of what humans can readily understand. In the biological literature, for example, linear discriminant functions are frequently used to communicate domain knowledge about sequences of interest. These functions frequently involve more weights than the perceptrons produced by our algorithm. We conclude, therefore, that our algorithm produces hypotheses that are not only accurate, but also comprehensible. We believe that the results on the protein-coding domain are especially interesting. The input representation for this problem consists of 15 nominal features representing 15 consecutive bases in a DNA sequence. In the regions of DNA that encode proteins (the positive examples in our task), non-overlapping triplets of consecu660 J. C. JACKSON, M. W. eRA VEN tive bases represent meaningful "words" called codons. In previous work (Craven & Shavlik, 1993), it has been found that a feature set that explicitly represents codons results in better generalization than a representation of just bases. However, we used the bases representation in our experiments in order to investigate the ability of our algorithm to select the "right" second-order features. Interestingly, nearly all of the second-order features included in our sparse perceptrons represent conjunctions of bases that are in the same codon. This result suggests that our algorithm is especially good at selecting relevant features from large feature sets. 5 Future Work Our present algorithm has a number of limitations which we plan to address. Two areas of current research are generalizing the algorithm for application to problems with real-valued features and developing methods for automatically suggesting highorder features to be included in our algorithm's feature set. Acknowledgements Mark Craven was partially supported by ONR grant N00014-93-1-0998. Jeff Jackson was partially supported by NSF grant CCR-9119319. References Aslam, J. A. & Decatur, S. E. (1993). General bounds on statistical query learning and PAC learning with noise via hypothesis boosting. In Proc. of the 34th Annual Annual Symposium on Foundations of Computer Science, (pp. 282-291). Bruck, J . (1990). Harmonic analysis of polynomial threshold functions. SIAM Journal of Discrete Mathematics, 3(2):168-177. Craven, M. W. & Shavlik, J. W. (1993). Learning to represent codons: A challenge problem for constructive induction. In Proc. of the 13th International Joint Conf. on Artificial Intelligence, (pp. 1319-1324), Chambery, France. Freund, Y. (1993). Data Filtering and Distribution Modeling Algorithms for Machine Learning. PhD thesis, University of California at Santa Cruz. Freund, Y. & Schapire, R. E. (1995). A decision-theoretic generalization of on-line learning and an application to boosting. In Proc. of the ~nd Annual European Conf. on Computational Learning Theory. Goldmann, M., Hastad, J., & Razborov, A. (1992). Majority gates vs. general weighted threshold gates. In Proc. of the 7th IEEE Conf. on Structure in Complexity Theory. Haussler, D. (1988). Quantifying inductive bias: AI learning algorithms and Valiant's learning framework. Artificial Intelligence, (pp. 177-221). Le Cun, Y., Denker, J. S., & Solla, S. A. (1989). Optimal brain damage. In Touretzky, D., editor, Advances in Neural Information Processing Systems (volume ~) . Quinlan, J. R. (1993). C4.5: Programs for Machine Learning. Morgan Kaufmann. Rumelhart, D., Hinton, G., & Williams, R. (1986). Learning internal representations by error propagation. In Rumelhart, D. & McClelland, J., editors, Parallel Distributed Processing: Explorations in the microstructure of cognition. Volume 1. MIT Press. Spackman, K. A. (1988). Learning categorical decision criteria. In Proc. of the 5th International Conf. on Machine Learning, (pp. 36-46), Ann Arbor, MI. Stormo, G. (1987). Identifying coding sequences. In Bishop, M. J. & Rawlings, C. J., editors, Nucleic Acid and Protein Sequence Analysis: A Practical Approach. IRL Press. Valiant,1. G. (1984). A theory of the learnable. Comm. of the ACM, 27(11):1134-1142.
|
1995
|
83
|
1,132
|
The Role of Activity in Synaptic Competition at the Neuromuscular Junction Samuel R. H. Joseph Centre for Cognitive Science Edinburgh University Edinburgh, U.K. email: sam@cns.ed.ac.uk David J. Willshaw Centre for Cognitive Science Edinburgh University Edinburgh, U.K. email: david@cns.ed.ac.uk Abstract An extended version of the dual constraint model of motor endplate morphogenesis is presented that includes activity dependent and independent competition. It is supported by a wide range of recent neurophysiological evidence that indicates a strong relationship between synaptic efficacy and survival. The computational model is justified at the molecular level and its predictions match the developmental and regenerative behaviour of real synapses. 1 INTRODUCTION The neuromuscular junction (NMJ) of mammalian skeletal muscle is one of the most extensively studied areas of the nervous system. One aspect of its development that it shares with many other parts of the nervous system is its achievement of single innervation, one axon terminal connecting to one muscle fibre, after an initial state of polyinnervation. The presence of electrical activity is associated with this transition, but the exact relationship is far from clear. Understanding how activity interacts with the morphogenesis of neural systems could provide us with insights into methods for constructing artificial neural networks. With that in mind, this paper examines how some of the conflicting ideas about the development of neuromuscular connections can be resolved. The Role of Activity in Synaptic Competition at the Neuromuscular Junction 97 2 EXPERIMENTAL FINDINGS The extent to which a muscle is innervated can be expressed in terms of the motor unit size - the number of fibres contacted by a given motor axon. Following removal of some motor axons at birth, the average size of the remaining motor units after withdrawal of poly innervation is larger than normal (Fladby & Jansen, 1987). This strongly suggests that individual motor axons successfully innervate more fibres as a result of the absence of their neighbours. It is appealing to interpret this as a competitive process where terminals from different axons compete for the same muscle endplate. Since each terminal is made up of a number of synapses the process can be viewed as the co-existence of synapses from the same terminal and the elimination of synapses from different terminals on the same end plate. 2.1 THE EFFECTS OF ELECTRICAL ACTIVITY There is a strong activity dependent component to synapse elimination. Paralysis or stimulation of selected motor units appears to favour the more active motor terminals (Colman & Lichtman, 1992), while inactive axon terminals tend to coexist. Recent work also shows that active synaptic sites can destabilise inactive synapses in their vicinity (Balice-Gordon & Lichtman, 1994). These findings support the idea that more active terminals have a competitive advantage over their inactive fellows, and that this competition takes place at a synaptic level. Activity independent competition has been demonstrated in the rat lumbrical muscle (Ribchester, 1993). This muscle is innervated by the sural and the lateral plantar nerves. If the sural nerve is damaged the lateral plantar nerve will expand its territory to the extent that it innervates the entire muscle. On subsequent reinnervation the regenerating sural nerve may displace some of the lateral plantar nerve terminals. If the muscle is paralysed during reinnervation more lateral plantar nerve terminals are displaced than in the normal case, indicating that competition between inactive terminals does take place, and that paralysis can give an advantage to some terminals. 3 MODELS AND MECHANISMS If the nerve terminals are competing with each other for dominance of motor endplates, what is the mechanism behind it? As mentioned above, activity is thought to play an important role in affecting the competitive chances of a terminal, but in most models the terminals compete for some kind of trophic resource (Gouze et aI., 1983; Willshaw, 1981). It is possible to create models that use competition for either a postsynaptic (endplate) resource or a presynaptic (motor axon) resource. Both types of model have advantages and disadvantages, which leads naturally to the possibility of combining the two into a single model. 3.1 BENNET AND ROBINSON'S DUAL CONSTRAINT MODEL The dual constraint model (DCM) (Bennet & Robinson, 1989), as extended by Rasmussen & Willshaw (1993), is based on a reversible reaction between molecules from a presynaptic resource A and a postsynaptic resource B. This reaction takes place in the synaptic cleft and produces a binding complex C which is essential for 98 S. R. H. JOSEPH, D. J. Wll...LSHAW the terminal's survival. Each motor axon and muscle fibre has a limited amount of their particular resource and the size of each terminal is proportional to the amount of the binding complex at that terminaL The model achieves single innervation and a perturbation analysis performed by Rasmussen & Willshaw (1993) showed that this single innervation state is stable. However, for the DCM to function the forward rate of the reaction had to be made proportional to the size of the terminal, which was difficult to justify other than suggesting it was related to electrical activity. 3.2 SELECTIVE MECHANISMS While the synapses in the surviving presynaptic terminal are allowed to coexist, synapses from other axons are eliminated. How do synapses make a distinction between synapses in their own terminal and those in others? There are two possibilities: (i) Synchronous transmitter release in the synaptic boutons of a motor neuron could distinguish synapses, allowing them to compete as cartels rather than individuals (Colman & Lichtman, 1992). (ii) The synapses could be employing selective recognition mechanisms, e.g the 'induced-fit' model (Rib chester & Barry, 1994). A selective mechanism implies that all the synapses of a given motor neuron can be identified by a molecular substrate. In the induced-fit model each motor neuron is associated with a specific isoform of a cellular adhesion molecule (CAM); the synapses compete by attempting to induce all the CAMs on the end plate into the conformation associated with their neuron. This kind of model can be used to account for much of the developmental and regenerative processes of the NMJ. However, it has difficulty explaining Balice-Gordon & Lichtman's (1994) focal blockade experiments which show competition between synapses distinguished only by the presence of activity. If, instead, activity is responsible for the distinction of friend from foe, how can competition take place at the terminal level when activity is not present? Could we resolve this dilemma by extending the dual constraint model? 4 EXTENDING THE DUAL CONSTRAINT MODEL Tentative suggestions can be made for the identity of the 'mystery molecules' in the DCM. According to McMahan (1990) a protein called agrin is synthesised in the cell bodies of motor neurons and transported down their axons to the muscle. When this protein binds to the surface of the developing muscle, it causes acetylcholine receptors (AChRs), and other components of the postsynaptic apparatus, to aggregate on the myotube surface in the vicinity of the activated agrin. Other work (Wallace, 1988) has provided insights into the mechanism used by agrin to cause the aggregation of the postsynaptic apparatus. Initially, AChR aggregates, or 'speckles', are free to diffuse laterally in the myotube plasma membrane (Axelrod et aI., 1976). When agrin binds to an agrin-specific receptor, AChR speckles in the immediate vicinity of the agrin-receptor complex are immobilised. As more speckles are trapped larger patches are formed, until a steady state is reached. Such a patch will remain so long as agrin is bound to its receptor and Ca++ and energy supplies are available. Following AChR activation by acetylcholine, Ca++ enters the postsynaptic cell. Since Ca++ is required for both the formation and maintenance of AChR aggregates, The Role of Activity in Synaptic Competition at the Neuromuscular Junction 99 a feedback loop is possible whereby the bigger a patch is the more Ca++ it will have available when the receptors are activated. Crucially, depolarisation of nonjunctional regions blocks AChR expression (Andreose et al., 1995) and it is AChR activation at the NMJ that causes depolarisation of the postsynaptic cell. So it seems that agrin is a candidate for molecule A, but what about B or C? It is tempting to posit AChR as molecule B since it is the critical postsynaptic resource. However, since agrin does not bind directly to the acetylcholine receptor, a different sort of reaction is required. 4.1 A DIFFERENT SORT OF REACTION If AChR is molecule B, and one agrin molecule can attract at least 160 AChRs (Nitkin et al., 1987) the simple reversible reaction of the DCM is ruled out. Alternatively, AChR could exist in either free, B f' or bound, Bb states, being converted through the mediation of A. Bb would now play the role of C in the DCM. It is possible to devise a rate equation for the change in the number of receptors at a nerve terminal over time: dBb = nABf - (3Bb dt (1) where n and (3 are rate constants. The increase in bound AChR over time is proportional to the amount of agrin at a junction and the number of free receptors in the endplate area, while the decrease is proportional to the amount of bound AChRs. The rate equation (1) can be used as the basis of an extended DCM if four other factors are considered: (i) Agrin stays active as receptors accumulate, so the conservation equations for A and Bare: M Ao = An + LAnj j=1 N Bo = Bmf + LBimb i=1 (2) where the subscript 0 indicates the fixed resource available to each muscle or neuron, the lettered subscripts indicate the amount of that substance that is present in the neuron n, muscle fibre m and terminal nm, and there are N motor neurons and M muscle fibres. (ii) The size of a terminal is proportional to the number of bound AChRs, so if we assume the anterograde flow is evenly divided between the lin terminals of neuron n, the transport equation for agrin is: (3) where>. and IS are transport rate constants and the retrograde flow is assumed proportional to the amount of agrin at the terminal and inversely proportional to the size of the terminal. (iii) AChRs are free to diffuse laterally across the surface of the muscle, so the forward reaction rate will be related to the probability of an AChR speckle intersecting a terminal, which is itself proportional to the terminal diameter. (iv) The influx of Ca++ through AChRs on the surface of the endplate will also affect the forward reaction rate in proportion to the area of the terminal. Taking Bb to be proportional to the volume of the postsynaptic apparatus, these last two terms are proportional to B~/3 and B;/3 respectively. This gives the final rate equation: (4) 100 S. R. H. JOSEPH, D. J. WILLSHA W Equations (3) and (4) are similar to those in the original DCM, only now we have been able to justify the dependence of the forward reaction rate on the size of the terminal, Bnmb . We can also resolve the distinction paradox, as follows. 4.2 RESOLVING THE DISTINCTION PARADOX In terms of distinguishing between synapses it seems plausible that concurrently active synapses (Le. those belonging to the same neuron) will protect themselves from the negative effects of depolarisation. In paralysed systems, synapses will benefit from the AChR accumulating affects of the agrin molecules in those synapses nearby (i.e. those in the same terminal). It was suggested (Jennings, 1994) that competition between synapses of the same terminal was seen after focal blockade because active AChRs help stabilise the receptors around them and suppress those further away. This fits in with the stabilisation role of Ca++ in this model and the suppressive effects of depolarisation, as well as the physical range of these effects during 'heterosynaptic suppression' (Lo & Poo, 1991). It seems that Jenning's mechanism, although originally speculative, is actually quite a plausible explanation and one that fits in well with the extended DCM. The critical effect in the XDCM is that if the system is paralysed during development there is a change in the dependency of the forward reaction rate on the size of an individual terminal. This gives the reinnervating terminals a small initial advantage due to their more competitive diameter/volume ratios. As we shall see in the next section, this allows us to demonstrate activity independent competition. 5 SIMULATING THE EXTENDED DCM In terms of achieving single innervation the extended DCM performs just as well as the original, and when subjected to the same perturbation analysis it has been demonstrated to be stable. Simulating a number of systems with as many muscle fibres and motor neurons as found in real muscles allowed a direct comparison of model findings with experimental data (figure 1) . ..... -~ ... '; ,-\-. . t\_ '. .~\\ " " .... • E""pcrimental + Simulation ~,~ •• 1 ............. __ ..... . ~.----~----+. .. ----~--~~ Days after birth Figure 1: Elimination of Polyinnervation in Rat soleus muscle and Simulation Figure 2 shows nerve dominance histograms of reinnervation in both the rat lumbrical muscle and its extended DCM simulation. Both compare the results produced when the system is paralysed from the outset of reinnervation (removal of B~~b The Role of Activity in Synaptic Competition at the Neuromuscular Junction 101 term from equation (4)) with the normal situation. Note that in both the simulation and the experiment the percentage of fibres singly innervated by the reinnervating sural nerve is increased in the paralysis case. Inactive sural nerve terminals are displacing more inactive lateral plantar nerve terminals (activity independent competition). They can achieve this because during paralysis the terminals with the largest diameters capture more receptors, while the terminals with the largest volumes lose more agrin; so small reinnervating terminals do a little better. However, if activity is present the receptors are captured in proportion to a terminal's volume, so there's no advantage to a small terminal's larger diameter/volume ratio. I Nerve Dominance Histogram (Experimental) I I I, , I 1:::11 I I I Nerve Dominance Histogram (Simulation) I SingleLPN Multi SingleSN I SingieLPN Mull! SmglcSN ________________________ ~ ~ I ____________ __ Figure 2: Types of Innervation by Lateral Plantar and Sural Nerves 6 DISCUSSION The extensions to the DCM outlined here demonstrate both activity dependent and independent competition and provide greater biochemical plausibility. However this is still only a phenomenological demonstration and further experimental work is required to ascertain its validity. There is a need for illumination concerning the specific chemical mechanisms that underlie agrin's aggregational effects and the roles that both Ca++ and depolarisation play in junctional dynamics. An important connection made here is one between synaptic efficiency and junctional survival. Ca++ and NO have both been implicated in Hebbian mechanisms (Bliss & Collingridge, 1993) and perhaps some of the principles uncovered here may be applicable to neuroneuronic synapses. This work should be followed up with a direct model of synaptic interaction at the NMJ that includes the presynaptic effects of depolarisation, allowing the efficacy of the synapse to be related to its biochemistry; an important step forward in our understanding of nervous system plasticity. Relating changes in synaptic efficiency to neural morphogenesis may also give insights into the construction of artificial neural networks. Acknowledgements We are grateful to Michael Joseph and Bruce Graham for critical reading of the manuscript and to the M.R.C. for funding this work. 102 S. R. H. JOSEPH, D. J. WILLS HAW References Andreose J. S., Fumagalli G. & L0mo T. (1995) Number of junctional acetylcholine receptors: control by neural and muscular influences in the rat. Journal of Physiology 483.2:397-406. Axelrod D., Ravdin P., Koppel D. E., Schlessinger J., Webb W. W., Elson E. L. & Podleski T. R. (1976) Lateral motion offluorescently labelled acetylcholine receptors in membranes of developing muscle fibers. Proc. Natl. Acad. Sci. USA 73:45944598. Balice-Gordon R. J. & Lichtman J. W. (1994) Long-term synapse loss induced by focal blockade of postsynaptic receptors. Nature 372:519-524. Bennett M. R. & Robinson J. (1989) Growth and elimination of nerve terminals during polyneuronal innervation of muscle cells: a trophic hypothesis. Proc. Royal Soc. Lond. [Biol] 235:299-320. Bliss T. V. P. & Collingridge G. L. (1993) A synaptic model of memory: long-term potentiation in the hippocampus. Nature 361:31-39. Colman H. & Lichtman J. W. (1992) 'Cartellian' competition at the neuromuscular junction. Trends in Neuroscience 15, 6:197-199. Fladby T. & Jansen J. K. S. (1987) Postnatal loss of synaptic terminals in the partially denervated mouse soleus muscle. Acta. Physiol. Scand 129:239-246. Gouze J. L., Lasry J. M. & Changeux J. -Po (1983) Selective stabilization of muscle innervation during development: A mathematical model. Biol Cybern. 46:207-215. Jennings C. (1994) Death of a synapse. Nature 372:498-499. Lo Y. J. & Poo M. M. (1991) Activity-dependent synapse competition in vitro: heterosynaptic suppression of developing synapses. Science 254:1019-1022. McMahan U. J. (1990) The Agrin Hypothesis. Cold Spring Harbour Symp. Quant. Biol. 55:407-419. Nitkin R. M., Smith M. A., Magill C., Fallon J. R., Yao Y. -M. M., Wallace B. G. & McMahan U. J. (1987) Identification of agrin, a synaptic organising protein from Torpedo electric organ. Journal Cell Biology 105:2471-2478. Rasmussen C. E. & Willshaw D. J. (1993) Presynaptic and postsynatic competition in models for the development of neuromuscular connections. B. Cyb. 68:409-419. Ribchester R. R. (1993) Co-existence and elimination of convergent motor nerve terminals in reinnervated and paralysed adult rat skeletal muscle. J. Phys. 466: 421-441. Ribchester R. R. & Barry J. A. (1994) Spatial Versus Consumptive Competition at Polyneuronally Innervated Neuromuscular Junctions. Exp. Physiology 79:465-494. Wallace B. G. (1988) Regulation of agrin-induced acetylcholine receptor aggregation by Ca++ and phorbol ester. Journal of Cell Biol. 107:267-278. Willshaw D. J. (1981) The establishment and the subsequent elimination of polyneuronal innervation of developing muscle: theoretical considerations. Proc. Royal Soc. Lond. B212: 233-252.
|
1995
|
84
|
1,133
|