text
stringlengths 35
1.54k
| source
stringclasses 1
value | page
int64 1
800
| book
stringclasses 1
value | chunk_index
int64 0
0
|
|---|---|---|---|---|
mass of the clean pointsx which could have given rise to [UNK]. the autoencoder thus learns a vector field g ( f ( x ) ) −x indicated by the green arrows. this vector field estimates the score ∇xlog pdata ( x ) up to a multiplicative factor that is the average root mean square reconstruction error. 512
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 527
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders 14. 5. 1 estimating the score score matching (, ) is an alternative to maximum likelihood. it hyvarinen 2005 provides a consistent estimator of probability distributions based on encouraging the model to have the same score as the data distribution at every training point x. in this context, the score is a particular gradient field : ∇x log ( ) p x. ( 14. 15 ) score matching is discussed further in section. for the present discussion 18. 4 regarding autoencoders, it is [UNK] to understand that learning the gradient field of log pdata is one way to learn the structure of pdata itself. a very important property of daes is that their training criterion ( with conditionally gaussian p ( x h | ) ) makes the autoencoder learn a vector field ( g ( f ( x ) ) −x ) that estimates the score of the data distribution. this is illustrated in figure. 14. 4 denoising training of a specific kind of autoencoder ( sigmoidal hidden units, linear reconstruction units ) using gaussian noise and mean squared error as the reconstruction cost is equivalent (, ) to training a speci
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 528
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
##oising training of a specific kind of autoencoder ( sigmoidal hidden units, linear reconstruction units ) using gaussian noise and mean squared error as the reconstruction cost is equivalent (, ) to training a specific kind vincent 2011 of undirected probabilistic model called an rbm with gaussian visible units. this kind of model will be described in detail in section ; for the present 20. 5. 1 discussion it [UNK] to know that it is a model that provides an explicit pmodel ( x ; θ ). when the rbm is trained using denoising score matching (, kingma and lecun 2010 ), its learning algorithm is equivalent to denoising training in the corresponding autoencoder. with a fixed noise level, regularized score matching is not a consistent estimator ; it instead recovers a blurred version of the distribution. however, if the noise level is chosen to approach 0 when the number of examples approaches infinity, then consistency is recovered. denoising score matching is discussed in more detail in section. 18. 5 other connections between autoencoders and rbms exist. score matching applied to rbms yields a cost function that is identical
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 528
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
infinity, then consistency is recovered. denoising score matching is discussed in more detail in section. 18. 5 other connections between autoencoders and rbms exist. score matching applied to rbms yields a cost function that is identical to reconstruction error combined with a regularization term similar to the contractive penalty of the cae ( swersky 2011 bengio and delalleau 2009 et al., ). ( ) showed that an autoen - coder gradient provides an approximation to contrastive divergence training of rbms. for continuous - valued x, the denoising criterion with gaussian corruption and reconstruction distribution yields an estimator of the score that is applicable to general encoder and decoder parametrizations (, ). this alain and bengio 2013 means a generic encoder - decoder architecture may be made to estimate the score 513
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 528
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders by training with the squared error criterion | | g f ( ( [UNK] x ) ) − | | 2 ( 14. 16 ) and corruption c ( [UNK] = [UNK] x | ) = ( n [UNK] x ; = µ, σ σ = 2 i ) ( 14. 17 ) with noise variance σ2. see figure for an illustration of how this works. 14. 5 figure 14. 5 : vector field learned by a denoising autoencoder around a 1 - d curved manifold near which the data concentrates in a 2 - d space. each arrow is proportional to the reconstruction minus input vector of the autoencoder and points towards higher probability according to the implicitly estimated probability distribution. the vector field has zeros at both maxima of the estimated density function ( on the data manifolds ) and at minima of that density function. for example, the spiral arm forms a one - dimensional manifold of local maxima that are connected to each other. local minima appear near the middle of the gap between two arms. when the norm of reconstruction error ( shown by the length of the arrows ) is large, it means that probability can be significantly increased by moving in the direction
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 529
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
. local minima appear near the middle of the gap between two arms. when the norm of reconstruction error ( shown by the length of the arrows ) is large, it means that probability can be significantly increased by moving in the direction of the arrow, and that is mostly the case in places of low probability. the autoencoder maps these low probability points to higher probability reconstructions. where probability is maximal, the arrows shrink because the reconstruction becomes more accurate. figure reproduced with permission from ( ). alain and bengio 2013 in general, there is no guarantee that the reconstruction g ( f ( x ) ) minus the input x corresponds to the gradient of any function, let alone to the score. that is 514
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 529
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders why the early results (, ) are specialized to particular parametrizations vincent 2011 where g ( f ( x ) ) −x may be obtained by taking the derivative of another function. kamyshanska and memisevic 2015 vincent 2011 ( ) generalized the results of ( ) by identifying a family of shallow autoencoders such that g ( f ( x ) ) −x corresponds to a score for all members of the family. so far we have described only how the denoising autoencoder learns to represent a probability distribution. more generally, one may want to use the autoencoder as a generative model and draw samples from this distribution. this will be described later, in section. 20. 11 14. 5. 1. 1 historical perspective the idea of using mlps for denoising dates back to the work of ( ) lecun 1987 and ( ). ( ) also used recurrent networks to denoise gallinari et al. 1987 behnke 2001 images. denoising autoencoders are, in some sense, just mlps trained to denoise. however, the name “ denoising autoencoder ” refers to a model that is intended not merely
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 530
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
1987 behnke 2001 images. denoising autoencoders are, in some sense, just mlps trained to denoise. however, the name “ denoising autoencoder ” refers to a model that is intended not merely to learn to denoise its input but to learn a good internal representation as a side [UNK] of learning to denoise. this idea came much later ( vincent et al.,, ). the learned representation may then be used to pretrain a 2008 2010 deeper unsupervised network or a supervised network. like sparse autoencoders, sparse coding, contractive autoencoders and other regularized autoencoders, the motivation for daes was to allow the learning of a very high - capacity encoder while preventing the encoder and decoder from learning a useless identity function. prior to the introduction of the modern dae, inayoshi and kurita 2005 ( ) explored some of the same goals with some of the same methods. their approach minimizes reconstruction error in addition to a supervised objective while injecting noise in the hidden layer of a supervised mlp, with the objective to improve generalization by introducing the reconstruction error and the injected noise. however, their method was based on a
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 530
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
minimizes reconstruction error in addition to a supervised objective while injecting noise in the hidden layer of a supervised mlp, with the objective to improve generalization by introducing the reconstruction error and the injected noise. however, their method was based on a linear encoder and could not learn function families as powerful as can the modern dae. 14. 6 learning manifolds with autoencoders like many other machine learning algorithms, autoencoders exploit the idea that data concentrates around a low - dimensional manifold or a small set of such manifolds, as described in section. some machine learning algorithms exploit 5. 11. 3 this idea only insofar as that they learn a function that behaves correctly on the manifold but may have unusual behavior if given an input that is [UNK] manifold. 515
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 530
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders autoencoders take this idea further and aim to learn the structure of the manifold. to understand how autoencoders do this, we must present some important characteristics of manifolds. an important characterization of a manifold is the set of its tangent planes. at a point x on a d - dimensional manifold, the tangent plane is given by d basis vectors that span the local directions of variation allowed on the manifold. as illustrated in figure, these local directions specify how one can change 14. 6 x infinitesimally while staying on the manifold. all autoencoder training procedures involve a compromise between two forces : 1. learning a representation h of a training example x such that x can be approximately recovered from h through a decoder. the fact that x is drawn from the training data is crucial, because it means the autoencoder need not successfully reconstruct inputs that are not probable under the data generating distribution. 2. satisfying the constraint or regularization penalty. this can be an architec - tural constraint that limits the capacity of the autoencoder, or it can be a regularization term added to the reconstruction cost. these techniques generally prefer solutions that are less sensitive to the
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 531
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
##ization penalty. this can be an architec - tural constraint that limits the capacity of the autoencoder, or it can be a regularization term added to the reconstruction cost. these techniques generally prefer solutions that are less sensitive to the input. clearly, neither force alone would be useful — copying the input to the output is not useful on its own, nor is ignoring the input. instead, the two forces together are useful because they force the hidden representation to capture information about the structure of the data generating distribution. the important principle is that the autoencoder can [UNK] to represent only the variations that are needed to reconstruct training examples. if the data generating distribution concentrates near a low - dimensional manifold, this yields representations that implicitly capture a local coordinate system for this manifold : only the variations tangent to the manifold around x need to correspond to changes in h = f ( x ). hence the encoder learns a mapping from the input space x to a representation space, a mapping that is only sensitive to changes along the manifold directions, but that is insensitive to changes orthogonal to the manifold. a one - dimensional example is illustrated in figure, showing that, by making 14. 7 the reconstruction function insensitive to per
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 531
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
only sensitive to changes along the manifold directions, but that is insensitive to changes orthogonal to the manifold. a one - dimensional example is illustrated in figure, showing that, by making 14. 7 the reconstruction function insensitive to perturbations of the input around the data points, we cause the autoencoder to recover the manifold structure. to understand why autoencoders are useful for manifold learning, it is in - structive to compare them to other approaches. what is most commonly learned to characterize a manifold is a representation of the data points on ( or near ) 516
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 531
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders figure 14. 6 : an illustration of the concept of a tangent hyperplane. here we create a one - dimensional manifold in 784 - dimensional space. we take an mnist image with 784 pixels and transform it by translating it vertically. the amount of vertical translation defines a coordinate along a one - dimensional manifold that traces out a curved path through image space. this plot shows a few points along this manifold. for visualization, we have projected the manifold into two dimensional space using pca. an n - dimensional manifold has an n - dimensional tangent plane at every point. this tangent plane touches the manifold exactly at that point and is oriented parallel to the surface at that point. it defines the space of directions in which it is possible to move while remaining on the manifold. this one - dimensional manifold has a single tangent line. we indicate an example tangent line at one point, with an image showing how this tangent direction appears in image space. gray pixels indicate pixels that do not change as we move along the tangent line, white pixels indicate pixels that brighten, and black pixels indicate pixels that darken. 517
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 532
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders x0 x 1 x2 x 0 0. 0 2. 0 4. 0 6. 0 8. 1 0. r x ( ) identity optimal reconstruction figure 14. 7 : if the autoencoder learns a reconstruction function that is invariant to small perturbations near the data points, it captures the manifold structure of the data. here the manifold structure is a collection of - dimensional manifolds. the dashed diagonal 0 line indicates the identity function target for reconstruction. the optimal reconstruction function crosses the identity function wherever there is a data point. the horizontal arrows at the bottom of the plot indicate the r ( x ) −x reconstruction direction vector at the base of the arrow, in input space, always pointing towards the nearest “ manifold ” ( a single datapoint, in the 1 - d case ). the denoising autoencoder explicitly tries to make the derivative of the reconstruction function r ( x ) small around the data points. the contractive autoencoder does the same for the encoder. although the derivative ofr ( x ) is asked to be small around the data points, it can be large between the data points. the space between the data points corresponds to the region between the manifolds
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 533
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
does the same for the encoder. although the derivative ofr ( x ) is asked to be small around the data points, it can be large between the data points. the space between the data points corresponds to the region between the manifolds, where the reconstruction function must have a large derivative in order to map corrupted points back onto the manifold. the manifold. such a representation for a particular example is also called its embedding. it is typically given by a low - dimensional vector, with less dimensions than the “ ambient ” space of which the manifold is a low - dimensional subset. some algorithms ( non - parametric manifold learning algorithms, discussed below ) directly learn an embedding for each training example, while others learn a more general mapping, sometimes called an encoder, or representation function, that maps any point in the ambient space ( the input space ) to its embedding. manifold learning has mostly focused on unsupervised learning procedures that attempt to capture these manifolds. most of the initial machine learning research on learning nonlinear manifolds has focused on non - parametric methods based on the nearest - neighbor graph. this graph has one node per training example and edges connecting near neighbors to each other. these methods ( scholkopf
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 533
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
the initial machine learning research on learning nonlinear manifolds has focused on non - parametric methods based on the nearest - neighbor graph. this graph has one node per training example and edges connecting near neighbors to each other. these methods ( scholkopf et al., ; 1998 roweis and saul 2000 tenenbaum 2000 brand 2003 belkin, ; et al., ;, ; 518
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 533
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders figure 14. 8 : non - parametric manifold learning procedures build a nearest neighbor graph in which nodes represent training examples a directed edges indicate nearest neighbor relationships. various procedures can thus obtain the tangent plane associated with a neighborhood of the graph as well as a coordinate system that associates each training example with a real - valued vector position, or embedding. it is possible to generalize such a representation to new examples by a form of interpolation. so long as the number of examples is large enough to cover the curvature and twists of the manifold, these approaches work well. images from the qmul multiview face dataset (, gong et al. 2000 ). and niyogi 2003 donoho and grimes 2003 weinberger and saul 2004 hinton, ;, ;, ; and roweis 2003 van der maaten and hinton 2008, ;, ) associate each of nodes with a tangent plane that spans the directions of variations associated with the [UNK] vectors between the example and its neighbors, as illustrated in figure. 14. 8 a global coordinate system can then be obtained through an optimization or solving a linear system. figure illustrates how a manifold can be tiled by a 14. 9 large number of locally linear gauss
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 534
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
its neighbors, as illustrated in figure. 14. 8 a global coordinate system can then be obtained through an optimization or solving a linear system. figure illustrates how a manifold can be tiled by a 14. 9 large number of locally linear gaussian - like patches ( or “ pancakes, ” because the gaussians are flat in the tangent directions ). however, there is a fundamental [UNK] with such local non - parametric approaches to manifold learning, raised in ( ) : if the bengio and monperrus 2005 manifolds are not very smooth ( they have many peaks and troughs and twists ), one may need a very large number of training examples to cover each one of 519
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 534
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders figure 14. 9 : if the tangent planes ( see figure ) at each location are known, then they 14. 6 can be tiled to form a global coordinate system or a density function. each local patch can be thought of as a local euclidean coordinate system or as a locally flat gaussian, or “ pancake, ” with a very small variance in the directions orthogonal to the pancake and a very large variance in the directions defining the coordinate system on the pancake. a mixture of these gaussians provides an estimated density function, as in the manifold parzen window algorithm (, ) or its non - local neural - net based vincent and bengio 2003 variant (, ). bengio et al. 2006c these variations, with no chance to generalize to unseen variations. indeed, these methods can only generalize the shape of the manifold by interpolating between neighboring examples. unfortunately, the manifolds involved in ai problems can have very complicated structure that can be [UNK] to capture from only local interpolation. consider for example the manifold resulting from translation shown in figure. if we watch just one coordinate within the input vector, 14. 6 xi, as the image is translated, we
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 535
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
that can be [UNK] to capture from only local interpolation. consider for example the manifold resulting from translation shown in figure. if we watch just one coordinate within the input vector, 14. 6 xi, as the image is translated, we will observe that one coordinate encounters a peak or a trough in its value once for every peak or trough in brightness in the image. in other words, the complexity of the patterns of brightness in an underlying image template drives the complexity of the manifolds that are generated by performing simple image transformations. this motivates the use of distributed representations and deep learning for capturing manifold structure. 520
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 535
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders 14. 7 contractive autoencoders the contractive autoencoder (,, ) introduces an explicit regularizer rifai et al. 2011a b on the code h = f ( x ), encouraging the derivatives of f to be as small as possible : ω ( ) = h λ ∂f ( ) x ∂x 2 f. ( 14. 18 ) the penalty ω ( h ) is the squared frobenius norm ( sum of squared elements ) of the jacobian matrix of partial derivatives associated with the encoder function. there is a connection between the denoising autoencoder and the contractive autoencoder : ( ) showed that in the limit of small gaussian alain and bengio 2013 input noise, the denoising reconstruction error is equivalent to a contractive penalty on the reconstruction function that maps x to r = g ( f ( x ) ). in other words, denoising autoencoders make the reconstruction function resist small but finite - sized perturbations of the input, while contractive autoencoders make the feature extraction function resist infinitesimal perturbations of the input. when using the jacobian -
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 536
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
function resist small but finite - sized perturbations of the input, while contractive autoencoders make the feature extraction function resist infinitesimal perturbations of the input. when using the jacobian - based contractive penalty to pretrain features f ( x ) for use with a classifier, the best classification accuracy usually results from applying the contractive penalty to f ( x ) rather than to g ( f ( x ) ). a contractive penalty on f ( x ) also has close connections to score matching, as discussed in section. 14. 5. 1 the name contractive arises from the way that the cae warps space. specifi - cally, because the cae is trained to resist perturbations of its input, it is encouraged to map a neighborhood of input points to a smaller neighborhood of output points. we can think of this as contracting the input neighborhood to a smaller output neighborhood. to clarify, the cae is contractive only locally — all perturbations of a training point x are mapped near to f ( x ). globally, two [UNK] points x and xmay be mapped to f ( x ) and f ( x ) points that
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 536
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
cae is contractive only locally — all perturbations of a training point x are mapped near to f ( x ). globally, two [UNK] points x and xmay be mapped to f ( x ) and f ( x ) points that are farther apart than the original points. it is plausible that f be expanding in - between or far from the data manifolds ( see for example what happens in the 1 - d toy example of figure ). when the 14. 7 ω ( h ) penalty is applied to sigmoidal units, one easy way to shrink the jacobian is to make the sigmoid units saturate to or. this encourages the cae to encode 0 1 input points with extreme values of the sigmoid that may be interpreted as a binary code. it also ensures that the cae will spread its code values throughout most of the hypercube that its sigmoidal hidden units can span. we can think of the jacobian matrix j at a point x as approximating the nonlinear encoder f ( x ) as being a linear operator. this allows us to use the word “ contractive ” more formally. in the theory of linear operators, a linear operator 521
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 536
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders is said to be contractive if the norm of jx remains less than or equal to for 1 all unit - norm x. in other words, j is contractive if it shrinks the unit sphere. we can think of the cae as penalizing the frobenius norm of the local linear approximation of f ( x ) at every training point x in order to encourage each of these local linear operator to become a contraction. as described in section, regularized autoencoders learn manifolds by 14. 6 balancing two opposing forces. in the case of the cae, these two forces are reconstruction error and the contractive penalty ω ( h ). reconstruction error alone would encourage the cae to learn an identity function. the contractive penalty alone would encourage the cae to learn features that are constant with respect tox. the compromise between these two forces yields an autoencoder whose derivatives ∂f ( ) x ∂x are mostly tiny. only a small number of hidden units, corresponding to a small number of directions in the input, may have significant derivatives. the goal of the cae is to learn the manifold structure of the data. directions x with large jx rapidly change h, so
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 537
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
hidden units, corresponding to a small number of directions in the input, may have significant derivatives. the goal of the cae is to learn the manifold structure of the data. directions x with large jx rapidly change h, so these are likely to be directions which approximate the tangent planes of the manifold. experiments by ( ) rifai et al. 2011a and ( ) show that training the cae results in most singular values rifai et al. 2011b of j dropping below in magnitude and therefore becoming contractive. however, 1 some singular values remain above, because the reconstruction error penalty 1 encourages the cae to encode the directions with the most local variance. the directions corresponding to the largest singular values are interpreted as the tangent directions that the contractive autoencoder has learned. ideally, these tangent directions should correspond to real variations in the data. for example, a cae applied to images should learn tangent vectors that show how the image changes as objects in the image gradually change pose, as shown in figure. visualizations 14. 6 of the experimentally obtained singular vectors do seem to correspond to meaningful transformations of the input image, as shown in figure. 14. 10 one practical issue with the cae
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 537
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
pose, as shown in figure. visualizations 14. 6 of the experimentally obtained singular vectors do seem to correspond to meaningful transformations of the input image, as shown in figure. 14. 10 one practical issue with the cae regularization criterion is that although it is cheap to compute in the case of a single hidden layer autoencoder, it becomes much more expensive in the case of deeper autoencoders. the strategy followed by rifai 2011a et al. ( ) is to separately train a series of single - layer autoencoders, each trained to reconstruct the previous autoencoder ’ s hidden layer. the composition of these autoencoders then forms a deep autoencoder. because each layer was separately trained to be locally contractive, the deep autoencoder is contractive as well. the result is not the same as what would be obtained by jointly training the entire architecture with a penalty on the jacobian of the deep model, but it captures many of the desirable qualitative characteristics. another practical issue is that the contraction penalty can obtain useless results 522
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 537
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders input point tangent vectors local pca ( no sharing across regions ) contractive autoencoder figure 14. 10 : illustration of tangent vectors of the manifold estimated by local pca and by a contractive autoencoder. the location on the manifold is defined by the input image of a dog drawn from the cifar - 10 dataset. the tangent vectors are estimated by the leading singular vectors of the jacobian matrix ∂h ∂x of the input - to - code mapping. although both local pca and the cae can capture local tangents, the cae is able to form more accurate estimates from limited training data because it exploits parameter sharing across [UNK] locations that share a subset of active hidden units. the cae tangent directions typically correspond to moving or changing parts of the object ( such as the head or legs ). images reproduced with permission from ( ). rifai et al. 2011c if we do not impose some sort of scale on the decoder. for example, the encoder could consist of multiplying the input by a small constant and the decoder could consist of dividing the code by. as approaches, the encoder drives the 0 contractive penalty ω ( h ) to
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 538
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
. for example, the encoder could consist of multiplying the input by a small constant and the decoder could consist of dividing the code by. as approaches, the encoder drives the 0 contractive penalty ω ( h ) to approach without having learned anything about the 0 distribution. meanwhile, the decoder maintains perfect reconstruction. in rifai et al. ( ), this is prevented by tying the weights of 2011a f and g. both f and g are standard neural network layers consisting of an [UNK] transformation followed by an element - wise nonlinearity, so it is straightforward to set the weight matrix of g to be the transpose of the weight matrix of. f 14. 8 predictive sparse decomposition predictive sparse decomposition ( psd ) is a model that is a hybrid of sparse coding and parametric autoencoders ( kavukcuoglu 2008 et al., ). a parametric encoder is trained to predict the output of iterative inference. psd has been applied to unsupervised feature learning for object recognition in images and video ( kavukcuoglu 2009 2010 jarrett 2009 farabet 2011 et al.,, ; et al., ; et al., ), as well
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 538
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
applied to unsupervised feature learning for object recognition in images and video ( kavukcuoglu 2009 2010 jarrett 2009 farabet 2011 et al.,, ; et al., ; et al., ), as well as for audio (, ). the model consists of an encoder [UNK] al. 2011 f ( x ) and a decoder g ( h ) that are both parametric. during training, h is controlled by the 523
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 538
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders optimization algorithm. training proceeds by minimizing | | − | | x g ( ) h 2 + λ | | h 1 + ( ) γ f | | − h x | | 2. ( 14. 19 ) like in sparse coding, the training algorithm alternates between minimization with respect to h and minimization with respect to the model parameters. minimization with respect to h is fast because f ( x ) provides a good initial value of h and the cost function constrains h to remain near f ( x ) anyway. simple gradient descent can obtain reasonable values of in as few as ten steps. h the training procedure used by psd is [UNK] from first training a sparse coding model and then training f ( x ) to predict the values of the sparse coding features. the psd training procedure regularizes the decoder to use parameters for which can infer good code values. f ( ) x predictive sparse coding is an example of learned approximate inference. in section, this topic is developed further. the tools presented in chapter 19. 5 19 make it clear that psd can be interpreted as training a directed sparse coding probabilistic model by maximizing a lower bound on the log - likelihood
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 539
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
. in section, this topic is developed further. the tools presented in chapter 19. 5 19 make it clear that psd can be interpreted as training a directed sparse coding probabilistic model by maximizing a lower bound on the log - likelihood of the model. in practical applications of psd, the iterative optimization is only used during training. the parametric encoder f is used to compute the learned features when the model is deployed. evaluating f is computationally inexpensive compared to inferring h via gradient descent. because f is a [UNK] parametric function, psd models may be stacked and used to initialize a deep network to be trained with another criterion. 14. 9 applications of autoencoders autoencoders have been successfully applied to dimensionality reduction and infor - mation retrieval tasks. dimensionality reduction was one of the first applications of representation learning and deep learning. it was one of the early motivations for studying autoencoders. for example, hinton and salakhutdinov 2006 ( ) trained a stack of rbms and then used their weights to initialize a deep autoencoder with gradually smaller hidden layers, culminating in a bottleneck of 30 units. the resulting code yielded less reconstruction error than
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 539
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
##akhutdinov 2006 ( ) trained a stack of rbms and then used their weights to initialize a deep autoencoder with gradually smaller hidden layers, culminating in a bottleneck of 30 units. the resulting code yielded less reconstruction error than pca into 30 dimensions and the learned representation was qualitatively easier to interpret and relate to the underlying categories, with these categories manifesting as well - separated clusters. lower - dimensional representations can improve performance on many tasks, such as classification. models of smaller spaces consume less memory and runtime. 524
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 539
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 14. autoencoders many forms of dimensionality reduction place semantically related examples near each other, as observed by salakhutdinov and hinton 2007b torralba ( ) and et al. ( ). the hints provided by the mapping to the lower - dimensional space aid 2008 generalization. one task that benefits even more than usual from dimensionality reduction is information retrieval, the task of finding entries in a database that resemble a query entry. this task derives the usual benefits from dimensionality reduction that other tasks do, but also derives the additional benefit that search can become extremely [UNK] in certain kinds of low dimensional spaces. specifically, if we train the dimensionality reduction algorithm to produce a code that is low - dimensional and, then we can store all database entries in a hash table binary mapping binary code vectors to entries. this hash table allows us to perform information retrieval by returning all database entries that have the same binary code as the query. we can also search over slightly less similar entries very [UNK], just by flipping individual bits from the encoding of the query. this approach to information retrieval via dimensionality reduction and binarization is called semantic hashing ( salakhutdin
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 540
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
. we can also search over slightly less similar entries very [UNK], just by flipping individual bits from the encoding of the query. this approach to information retrieval via dimensionality reduction and binarization is called semantic hashing ( salakhutdinov and hinton 2007b 2009b,, ), and has been applied to both textual input ( salakhutdinov and hinton 2007b 2009b,, ) and images ( torralba 2008 weiss 2008 krizhevsky and hinton 2011 et al., ; et al., ;, ). to produce binary codes for semantic hashing, one typically uses an encoding function with sigmoids on the final layer. the sigmoid units must be trained to be saturated to nearly 0 or nearly 1 for all input values. one trick that can accomplish this is simply to inject additive noise just before the sigmoid nonlinearity during training. the magnitude of the noise should increase over time. to fight that noise and preserve as much information as possible, the network must increase the magnitude of the inputs to the sigmoid function, until saturation occurs. the idea of learning a hashing function has been further explored in several directions, including the idea of training the representations so as
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 540
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
information as possible, the network must increase the magnitude of the inputs to the sigmoid function, until saturation occurs. the idea of learning a hashing function has been further explored in several directions, including the idea of training the representations so as to optimize a loss more directly linked to the task of finding nearby examples in the hash table (, ). norouzi and fleet 2011 525
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 540
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15 representation learning in this chapter, we first discuss what it means to learn representations and how the notion of representation can be useful to design deep architectures. we discuss how learning algorithms share statistical strength across [UNK] tasks, including using information from unsupervised tasks to perform supervised tasks. shared representations are useful to handle multiple modalities or domains, or to transfer learned knowledge to tasks for which few or no examples are given but a task representation exists. finally, we step back and argue about the reasons for the success of representation learning, starting with the theoretical advantages of distributed representations ( hinton 1986 et al., ) and deep representations and ending with the more general idea of underlying assumptions about the data generating process, in particular about underlying causes of the observed data. many information processing tasks can be very easy or very [UNK] depending on how the information is represented. this is a general principle applicable to daily life, computer science in general, and to machine learning. for example, it is straightforward for a person to divide 210 by 6 using long division. the task becomes considerably less straightforward if it is instead posed using the roman numeral representation of the numbers. most modern people asked to divide ccx by vi would begin by converting the numbers to the arabic numeral
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 541
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
divide 210 by 6 using long division. the task becomes considerably less straightforward if it is instead posed using the roman numeral representation of the numbers. most modern people asked to divide ccx by vi would begin by converting the numbers to the arabic numeral representation, permitting long division procedures that make use of the place value system. more concretely, we can quantify the asymptotic runtime of various operations using appropriate or inappropriate representations. for example, inserting a number into the correct position in a sorted list of numbers is an o ( n ) operation if the list is represented as a linked list, but only o ( log n ) if the list is represented as a red - black tree. in the context of machine learning, what makes one representation better than 526
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 541
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning another? generally speaking, a good representation is one that makes a subsequent learning task easier. the choice of representation will usually depend on the choice of the subsequent learning task. we can think of feedforward networks trained by supervised learning as per - forming a kind of representation learning. specifically, the last layer of the network is typically a linear classifier, such as a softmax regression classifier. the rest of the network learns to provide a representation to this classifier. training with a supervised criterion naturally leads to the representation at every hidden layer ( but more so near the top hidden layer ) taking on properties that make the classification task easier. for example, classes that were not linearly separable in the input features may become linearly separable in the last hidden layer. in principle, the last layer could be another kind of model, such as a nearest neighbor classifier ( salakhutdinov and hinton 2007a, ). the features in the penultimate layer should learn [UNK] properties depending on the type of the last layer. supervised training of feedforward networks does not involve explicitly imposing any condition on the learned intermediate features. other kinds of representation learning algorithms are
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 542
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
##a, ). the features in the penultimate layer should learn [UNK] properties depending on the type of the last layer. supervised training of feedforward networks does not involve explicitly imposing any condition on the learned intermediate features. other kinds of representation learning algorithms are often explicitly designed to shape the representation in some particular way. for example, suppose we want to learn a representation that makes density estimation easier. distributions with more independences are easier to model, so we could design an objective function that encourages the elements of the representation vector h to be independent. just like supervised networks, unsupervised deep learning algorithms have a main training objective but also learn a representation as a side [UNK]. regardless of how a representation was obtained, it can be used for another task. alternatively, multiple tasks ( some supervised, some unsupervised ) can be learned together with some shared internal representation. most representation learning problems face a [UNK] preserving as much information about the input as possible and attaining nice properties ( such as independence ). representation learning is particularly interesting because it provides one way to perform unsupervised and semi - supervised learning. we often have very large amounts of unlabeled training data and relatively little labeled training data. training with supervised learning techniques on the labeled subset often
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 542
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
particularly interesting because it provides one way to perform unsupervised and semi - supervised learning. we often have very large amounts of unlabeled training data and relatively little labeled training data. training with supervised learning techniques on the labeled subset often results in severe overfitting. semi - supervised learning [UNK] the chance to resolve this overfitting problem by also learning from the unlabeled data. specifically, we can learn good representations for the unlabeled data, and then use these representations to solve the supervised learning task. humans and animals are able to learn from very few labeled examples. we do 527
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 542
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning not yet know how this is possible. many factors could explain improved human performance — for example, the brain may use very large ensembles of classifiers or bayesian inference techniques. one popular hypothesis is that the brain is able to leverage unsupervised or semi - supervised learning. there are many ways to leverage unlabeled data. in this chapter, we focus on the hypothesis that the unlabeled data can be used to learn a good representation. 15. 1 greedy layer - wise unsupervised pretraining unsupervised learning played a key historical role in the revival of deep neural networks, enabling researchers for the first time to train a deep supervised network without requiring architectural specializations like convolution or recurrence. we call this procedure unsupervised pretraining, or more precisely, greedy layer - wise unsupervised pretraining. this procedure is a canonical example of how a representation learned for one task ( unsupervised learning, trying to capture the shape of the input distribution ) can sometimes be useful for another task ( supervised learning with the same input domain ). greedy layer - wise unsupervised pretraining relies on
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 543
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
task ( unsupervised learning, trying to capture the shape of the input distribution ) can sometimes be useful for another task ( supervised learning with the same input domain ). greedy layer - wise unsupervised pretraining relies on a single - layer represen - tation learning algorithm such as an rbm, a single - layer autoencoder, a sparse coding model, or another model that learns latent representations. each layer is pretrained using unsupervised learning, taking the output of the previous layer and producing as output a new representation of the data, whose distribution ( or its relation to other variables such as categories to predict ) is hopefully simpler. see algorithm for a formal description. 15. 1 greedy layer - wise training procedures based on unsupervised criteria have long been used to sidestep the [UNK] of jointly training the layers of a deep neural net for a supervised task. this approach dates back at least as far as the neocognitron ( fukushima 1975, ). the deep learning renaissance of 2006 began with the discovery that this greedy learning procedure could be used to find a good initialization for a joint learning procedure over all the layers, and that this approach could be used to successfully
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 543
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
fukushima 1975, ). the deep learning renaissance of 2006 began with the discovery that this greedy learning procedure could be used to find a good initialization for a joint learning procedure over all the layers, and that this approach could be used to successfully train even fully connected architectures ( hinton 2006 hinton et al., ; and salakhutdinov 2006 hinton 2006 bengio 2007 ranzato 2007a, ;, ; et al., ; et al., ). prior to this discovery, only convolutional deep networks or networks whose depth resulted from recurrence were regarded as feasible to train. today, we now know that greedy layer - wise pretraining is not required to train fully connected deep architectures, but the unsupervised pretraining approach was the first method to succeed. greedy layer - wise pretraining is called greedy because it is a greedy algo - 528
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 543
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning rithm, meaning that it optimizes each piece of the solution independently, one piece at a time, rather than jointly optimizing all pieces. it is called layer - wise because these independent pieces are the layers of the network. specifically, greedy layer - wise pretraining proceeds one layer at a time, training the k - th layer while keeping the previous ones fixed. in particular, the lower layers ( which are trained first ) are not adapted after the upper layers are introduced. it is called unsuper - vised because each layer is trained with an unsupervised representation learning algorithm. however it is also called pretraining, because it is supposed to be only a first step before a joint training algorithm is applied to fine - tune all the layers together. in the context of a supervised learning task, it can be viewed as a regularizer ( in some experiments, pretraining decreases test error without decreasing training error ) and a form of parameter initialization. it is common to use the word “ pretraining ” to refer not only to the pretraining stage itself but to the entire two phase protocol that combines the pretraining phase and a supervised learning phase
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 544
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
and a form of parameter initialization. it is common to use the word “ pretraining ” to refer not only to the pretraining stage itself but to the entire two phase protocol that combines the pretraining phase and a supervised learning phase. the supervised learning phase may involve training a simple classifier on top of the features learned in the pretraining phase, or it may involve supervised fine - tuning of the entire network learned in the pretraining phase. no matter what kind of unsupervised learning algorithm or what model type is employed, in the vast majority of cases, the overall training scheme is nearly the same. while the choice of unsupervised learning algorithm will obviously impact the details, most applications of unsupervised pretraining follow this basic protocol. greedy layer - wise unsupervised pretraining can also be used as initialization for other unsupervised learning algorithms, such as deep autoencoders ( hinton and salakhutdinov 2006, ) and probabilistic models with many layers of latent variables. such models include deep belief networks (, ) and deep hinton et al. 2006 boltzmann machines ( salakhutdinov and
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 544
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
salakhutdinov 2006, ) and probabilistic models with many layers of latent variables. such models include deep belief networks (, ) and deep hinton et al. 2006 boltzmann machines ( salakhutdinov and hinton 2009a, ). these deep generative models will be described in chapter. 20 as discussed in section, it is also possible to have greedy layer - wise 8. 7. 4 supervised pretraining. this builds on the premise that training a shallow network is easier than training a deep one, which seems to have been validated in several contexts (, ). erhan et al. 2010 15. 1. 1 when and why does unsupervised pretraining work? on many tasks, greedy layer - wise unsupervised pretraining can yield substantial improvements in test error for classification tasks. this observation was responsible for the renewed interested in deep neural networks starting in 2006 ( hinton et al., 529
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 544
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning algorithm 15. 1 greedy layer - wise unsupervised pretraining protocol. given the following : unsupervised feature learning algorithm l, which takes a training set of examples and returns an encoder or feature function f. the raw input data is x, with one row per example and f ( 1 ) ( x ) is the output of the first stage encoder on x. in the case where fine - tuning is performed, we use a learner t which takes an initial function f, input examples x ( and in the supervised fine - tuning case, associated targets y ), and returns a tuned function. the number of stages is. m f ←identity function [UNK] x x = for do k,..., m = 1 f ( ) k = ( l [UNK] x ) f f ← ( ) k [UNK] [UNK] x ←f ( ) k ( [UNK] ) end for if fine - tuning then f f,, ←t ( x y ) end if return f 2006 bengio 2007 ranzato 2007a ; et al., ; et al., ). on many other tasks, however, unsupervised pretraining either does not confer
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 545
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
##t ( x y ) end if return f 2006 bengio 2007 ranzato 2007a ; et al., ; et al., ). on many other tasks, however, unsupervised pretraining either does not confer a benefit or even causes noticeable harm. ( ) studied the [UNK] of pretraining on machine learning ma et al. 2015 models for chemical activity prediction and found that, on average, pretraining was slightly harmful, but for many tasks was significantly helpful. because unsupervised pretraining is sometimes helpful but often harmful it is important to understand when and why it works in order to determine whether it is applicable to a particular task. at the outset, it is important to clarify that most of this discussion is restricted to greedy unsupervised pretraining in particular. there are other, completely [UNK] paradigms for performing semi - supervised learning with neural networks, such as virtual adversarial training described in section. it is also possible to 7. 13 train an autoencoder or generative model at the same time as the supervised model. examples of this single - stage approach include the discriminative rbm ( larochelle and bengio
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 545
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
it is also possible to 7. 13 train an autoencoder or generative model at the same time as the supervised model. examples of this single - stage approach include the discriminative rbm ( larochelle and bengio 2008, ) and the ladder network (, ), in which the total rasmus et al. 2015 objective is an explicit sum of the two terms ( one using the labels and one only using the input ). unsupervised pretraining combines two [UNK] ideas. first, it makes use of 530
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 545
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning the idea that the choice of initial parameters for a deep neural network can have a significant regularizing [UNK] on the model ( and, to a lesser extent, that it can improve optimization ). second, it makes use of the more general idea that learning about the input distribution can help to learn about the mapping from inputs to outputs. both of these ideas involve many complicated interactions between several parts of the machine learning algorithm that are not entirely understood. the first idea, that the choice of initial parameters for a deep neural network can have a strong regularizing [UNK] on its performance, is the least well understood. at the time that pretraining became popular, it was understood as initializing the model in a location that would cause it to approach one local minimum rather than another. today, local minima are no longer considered to be a serious problem for neural network optimization. we now know that our standard neural network training procedures usually do not arrive at a critical point of any kind. it remains possible that pretraining initializes the model in a location that would otherwise be inaccessible — for example, a region that is surrounded by areas where the cost function varies so much from one example to another that minibatches give only a very noisy
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 546
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
remains possible that pretraining initializes the model in a location that would otherwise be inaccessible — for example, a region that is surrounded by areas where the cost function varies so much from one example to another that minibatches give only a very noisy estimate of the gradient, or a region surrounded by areas where the hessian matrix is so poorly conditioned that gradient descent methods must use very small steps. however, our ability to characterize exactly what aspects of the pretrained parameters are retained during the supervised training stage is limited. this is one reason that modern approaches typically use simultaneous unsupervised learning and supervised learning rather than two sequential stages. one may also avoid struggling with these complicated ideas about how optimization in the supervised learning stage preserves information from the unsupervised learning stage by simply freezing the parameters for the feature extractors and using supervised learning only to add a classifier on top of the learned features. the other idea, that a learning algorithm can use information learned in the unsupervised phase to perform better in the supervised learning stage, is better understood. the basic idea is that some features that are useful for the unsupervised task may also be useful for the supervised learning task. for example, if we train a
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 546
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
##ed phase to perform better in the supervised learning stage, is better understood. the basic idea is that some features that are useful for the unsupervised task may also be useful for the supervised learning task. for example, if we train a generative model of images of cars and motorcycles, it will need to know about wheels, and about how many wheels should be in an image. if we are fortunate, the representation of the wheels will take on a form that is easy for the supervised learner to access. this is not yet understood at a mathematical, theoretical level, so it is not always possible to predict which tasks will benefit from unsupervised learning in this way. many aspects of this approach are highly dependent on the specific models used. for example, if we wish to add a linear classifier on 531
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 546
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning top of pretrained features, the features must make the underlying classes linearly separable. these properties often occur naturally but do not always do so. this is another reason that simultaneous supervised and unsupervised learning can be preferable — the constraints imposed by the output layer are naturally included from the start. from the point of view of unsupervised pretraining as learning a representation, we can expect unsupervised pretraining to be more [UNK] when the initial representation is poor. one key example of this is the use of word embeddings. words represented by one - hot vectors are not very informative because every two distinct one - hot vectors are the same distance from each other ( squaredl2 distance of ). learned word embeddings naturally encode similarity between words by their 2 distance from each other. because of this, unsupervised pretraining is especially useful when processing words. it is less useful when processing images, perhaps because images already lie in a rich vector space where distances provide a low quality similarity metric. from the point of view of unsupervised pretraining as a regularizer, we can expect unsupervised pretraining
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 547
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
, perhaps because images already lie in a rich vector space where distances provide a low quality similarity metric. from the point of view of unsupervised pretraining as a regularizer, we can expect unsupervised pretraining to be most helpful when the number of labeled examples is very small. because the source of information added by unsupervised pretraining is the unlabeled data, we may also expect unsupervised pretraining to perform best when the number of unlabeled examples is very large. the advantage of semi - supervised learning via unsupervised pretraining with many unlabeled examples and few labeled examples was made particularly clear in 2011 with unsupervised pretraining winning two international transfer learning competitions (, ;, ), in settings where the mesnil et al. 2011 goodfellow et al. 2011 number of labeled examples in the target task was small ( from a handful to dozens of examples per class ). these [UNK] were also documented in carefully controlled experiments by paine 2014 et al. ( ). other factors are likely to be involved. for example, unsupervised pretraining is likely to be most useful when the
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 547
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
per class ). these [UNK] were also documented in carefully controlled experiments by paine 2014 et al. ( ). other factors are likely to be involved. for example, unsupervised pretraining is likely to be most useful when the function to be learned is extremely complicated. unsupervised learning [UNK] from regularizers like weight decay because it does not bias the learner toward discovering a simple function but rather toward discovering feature functions that are useful for the unsupervised learning task. if the true underlying functions are complicated and shaped by regularities of the input distribution, unsupervised learning can be a more appropriate regularizer. these caveats aside, we now analyze some success cases where unsupervised pretraining is known to cause an improvement, and explain what is known about why this improvement occurs. unsupervised pretraining has usually been used to improve classifiers, and is usually most interesting from the point of view of 532
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 547
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning figure 15. 1 : visualization via nonlinear projection of the learning trajectories of [UNK] neural networks in function space ( not parameter space, to avoid the issue of many - to - one mappings from parameter vectors to functions ), with [UNK] random initializations and with or without unsupervised pretraining. each point corresponds to a [UNK] neural network, at a particular time during its training process. this figure is adapted with permission from ( ). a coordinate in function space is an infinite - erhan et al. 2010 dimensional vector associating every input x with an output y. ( ) made erhan et al. 2010 a linear projection to high - dimensional space by concatenating they for many specific x points. they then made a further nonlinear projection to 2 - d by isomap ( tenenbaum et al., ). color indicates time. all networks are initialized near the center of the plot 2000 ( corresponding to the region of functions that produce approximately uniform distributions over the class y for most inputs ). over time, learning moves the function outward, to points that make strong predictions. training consistently terminates in one region when using pretraining and in another
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 548
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
corresponding to the region of functions that produce approximately uniform distributions over the class y for most inputs ). over time, learning moves the function outward, to points that make strong predictions. training consistently terminates in one region when using pretraining and in another, non - overlapping region when not using pretraining. isomap tries to preserve global relative distances ( and hence volumes ) so the small region corresponding to pretrained models may indicate that the pretraining - based estimator has reduced variance. 533
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 548
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning reducing test set error. however, unsupervised pretraining can help tasks other than classification, and can act to improve optimization rather than being merely a regularizer. for example, it can improve both train and test reconstruction error for deep autoencoders ( hinton and salakhutdinov 2006, ). erhan 2010 et al. ( ) performed many experiments to explain several successes of unsupervised pretraining. both improvements to training error and improvements to test error may be explained in terms of unsupervised pretraining taking the parameters into a region that would otherwise be inaccessible. neural network training is non - deterministic, and converges to a [UNK] function every time it is run. training may halt at a point where the gradient becomes small, a point where early stopping ends training to prevent overfitting, or at a point where the gradient is large but it is [UNK] to find a downhill step due to problems such as stochasticity or poor conditioning of the hessian. neural networks that receive unsupervised pretraining consistently halt in the same region of function space, while neural networks without pretraining consistently halt in another region. see
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 549
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
such as stochasticity or poor conditioning of the hessian. neural networks that receive unsupervised pretraining consistently halt in the same region of function space, while neural networks without pretraining consistently halt in another region. see figure for a visualization of this phenomenon. the region where pretrained 15. 1 networks arrive is smaller, suggesting that pretraining reduces the variance of the estimation process, which can in turn reduce the risk of severe over - fitting. in other words, unsupervised pretraining initializes neural network parameters into a region that they do not escape, and the results following this initialization are more consistent and less likely to be very bad than without this initialization. erhan 2010 et al. ( ) also provide some answers as to pretraining works when best — the mean and variance of the test error were most reduced by pretraining for deeper networks. keep in mind that these experiments were performed before the invention and popularization of modern techniques for training very deep networks ( rectified linear units, dropout and batch normalization ) so less is known about the [UNK] of unsupervised pretraining in conjunction with contemporary approaches. an important question is how un
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 549
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
modern techniques for training very deep networks ( rectified linear units, dropout and batch normalization ) so less is known about the [UNK] of unsupervised pretraining in conjunction with contemporary approaches. an important question is how unsupervised pretraining can act as a regularizer. one hypothesis is that pretraining encourages the learning algorithm to discover features that relate to the underlying causes that generate the observed data. this is an important idea motivating many other algorithms besides unsupervised pretraining, and is described further in section. 15. 3 compared to other forms of unsupervised learning, unsupervised pretraining has the disadvantage that it operates with two separate training phases. many regularization strategies have the advantage of allowing the user to control the strength of the regularization by adjusting the value of a single hyperparameter. unsupervised pretraining does not [UNK] a clear way to adjust the the strength of the regularization arising from the unsupervised stage. instead, there are 534
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 549
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning very many hyperparameters, whose [UNK] may be measured after the fact but is often [UNK] to predict ahead of time. when we perform unsupervised and supervised learning simultaneously, instead of using the pretraining strategy, there is a single hyperparameter, usually a [UNK] attached to the unsupervised cost, that determines how strongly the unsupervised objective will regularize the supervised model. one can always predictably obtain less regularization by decreasing this [UNK]. in the case of unsupervised pretraining, there is not a way of flexibly adapting the strength of the regularization — either the supervised model is initialized to pretrained parameters, or it is not. another disadvantage of having two separate training phases is that each phase has its own hyperparameters. the performance of the second phase usually cannot be predicted during the first phase, so there is a long delay between proposing hyperparameters for the first phase and being able to update them using feedback from the second phase. the most principled approach is to use validation set error in the supervised phase in order to select the hyperparameters of the pretraining phase, as discussed in ( ).
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 550
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
and being able to update them using feedback from the second phase. the most principled approach is to use validation set error in the supervised phase in order to select the hyperparameters of the pretraining phase, as discussed in ( ). in practice, some hyperparameters, larochelle et al. 2009 like the number of pretraining iterations, are more conveniently set during the pretraining phase, using early stopping on the unsupervised objective, which is not ideal but computationally much cheaper than using the supervised objective. today, unsupervised pretraining has been largely abandoned, except in the field of natural language processing, where the natural representation of words as one - hot vectors conveys no similarity information and where very large unlabeled sets are available. in that case, the advantage of pretraining is that one can pretrain once on a huge unlabeled set ( for example with a corpus containing billions of words ), learn a good representation ( typically of words, but also of sentences ), and then use this representation or fine - tune it for a supervised task for which the training set contains substantially fewer examples. this approach was pioneered by by collobert and
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 550
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
learn a good representation ( typically of words, but also of sentences ), and then use this representation or fine - tune it for a supervised task for which the training set contains substantially fewer examples. this approach was pioneered by by collobert and weston 2008b turian 2010 collobert ( ), et al. ( ), and et al. ( ) and remains in common use today. 2011a deep learning techniques based on supervised learning, regularized with dropout or batch normalization, are able to achieve human - level performance on very many tasks, but only with extremely large labeled datasets. these same techniques out - perform unsupervised pretraining on medium - sized datasets such as cifar - 10 and mnist, which have roughly 5, 000 labeled examples per class. on extremely small datasets, such as the alternative splicing dataset, bayesian methods outperform methods based on unsupervised pretraining ( srivastava 2013, ). for these reasons, the popularity of unsupervised pretraining has declined. nevertheless, unsupervised pretraining remains an important milestone in the history of deep learning research 535
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 550
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning and continues to influence contemporary approaches. the idea of pretraining has been generalized to supervised pretraining discussed in section, as a very 8. 7. 4 common approach for transfer learning. supervised pretraining for transfer learning is popular (, ; oquab et al. 2014 yosinski 2014 et al., ) for use with convolutional networks pretrained on the imagenet dataset. practitioners publish the parameters of these trained networks for this purpose, just like pretrained word vectors are published for natural language tasks (, ; collobert et al. 2011a mikolov 2013a et al., ). 15. 2 transfer learning and domain adaptation transfer learning and domain adaptation refer to the situation where what has been learned in one setting ( i. e., distribution p1 ) is exploited to improve generalization in another setting ( say distribution p2 ). this generalizes the idea presented in the previous section, where we transferred representations between an unsupervised learning task and a supervised learning task. in transfer learning, the learner must perform two or more [UNK] tasks, but we assume that many of the factors that explain the variations in p1 are relevant to the
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 551
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
representations between an unsupervised learning task and a supervised learning task. in transfer learning, the learner must perform two or more [UNK] tasks, but we assume that many of the factors that explain the variations in p1 are relevant to the variations that need to be captured for learning p2. this is typically understood in a supervised learning context, where the input is the same but the target may be of a [UNK] nature. for example, we may learn about one set of visual categories, such as cats and dogs, in the first setting, then learn about a [UNK] set of visual categories, such as ants and wasps, in the second setting. if there is significantly more data in the first setting ( sampled from p1 ), then that may help to learn representations that are useful to quickly generalize from only very few examples drawn from p2. many visual categories share low - level notions of edges and visual shapes, the [UNK] of geometric changes, changes in lighting, etc. in general, transfer learning, multi - task learning ( section ), and domain 7. 7 adaptation can be achieved via representation learning when there exist features that are useful for the [UNK] settings or tasks, corresponding to underlying factors that appear in more than one setting
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 551
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
general, transfer learning, multi - task learning ( section ), and domain 7. 7 adaptation can be achieved via representation learning when there exist features that are useful for the [UNK] settings or tasks, corresponding to underlying factors that appear in more than one setting. this is illustrated in figure, with 7. 2 shared lower layers and task - dependent upper layers. however, sometimes, what is shared among the [UNK] tasks is not the semantics of the input but the semantics of the output. for example, a speech recognition system needs to produce valid sentences at the output layer, but the earlier layers near the input may need to recognize very [UNK] versions of the same phonemes or sub - phonemic vocalizations depending on which person is speaking. in cases like these, it makes more sense to share the upper layers ( near the output ) of the neural network, and have a task - specific preprocessing, as 536
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 551
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning illustrated in figure. 15. 2 selection switch h ( 1 ) h ( 1 ) h ( 2 ) h ( 2 ) h ( 3 ) h ( 3 ) y h ( shared ) h ( shared ) x ( 1 ) x ( 1 ) x ( 2 ) x ( 2 ) x ( 3 ) x ( 3 ) figure 15. 2 : example architecture for multi - task or transfer learning when the output variable has the same semantics for all tasks while the input variable has a [UNK] y x meaning ( and possibly even a [UNK] dimension ) for each task ( or, for example, each user ), called x ( 1 ), x ( 2 ) and x ( 3 ) for three tasks. the lower levels ( up to the selection switch ) are task - specific, while the upper levels are shared. the lower levels learn to translate their task - specific input into a generic set of features. in the related case of domain adaptation, the task ( and the optimal input - to - output mapping ) remains the same between each setting, but the input distribution is slightly [UNK]. for example, consider the task of sentiment analysis, which consists of determining whether a comment expresses positive or negative sentiment. comments posted on
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 552
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
the optimal input - to - output mapping ) remains the same between each setting, but the input distribution is slightly [UNK]. for example, consider the task of sentiment analysis, which consists of determining whether a comment expresses positive or negative sentiment. comments posted on the web come from many categories. a domain adaptation scenario can arise when a sentiment predictor trained on customer reviews of media content such as books, videos and music is later used to analyze comments about consumer electronics such as televisions or smartphones. one can imagine that there is an underlying function that tells whether any statement is positive, neutral or negative, but of course the vocabulary and style may vary from one domain to another, making it more [UNK] to generalize across domains. simple unsupervised pretraining ( with denoising autoencoders ) has been found to be very successful for sentiment analysis with domain adaptation (, ). glorot et al. 2011b a related problem is that of concept drift, which we can view as a form of transfer learning due to gradual changes in the data distribution over time. both concept drift and transfer learning can be viewed as particular forms of 537
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 552
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning multi - task learning. while the phrase “ multi - task learning ” typically refers to supervised learning tasks, the more general notion of transfer learning is applicable to unsupervised learning and reinforcement learning as well. in all of these cases, the objective is to take advantage of data from the first setting to extract information that may be useful when learning or even when directly making predictions in the second setting. the core idea of representation learning is that the same representation may be useful in both settings. using the same representation in both settings allows the representation to benefit from the training data that is available for both tasks. as mentioned before, unsupervised deep learning for transfer learning has found success in some machine learning competitions (, ; mesnil et al. 2011 goodfellow et al., ). in the first of these competitions, the experimental setup is the 2011 following. each participant is first given a dataset from the first setting ( from distribution p1 ), illustrating examples of some set of categories. the participants must use this to learn a good feature space ( mapping the raw input to some representation ), such that when we apply this learned transformation to inputs from the transfer setting ( distribution p
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 553
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
p1 ), illustrating examples of some set of categories. the participants must use this to learn a good feature space ( mapping the raw input to some representation ), such that when we apply this learned transformation to inputs from the transfer setting ( distribution p2 ), a linear classifier can be trained and generalize well from very few labeled examples. one of the most striking results found in this competition is that as an architecture makes use of deeper and deeper representations ( learned in a purely unsupervised way from data collected in the first setting, p1 ), the learning curve on the new categories of the second ( transfer ) setting p2 becomes much better. for deep representations, fewer labeled examples of the transfer tasks are necessary to achieve the apparently asymptotic generalization performance. two extreme forms of transfer learning are one - shot learning and zero - shot learning, sometimes also called zero - data learning. only one labeled example of the transfer task is given for one - shot learning, while no labeled examples are given at all for the zero - shot learning task. one - shot learning ( fei - fei 2006 et al., ) is possible because the representation learns to cleanly separate the underlying classes during the first stage. during the transfer
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 553
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
labeled examples are given at all for the zero - shot learning task. one - shot learning ( fei - fei 2006 et al., ) is possible because the representation learns to cleanly separate the underlying classes during the first stage. during the transfer learning stage, only one labeled example is needed to infer the label of many possible test examples that all cluster around the same point in representation space. this works to the extent that the factors of variation corresponding to these invariances have been cleanly separated from the other factors, in the learned representation space, and we have somehow learned which factors do and do not matter when discriminating objects of certain categories. as an example of a zero - shot learning setting, consider the problem of having a learner read a large collection of text and then solve object recognition problems. 538
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 553
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning it may be possible to recognize a specific object class even without having seen an image of that object, if the text describes the object well enough. for example, having read that a cat has four legs and pointy ears, the learner might be able to guess that an image is a cat, without having seen a cat before. zero - data learning ( larochelle 2008 palatucci et al., ) and zero - shot learning ( et al., ; 2009 socher 2013b et al., ) are only possible because additional information has been exploited during training. we can think of the zero - data learning scenario as including three random variables : the traditional inputs x, the traditional outputs or targets y, and an additional random variable describing the task, t. the model is trained to estimate the conditional distribution p ( y x |, t ) where t is a description of the task we wish the model to perform. in our example of recognizing cats after having read about cats, the output is a binary variable y with y = 1 indicating “ yes ” and y = 0 indicating “ no. ” the task variable t then represents questions to be answered such as “ is there a cat in this image? ” if
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 554
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
about cats, the output is a binary variable y with y = 1 indicating “ yes ” and y = 0 indicating “ no. ” the task variable t then represents questions to be answered such as “ is there a cat in this image? ” if we have a training set containing unsupervised examples of objects that live in the same space as t, we may be able to infer the meaning of unseen instances of t. in our example of recognizing cats without having seen an image of the cat, it is important that we have had unlabeled text data containing sentences such as “ cats have four legs ” or “ cats have pointy ears. ” zero - shot learning requires t to be represented in a way that allows some sort of generalization. for example, t cannot be just a one - hot code indicating an object category. ( ) provide instead a distributed representation socher et al. 2013b of object categories by using a learned word embedding for the word associated with each category. a similar phenomenon happens in machine translation ( klementiev 2012 et al., ; mikolov 2013b gouws 2014 et al., ; et al., ) : we have words in one language, and the relationships between words
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 554
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
similar phenomenon happens in machine translation ( klementiev 2012 et al., ; mikolov 2013b gouws 2014 et al., ; et al., ) : we have words in one language, and the relationships between words can be learned from unilingual corpora ; on the other hand, we have translated sentences which relate words in one language with words in the other. even though we may not have labeled examples translating word a in language x to word b in language y, we can generalize and guess a translation for word a because we have learned a distributed representation for words in language x, a distributed representation for words in language y, and created a link ( possibly two - way ) relating the two spaces, via training examples consisting of matched pairs of sentences in both languages. this transfer will be most successful if all three ingredients ( the two representations and the relations between them ) are learned jointly. zero - shot learning is a particular form of transfer learning. the same principle explains how one can perform multi - modal learning, capturing a representation 539
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 554
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning hx = fx ( ) x xtest ytest hy = fy ( ) y y−space relationship between embedded points within one of the domains maps between representation spaces fx fy x−space ( ) pairs in the training set x y, fx : encoder function for x fy : encoder function for y figure 15. 3 : transfer learning between two domains x and y enables zero - shot learning. labeled or unlabeled examples of x allow one to learn a representation function fx and similarly with examples of y to learn fy. each application of the f x and f y functions appears as an upward arrow, with the style of the arrows indicating which function is applied. distance in hx space provides a similarity metric between any pair of points in x space that may be more meaningful than distance in x space. likewise, distance in hy space provides a similarity metric between any pair of points in y space. both of these similarity functions are indicated with dotted bidirectional arrows. labeled examples ( dashed horizontal lines ) are pairs ( x y, ) which allow one to learn a one - way or two - way map ( solid bidirectional arrow ) between the representationsfx ( x ) and
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 555
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
##irectional arrows. labeled examples ( dashed horizontal lines ) are pairs ( x y, ) which allow one to learn a one - way or two - way map ( solid bidirectional arrow ) between the representationsfx ( x ) and the representations f y ( y ) and anchor these representations to each other. zero - data learning is then enabled as follows. one can associate an image xtest to a word ytest, even if no image of that word was ever presented, simply because word - representationsfy ( ytest ) and image - representations fx ( xtest ) can be related to each other via the maps between representation spaces. it works because, although that image and that word were never paired, their respective feature vectors fx ( xtest ) and fy ( ytest ) have been related to each other. figure inspired from suggestion by hrant khachatrian. 540
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 555
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning in one modality, a representation in the other, and the relationship ( in general a joint distribution ) between pairs ( x y, ) consisting of one observation x in one modality and another observation y in the other modality ( srivastava and salakhutdinov, 2012 ). by learning all three sets of parameters ( from x to its representation, from y to its representation, and the relationship between the two representations ), concepts in one representation are anchored in the other, and vice - versa, allowing one to meaningfully generalize to new pairs. the procedure is illustrated in figure. 15. 3 15. 3 semi - supervised disentangling of causal factors an important question about representation learning is “ what makes one repre - sentation better than another? ” one hypothesis is that an ideal representation is one in which the features within the representation correspond to the under - lying causes of the observed data, with separate features or directions in feature space corresponding to [UNK] causes, so that the representation disentangles the causes from one another. this hypothesis motivates approaches in which we first seek a good representation for p ( x ). such a representation may also be a good representation for computing p (
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 556
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
causes, so that the representation disentangles the causes from one another. this hypothesis motivates approaches in which we first seek a good representation for p ( x ). such a representation may also be a good representation for computing p ( y x | ) if y is among the most salient causes of x. this idea has guided a large amount of deep learning research since at least the 1990s ( becker and hinton 1992 hinton and sejnowski 1999, ;, ), in more detail. for other arguments about when semi - supervised learning can outperform pure supervised learning, we refer the reader to section 1. 2 of ( ). chapelle et al. 2006 in other approaches to representation learning, we have often been concerned with a representation that is easy to model — for example, one whose entries are sparse, or independent from each other. a representation that cleanly separates the underlying causal factors may not necessarily be one that is easy to model. however, a further part of the hypothesis motivating semi - supervised learning via unsupervised representation learning is that for many ai tasks, these two properties coincide : once we are able to obtain the underlying explanations for what we observe, it generally becomes easy to isolate individual attributes from the
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 556
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
##tivating semi - supervised learning via unsupervised representation learning is that for many ai tasks, these two properties coincide : once we are able to obtain the underlying explanations for what we observe, it generally becomes easy to isolate individual attributes from the others. specifically, if a representation h represents many of the underlying causes of the observed x, and the outputs y are among the most salient causes, then it is easy to predict from. y h first, let us see how semi - supervised learning can fail because unsupervised learning of p ( x ) is of no help to learn p ( y x | ). consider for example the case where p ( x ) is uniformly distributed and we want to learn f ( x ) = e [ y | x ]. clearly, observing a training set of values alone gives us no information about. x p ( ) y x | 541
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 556
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning x p x ( ) y = 1 y = 2 y = 3 figure 15. 4 : example of a density over x that is a mixture over three components. the component identity is an underlying explanatory factor, y. because the mixture components ( e. g., natural object classes in image data ) are statistically salient, just modeling p ( x ) in an unsupervised way with no labeled example already reveals the factor y. next, let us see a simple example of how semi - supervised learning can succeed. consider the situation where x arises from a mixture, with one mixture component per value of y, as illustrated in figure. if the mixture components are well - 15. 4 separated, then modeling p ( x ) reveals precisely where each component is, and a single labeled example of each class will then be enough to perfectly learn p ( y x | ). but more generally, what could make and be tied together? p ( ) y x | p ( ) x if y is closely associated with one of the causal factors of x, then p ( x ) and p ( y x | ) will be strongly tied, and unsupervised representation learning that tries to disentangle
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 557
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
| p ( ) x if y is closely associated with one of the causal factors of x, then p ( x ) and p ( y x | ) will be strongly tied, and unsupervised representation learning that tries to disentangle the underlying factors of variation is likely to be useful as a semi - supervised learning strategy. consider the assumption that y is one of the causal factors of x, and let h represent all those factors. the true generative process can be conceived as structured according to this directed graphical model, with as the parent of : h x p, p p. ( h x ) = ( ) x h | ( ) h ( 15. 1 ) as a consequence, the data has marginal probability p ( ) = x ehp. ( ) x h | ( 15. 2 ) from this straightforward observation, we conclude that the best possible model of x ( from a generalization point of view ) is the one that uncovers the above “ true ” 542
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 557
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning structure, with h as a latent variable that explains the observed variations in x. the “ ideal ” representation learning discussed above should thus recover these latent factors. if y is one of these ( or closely related to one of them ), then it will be very easy to learn to predict y from such a representation. we also see that the conditional distribution of y given x is tied by bayes ’ rule to the components in the above equation : p ( ) = y x | p p ( ) x y | ( ) y p ( ) x. ( 15. 3 ) thus the marginal p ( x ) is intimately tied to the conditional p ( y x | ) and knowledge of the structure of the former should be helpful to learn the latter. therefore, in situations respecting these assumptions, semi - supervised learning should improve performance. an important research problem regards the fact that most observations are formed by an extremely large number of underlying causes. suppose y = hi, but the unsupervised learner does not know which hi. the brute force solution is for an unsupervised learner to learn a representation that captures the reasonably all salient generative factors hj and disentangles them from each other, thus
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 558
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
##ed learner does not know which hi. the brute force solution is for an unsupervised learner to learn a representation that captures the reasonably all salient generative factors hj and disentangles them from each other, thus making it easy to predict from, regardless of which h y h i is associated with. y in practice, the brute force solution is not feasible because it is not possible to capture all or most of the factors of variation that influence an observation. for example, in a visual scene, should the representation always encode all of the smallest objects in the background? it is a well - documented psychological phenomenon that human beings fail to perceive changes in their environment that are not immediately relevant to the task they are performing — see, e. g., simons and levin 1998 ( ). an important research frontier in semi - supervised learning is determining to encode in each situation. currently, two of the main strategies what for dealing with a large number of underlying causes are to use a supervised learning signal at the same time as the unsupervised learning signal so that the model will choose to capture the most relevant factors of variation, or to use much larger representations if using purely unsupervised learning.
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 558
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
to use a supervised learning signal at the same time as the unsupervised learning signal so that the model will choose to capture the most relevant factors of variation, or to use much larger representations if using purely unsupervised learning. an emerging strategy for unsupervised learning is to modify the definition of which underlying causes are most salient. historically, autoencoders and generative models have been trained to optimize a fixed criterion, often similar to mean squared error. these fixed criteria determine which causes are considered salient. for example, mean squared error applied to the pixels of an image implicitly specifies that an underlying cause is only salient if it significantly changes the brightness of a large number of pixels. this can be problematic if the task we wish to solve involves interacting with small objects. see figure for an example 15. 5 543
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 558
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning input reconstruction figure 15. 5 : an autoencoder trained with mean squared error for a robotics task has failed to reconstruct a ping pong ball. the existence of the ping pong ball and all of its spatial coordinates are important underlying causal factors that generate the image and are relevant to the robotics task. unfortunately, the autoencoder has limited capacity, and the training with mean squared error did not identify the ping pong ball as being salient enough to encode. images graciously provided by chelsea finn. of a robotics task in which an autoencoder has failed to learn to encode a small ping pong ball. this same robot is capable of successfully interacting with larger objects, such as baseballs, which are more salient according to mean squared error. other definitions of salience are possible. for example, if a group of pixels follow a highly recognizable pattern, even if that pattern does not involve extreme brightness or darkness, then that pattern could be considered extremely salient. one way to implement such a definition of salience is to use a recently developed approach called generative adversarial networks (, ). goodfellow et al. 2014c in this approach
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 559
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
could be considered extremely salient. one way to implement such a definition of salience is to use a recently developed approach called generative adversarial networks (, ). goodfellow et al. 2014c in this approach, a generative model is trained to fool a feedforward classifier. the feedforward classifier attempts to recognize all samples from the generative model as being fake, and all samples from the training set as being real. in this framework, any structured pattern that the feedforward network can recognize is highly salient. the generative adversarial network will be described in more detail in section. for the purposes of the present discussion, it is [UNK] to 20. 10. 4 understand that they learn how to determine what is salient. ( ) lotter et al. 2015 showed that models trained to generate images of human heads will often neglect to generate the ears when trained with mean squared error, but will successfully generate the ears when trained with the adversarial framework. because the ears are not extremely bright or dark compared to the surrounding skin, they are not especially salient according to mean squared error loss, but their highly 544
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 559
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning ground truth mse adversarial figure 15. 6 : predictive generative networks provide an example of the importance of learning which features are salient. in this example, the predictive generative network has been trained to predict the appearance of a 3 - d model of a human head at a specific viewing angle. ( left ) ground truth. this is the correct image, that the network should emit. image produced by a predictive generative network trained with mean ( center ) squared error alone. because the ears do not cause an extreme [UNK] in brightness compared to the neighboring skin, they were not [UNK] salient for the model to learn to represent them. ( right ) image produced by a model trained with a combination of mean squared error and adversarial loss. using this learned cost function, the ears are salient because they follow a predictable pattern. learning which underlying causes are important and relevant enough to model is an important active area of research. figures graciously provided by ( ). lotter et al. 2015 recognizable shape and consistent position means that a feedforward network can easily learn to detect them, making them highly salient under the generative adversarial framework. see figure
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 560
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
##ly provided by ( ). lotter et al. 2015 recognizable shape and consistent position means that a feedforward network can easily learn to detect them, making them highly salient under the generative adversarial framework. see figure for example images. generative adversarial 15. 6 networks are only one step toward determining which factors should be represented. we expect that future research will discover better ways of determining which factors to represent, and develop mechanisms for representing [UNK] factors depending on the task. a benefit of learning the underlying causal factors, as pointed out by scholkopf et al. ( ), is that if the true generative process has 2012 x as an [UNK] and y as a cause, then modeling p ( x y | ) is robust to changes in p ( y ). if the cause - [UNK] relationship was reversed, this would not be true, since by bayes ’ rule, p ( x y | ) would be sensitive to changes in p ( y ). very often, when we consider changes in distribution due to [UNK] domains, temporal non - stationarity, or changes in the nature of the task, the causal mechanisms remain invariant ( the laws of the universe are constant ) while the marginal distribution over
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 560
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
. very often, when we consider changes in distribution due to [UNK] domains, temporal non - stationarity, or changes in the nature of the task, the causal mechanisms remain invariant ( the laws of the universe are constant ) while the marginal distribution over the underlying causes can change. hence, better generalization and robustness to all kinds of changes can 545
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 560
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning be expected via learning a generative model that attempts to recover the causal factors and. h p ( ) x h | 15. 4 distributed representation distributed representations of concepts — representations composed of many ele - ments that can be set separately from each other — are one of the most important tools for representation learning. distributed representations are powerful because they can use n features with k values to describe kn [UNK] concepts. as we have seen throughout this book, both neural networks with multiple hidden units and probabilistic models with multiple latent variables make use of the strategy of distributed representation. we now introduce an additional observation. many deep learning algorithms are motivated by the assumption that the hidden units can learn to represent the underlying causal factors that explain the data, as discussed in section. distributed representations are natural for this approach, 15. 3 because each direction in representation space can correspond to the value of a [UNK] underlying configuration variable. an example of a distributed representation is a vector of n binary features, which can take 2n configurations, each potentially corresponding to a [UNK] region in input space, as illustrated in figure. this can be compared with 15. 7 a symbolic representation, where the input is associated with a single symbol or
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 561
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
which can take 2n configurations, each potentially corresponding to a [UNK] region in input space, as illustrated in figure. this can be compared with 15. 7 a symbolic representation, where the input is associated with a single symbol or category. if there are n symbols in the dictionary, one can imagine n feature detectors, each corresponding to the detection of the presence of the associated category. in that case only n [UNK] configurations of the representation space are possible, carving n [UNK] regions in input space, as illustrated in figure. 15. 8 such a symbolic representation is also called a one - hot representation, since it can be captured by a binary vector with n bits that are mutually exclusive ( only one of them can be active ). a symbolic representation is a specific example of the broader class of non - distributed representations, which are representations that may contain many entries but without significant meaningful separate control over each entry. examples of learning algorithms based on non - distributed representations include : • clustering methods, including the k - means algorithm : each input point is assigned to exactly one cluster. • k - nearest neighbors algorithms : one or a few templates or prototype examples are associated with a given input. in the
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 561
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
representations include : • clustering methods, including the k - means algorithm : each input point is assigned to exactly one cluster. • k - nearest neighbors algorithms : one or a few templates or prototype examples are associated with a given input. in the case of k > 1, there are multiple 546
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 561
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning h1 h2 h3 h = [ 1,, 1 1 ] h = [ 0,, 1 1 ] h = [ 1,, 0 1 ] h = [ 1,, 1 0 ] h = [ 0,, 1 0 ] h = [ 0,, 0 1 ] h = [ 1,, 0 0 ] figure 15. 7 : illustration of how a learning algorithm based on a distributed representation breaks up the input space into regions. in this example, there are three binary features h1, h2, and h3. each feature is defined by thresholding the output of a learned, linear transformation. each feature divides r2 into two half - planes. let h + i be the set of input points for which hi = 1 and h− i be the set of input points for which hi = 0. in this illustration, each line represents the decision boundary for onehi, with the corresponding arrow pointing to the h + i side of the boundary. the representation as a whole takes on a unique value at each possible intersection of these half - planes. for example, the representation value [ 1, 1, 1 ] corresponds to the region h + 1 ∩h + 2 ∩h + 3. compare
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 562
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
the representation as a whole takes on a unique value at each possible intersection of these half - planes. for example, the representation value [ 1, 1, 1 ] corresponds to the region h + 1 ∩h + 2 ∩h + 3. compare this to the non - distributed representations in figure. in the general case of 15. 8 d input dimensions, a distributed representation divides rd by intersecting half - spaces rather than half - planes. the distributed representation with n features assigns unique codes to o ( nd ) [UNK] regions, while the nearest neighbor algorithm withn examples assigns unique codes to only n regions. the distributed representation is thus able to distinguish exponentially many more regions than the non - distributed one. keep in mind that not allh values are feasible ( there is no h = 0 in this example ) and that a linear classifier on top of the distributed representation is not able to assign [UNK] class identities to every neighboring region ; even a deep linear - threshold network has a vc dimension of onlyo ( w w log ) where w is the number of weights (, ). the combination of a powerful representation sontag 1998 layer and a weak classifier layer can be a strong regularizer ; a classifier trying to learn the
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 562
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
( w w log ) where w is the number of weights (, ). the combination of a powerful representation sontag 1998 layer and a weak classifier layer can be a strong regularizer ; a classifier trying to learn the concept of “ person ” versus “ not a person ” does not need to assign a [UNK] class to an input represented as “ woman with glasses ” than it assigns to an input represented as “ man without glasses. ” this capacity constraint encourages each classifier to focus on few hi and encourages to learn to represent the classes in a linearly separable way. h 547
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 562
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning values describing each input, but they can not be controlled separately from each other, so this does not qualify as a true distributed representation. • decision trees : only one leaf ( and the nodes on the path from root to leaf ) is activated when an input is given. • gaussian mixtures and mixtures of experts : the templates ( cluster centers ) or experts are now associated with a degree of activation. as with the k - nearest neighbors algorithm, each input is represented with multiple values, but those values cannot readily be controlled separately from each other. • kernel machines with a gaussian kernel ( or other similarly local kernel ) : although the degree of activation of each “ support vector ” or template example is now continuous - valued, the same issue arises as with gaussian mixtures. • language or translation models based on n - grams. the set of contexts ( sequences of symbols ) is partitioned according to a tree structure of [UNK]. a leaf may correspond to the last two words being w1 and w2, for example. separate parameters are estimated for each leaf of the tree ( with some sharing being possible ). for some of these non - distributed algorithms, the output is not constant by parts but instead interpolates
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 563
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
words being w1 and w2, for example. separate parameters are estimated for each leaf of the tree ( with some sharing being possible ). for some of these non - distributed algorithms, the output is not constant by parts but instead interpolates between neighboring regions. the relationship between the number of parameters ( or examples ) and the number of regions they can define remains linear. an important related concept that distinguishes a distributed representation from a symbolic one is that generalization arises due to shared attributes between [UNK] concepts. as pure symbols, “ cat ” and “ dog ” are as far from each other as any other two symbols. however, if one associates them with a meaningful distributed representation, then many of the things that can be said about cats can generalize to dogs and vice - versa. for example, our distributed representation may contain entries such as “ has _ fur ” or “ number _ of _ legs ” that have the same value for the embedding of both “ cat ” and “ dog. ” neural language models that operate on distributed representations of words generalize much better than other models that operate directly on one - hot representations of words, as discussed in section. distributed representations induce a rich 12. 4 similarity space, in which semantically close concepts ( or
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 563
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
models that operate on distributed representations of words generalize much better than other models that operate directly on one - hot representations of words, as discussed in section. distributed representations induce a rich 12. 4 similarity space, in which semantically close concepts ( or inputs ) are close in distance, a property that is absent from purely symbolic representations. when and why can there be a statistical advantage from using a distributed representation as part of a learning algorithm? distributed representations can 548
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 563
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning figure 15. 8 : illustration of how the nearest neighbor algorithm breaks up the input space into [UNK] regions. the nearest neighbor algorithm provides an example of a learning algorithm based on a non - distributed representation. [UNK] non - distributed algorithms may have [UNK] geometry, but they typically break the input space into regions, with a separate set of parameters for each region. the advantage of a non - distributed approach is that, given enough parameters, it can fit the training set without solving a [UNK] optimization algorithm, because it is straightforward to choose a [UNK] output independently for each region. the disadvantage is that such non - distributed models generalize only locally via the smoothness prior, making it [UNK] to learn a complicated function with more peaks and troughs than the available number of examples. contrast this with a distributed representation, figure. 15. 7 549
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 564
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
chapter 15. representation learning have a statistical advantage when an apparently complicated structure can be compactly represented using a small number of parameters. some traditional non - distributed learning algorithms generalize only due to the smoothness assumption, which states that if u v ≈, then the target function f to be learned has the property that f ( u ) ≈f ( v ), in general. there are many ways of formalizing such an assumption, but the end result is that if we have an example ( x, y ) for which we know that f ( x ) ≈y, then we choose an estimator [UNK] that approximately satisfies these constraints while changing as little as possible when we move to a nearby input x +. this assumption is clearly very useful, but it [UNK] from the curse of dimensionality : in order to learn a target function that increases and decreases many times in many [UNK] regions, 1 we may need a number of examples that is at least as large as the number of distinguishable regions. one can think of each of these regions as a category or symbol : by having a separate degree of freedom for each symbol ( or region ), we can learn an arbitrary decoder mapping from symbol to value. however, this does not allow us to generalize
|
/home/ricoiban/GEMMA/mnlp_chatsplaining/RAG/Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org).pdf
| 565
|
Deep Learning by Ian Goodfellow, Yoshua Bengio, Aaron Courville (z-lib.org)
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.