text stringlengths 0 820 |
|---|
samples of the target domain. In this case, the data distributions remain unchanged, and the classifier is |
adapted to the target data distribution using strategies |
based on semisupervised learning. ◗The adaptation of the classifier by active learning (AL): This adaptation is performed by providing a lim - |
ited amount of well-chosen labeled samples from the target domain. This is a special case of the previous fam - |
ily, where we allow some new labeled examples to be |
sampled in the target domain to retrain the model itera - |
tively. Due to their acquisition cost, these samples need |
to be selected well, according to their potential to lead |
the model toward the desired target classifier. |
The rest of this section details recent advances in these |
four families. We will limit the discussion to approaches |
specific to the remote sensing literature and invite the in - |
terested reader to consult the specialized machine-learning |
and computer vision literature in [ 13], [14], and [16 ]. |
SELECTING INVARIANT FEATURESThe first family of DA methods is based on the selec - |
tion of invariant features that are usually a subset of the original set of features that are the most robust in |
the face of changes from the source to the target do - |
main. The main idea of the approach is to select features to reduce the difference between |
(, ) PX Ys and (, ). PX Yt An |
alternative strategy for encoding the invariance is based on |
the inclusion of additional synthetic labeled samples in |
the training set, a procedure known in machine learning |
as data augmentation . A meth - |
od adopting this strategy was |
studied in [ 29], where sam - |
ple-selection bias problems are addressed by enriching the training set with artificial |
examples that correspond to |
physically consistent varia - |
tions of the training samples |
(e.g., illumination, size, and |
rotation). To limit the num - |
ber of additional examples to be used by the support vector |
machine (SVM), variations are generated only for the train - |
ing samples considered as support vectors by the classifier |
trained on the source domain only. |
Let us consider the first strategy and focus on the analy- |
sis of hyperspectral images as an application of particular interest. Hyperspectral sensors are capable of capturing |
hundreds of narrow spectral bands from a wide range of |
the electromagnetic spectrum, which is why they are par - |
ticularly sensitive to subtle changes in image-acquisition |
conditions, leading to a nonstationary behavior of the spec - |
tral signature of the classes and, therefore, to problems that |
should be solved by transfer learning and DA approaches. |
Figure 4 shows an example of a shift in the signature of a |
hyperspectral image acquired by the Hyperion sensor over |
two areas of the Okavango Delta in Botswana. |
In [30], the authors propose an approach for selecting |
subsets of features that are both discriminative of the land-cover classes and invariant between the source and the target tHe coVaRiate SH iFt iS |
a pa RticuLaR ca Se oF |
Samp Le SeLection BiaS |
wHe Re tHe BiaS DepenDS |
onLY on t He input |
VaRiaBL e X (anD not |
on y). |
Authorized licensed use limited to: ASU Library. Downloaded on March 08,2024 at 03:13:37 UTC from IEEE Xplore. Restrictions apply. |
ieee Geoscience and remote sensin G ma Gazine june 201646 |
domain. The main idea of this |
approach is to explicitly con - |
sider two distinct terms in the criterion function for evaluat - |
ing both the discrimination |
capability |
D of the feature |
subset and the data set shift |
P of the features between the |
source and target domain. The first term is standard in filter |
methods for feature selection and provides high scores when |
the features selected show |
some kind of dependency with the desired output (e.g., the |
classes to be predicted). The |
second term has been introduced to evaluate the invariance |
of the feature subset between the two domains. The subset of |
features |
F is selected by jointly optimizing the two terms D |
and ,P i.e., by solving the following multiobjective optimiza - |
tion problem: |
(( ),()), argmin FP F |
||FlD- |
= (1)where l is the size of the feature subset. Both D and P |
are treated as functions of the subset of considered fea - |
tures .F The specific definitions of the terms D and P |
are reported in [ 30], considering their parametric esti - |
mation (assuming Gaussian distribution of the classes) |
in both the supervised and semisupervised DA settings. In [31], the two terms are defined considering kernel- |
based dependence estimators and kernel embedding of conditional distributions, resulting in a nonparamet - |
ric approach that does not require the estimation of the |
class distributions as an intermediate step. Equation (1) |
is solved by adopting a genetic multiobjective optimiza - |
tion algorithm. The solution results in features with high |
capability to discriminate classes (with a small value of |
)D- and high stability on the two domains (with a small |
data set shift ).P Adopting a multiobjective optimization |
approach instead of considering a linear combination of |
the two terms frees the user from specifying in advance |
the relative importance of the two terms D and .P The |
solution of the multiobjective problem allows one to find the solutions that represent the best tradeoffs of discrimi - |
native and stable feature subsets for the specific transfer- |
learning problem at hand.0 50 100 15001,0002,0003,0004,0005,0006,0007,000 |
Class 1 |
Class 2 |
Class 3 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.