text stringlengths 0 820 |
|---|
posed in recent remote sensing literature and discuss their |
strengths and weaknesses. We also provide a series of prac - |
tical examples about the use of DA in high- to very high-resolution image-processing tasks. However, we will not enter into the technical details of specific DA literature; for |
interested readers, we refer to the recent surveys published |
in [13], [14], [16], and [17 ]. |
tRanSFeR LeaRning anD Domain aDaptation |
Transfer-learning problems arise when inferences have to be made on processes that are not stationary over time or |
space. As previously discussed, this is the case in the anal - |
ysis of remote sensing images where different acquisitions |
are typically subject to different conditions (e.g., illumi - |
nation, viewing angle, soil moisture, and topography). Such differences can affect the observed spectral signa - |
tures of the land-cover types and, therefore, the distribu - |
tion of the information classes in the feature space [ 18]. |
Different transfer-learning problems (and the techniques to tackle them) have been considered in the literature, including DA, multitask learning, domain generaliza - |
tion, sample selection bias, and covariate shift [ 14]. In |
this article, we will focus on DA, which is a particular form of transfer learning. |
Let us consider two domains, called source domain and |
target domain , that are associated with two images ac - |
quired on different geographical areas (but with similar |
land-cover characteristics) or on the same area at differ - |
ent time instants. Figure 2 shows the DA problem in the |
context of remote sensing image classification. The source |
and target domains are associated with the joint probabil - |
ity distributions |
(, ) PX Ys and (, ), PX Yt respectively. The |
two joint probabilities define the classification problems |
on the two domains, where |
X is the input (vector) vari - |
able (i.e., spectral bands of |
the source image with pos - |
sible additional features used |
to characterize the contex - |
tual information of the single pixel) and |
Y is the output |
variable associated with a set |
of classes (i.e., land-cover or |
land-use information). The aim of DA methods is to adapt a classifier trained on the source domain to make predictions |
on the target domain. |
Supervised DA assumes that labeled samples are availa ble |
for both domains. The labeled sets |
xx{( ,),,(, )} Ty ys |
nn 11f = |
and xx{( ,),,(, )} Ty yt |
mm 11f = are the source- and target- |
domain training sets, respectively. Supervised DA meth - |
ods focus on challenging situations where labeled target- |
domain samples are less numerous than those available in |
the source domain (i.e., ). mn11 In such conditions, the |
proper use of source-domain information is very important |
in solving the target problem. Most of the work in DA as - |
sumes that source and target domains share the same set Source Image Target Image |
DA |
Ps(X,Y ) = Ps(X)Ps(Y|X) Pt(X,Y ) = Pt(X)Pt(Y|X) ≠ |
Figu Re 2. A graphical representation of the DA problem in the |
context of remote sensing image classification. Source and target images can be acquired on different geographical areas (but with |
similar land-cover characteristics) or on the same area at different times. The two images are associated with two different joint dis - |
tributions, which characterize the two classification problems. The two distributions can differ due to different acquisition conditions (e.g., illumination, viewing angle, soil moisture, and topography). |
tRanSFeR-LeaRning |
pRoBLemS aRiSe wHen |
inFeRence S HaVe to B e |
maDe on p RoceSSeS tHat |
aRe not Stationa RY o VeR |
time o R Space . |
Authorized licensed use limited to: ASU Library. Downloaded on March 08,2024 at 03:13:37 UTC from IEEE Xplore. Restrictions apply. |
ieee Geoscience and remote sensin G ma Gazine june 201644 |
of classes. There are only a few papers that address DA con - |
sidering differences in the set of classes between source and |
target [ 19]–[22 ]. |
Semisupervised DA methods assume that a training set |
is available only for the source domain, whereas target-do - |
main information is limited to a set of unlabeled samples |
xx{, ,} . Ut |
m 1f = This DA setting is more challenging than |
the supervised case and requires important assumptions on |
the relationship between source and target domains to make |
the algorithm converge to a consistent solution on the target |
domain. All DA methods are based on the assumption that |
(, ) PX Ys and (, ) PX Yt are different but close enough to en - |
sure that the source-domain |
information can be of help |
for solving the target-domain learning problem. On the one |
hand, if the source and target |
domain are arbitrarily differ - |
ent, there is no hope that the |
source-domain information |
will provide an advantage in solving the task in the target |
domain. On the other hand, |
if |
(, )( ,) , PX YP XYst= no ad - |
aptation is necessary, and the |
model trained on the source |
can be readily applied to the |
target. Semisupervised DA methods are effective in situa - |
tions that lie in between these two extreme cases.Unsupervised DA methods are the last family, and they |
assume that two unlabeled domains have to be matched. This is the most difficult case because label information is not available for any domain. In this situation, DA methods |
aim to match the marginal distributions of the two domains |
()PXs and ()PXt without knowledge on the learning task |
(classification or regression). Unsupervised methods can be used as preprocessing of any analysis task (e.g., clustering or |
density estimation), but they imperatively need to have data sets with similar structural properties before adaptation. |
Unsupervised DA models are generally feature extractors or |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.