text
stringlengths 0
820
|
|---|
token represents the masked patches that were not encoded.
|
Another positional encoding vector is added to all patches
|
and a sequence of transformer blocks decodes these patches
|
to form the original input image, which is used as the learning
|
target.
|
Input Scale-MAE performs a super resolution reconstruc-
|
tion, where the input image Iis downsampled from a higher
|
resolution image Ihrat the ground truth GSD. Instead of
|
targeting the input image, Scale-MAE targets high frequency
|
and low frequency components of Ihr, which is common in
|
Laplacian pyramid super resolution models [64], where the
|
high frequency component is at the same resolution as the
|
ground truth image Ihrand the low frequency component
|
is at the same resolution as the input image I, as shown in
|
Figure 2. Following many works in super resolution [64], the
|
low frequency target image is obtained by interpolating Ihr
|
to a much lower resolution, rlowand then interpolating to the
|
same resolution as the input image I. The high frequency tar-
|
get image is obtained by downsampling Ihrto another lower
|
resolution rhigh-low , and then upsampling to the same resolu-
|
tion as the ground truth image Ihrand subtracting this image
|
Ihf=Ihr−Ihigh-low . The supplementary material provide
|
more information on the upsampling/downsampling method-
|
ology. The key components for Scale-MAE are described
|
next.
|
GSD Positional Encoding Images from scale-dependent
|
domains have a metric which defines the absolute scale for
|
the image. This metric has different names across domains
|
and is referred to as the Ground Sample Distance (GSD) in
|
remote sensing. The GSD is critical to understanding, con-
|
ceptually, the kinds of features that will be available in an
|
image. An image with finer GSD (lower number) will have
|
higher frequency details than an image with coarser GSD
|
(high number). Models are generally unaware of absolute
|
scale when learning over a set of data. Specifically, even if
|
they implicitly learn that all images in a dataset share a vary-
|
ing resolution from input-space augmentations, then these
|
models do not explicitly condition on the GSDs encountered
|
in unseen data.
|
We extend the positional encoding from Equation (2) to
|
include GSD by scaling the positional encoding relative to
|
the land area covered in an image as depicted in Figure 3
|
and mathematically:
|
vgsd,x(pos,2i) = sing
|
Gpos
|
100002i
|
D(3)
|
vgsd,y(pos,2i+ 1) = cosg
|
Gpos
|
100002i
|
D(4)
|
Figure 3. Ground Sample Distance Positional Encoding (GS-
|
DPE). (Left) Input images at the same pixel resolution but different
|
GSDs are shown. The image on the bottom is a subset of the image
|
on the top. (Center) This overlap in location, albeit at a different
|
resolution, is reflected in the GSDPE. The finer image with smaller
|
spatial extent is represented by a corresponding subsection of the
|
overall sine wave on the bottom. (Right) A standard positional
|
encoding is strictly dependent on the image resolution and uses the
|
same embedding for both. The colors behind the sine waves show
|
the intensity and quantization of the encoding.
|
where gis the GSD of the image and Gis a reference GSD,
|
nominally set to 1m. Intuitively, an object imaged at a finer
|
resolution has more pixels representing it. When imaging the
|
same object at a coarser resolution, those pixels must map to
|
fewer pixels. In Equation (4), we interpolate the positional
|
encoding by a factor ofG
|
gto account for the ordering of the
|
coarser set of pixels. This simple idea underpins the GSD
|
Positional Encoding, visualized in Figure 3.
|
Scale-MAE decoder The standard MAE learns represen-
|
tations by tasking a network with reconstructing an image
|
after masking out most of its pixels. While the standard
|
MAE decoder reconstructs the input image at the same scale
|
as its input, the objective of Scale-MAE is to learn multi-
|
scale representations. We draw on works from progressive
|
super-resolution such as [56], that learn a high resolution,
|
high frequency image and a lower resolution low frequency
|
image, that when combined together, yield the input image
|
at a higher resolution.
|
The Scale-MAE introduces a novel decoder which de-
|
codes to multiple scales with a progressive Laplacian de-
|
coder architecture, replacing the traditional MAE “decoder”,
|
which is really a Transfomer encoder. This architecture
|
consists of three stages: decoding, upsampling, and recon-
|
struction, which are shown in Figure 2 and detailed below.
|
Decoding follows the standard MAE decoder where fol-
|
lowing the encoder, the removed mpatches are then placed
|
back into their original location in the sequence of patches
|
where a learned mask token represents the masked patches
|
that were not encoded, a positional encoding is added, and
|
then a series of transformer layers decode all patches. In
|
contrast to the standard MAE decoder, the Scale-MAE de-
|
coder uses fewer transformer layers (e.g. 3 layers instead of
|
8), which reduces the parameter complexity as quantified
|
in Section 5. The output of these layers is then fed into the
|
upsampling stage.
|
Upsampling The latent feature maps from the decoding
|
stage are progressively upsampled to 2x and 4x resolution
|
using deconvolution blocks, where the first deconvolution
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.