text stringlengths 0 820 |
|---|
Many problems can be viewed as forms of geospatial |
search aided by aerial imagery, with examples ranging from |
detecting poaching activity to human trafficking. We model |
this class of problems in a visual active search (VAS) frame- |
work, which has three key inputs: (1) an image of the entire |
search area, which is subdivided into regions, (2) a local |
search function, which determines whether a previously un- |
seen object class is present in a given region, and (3) a |
fixed search budget, which limits the number of times the |
local search function can be evaluated. The goal is to max- |
imize the number of objects found within the search budget. |
We propose a reinforcement learning approach for VAS that |
learns a meta-search policy from a collection of fully anno- |
tated search tasks. This meta-search policy is then used to |
dynamically search for a novel target-object class, lever- |
aging the outcome of any previous queries to determine |
where to query next. Through extensive experiments on sev- |
eral large-scale satellite imagery datasets, we show that the |
proposed approach significantly outperforms several strong |
baselines. We also propose novel domain adaptation tech- |
niques that improve the policy at decision time when there |
is a significant domain gap with the training data. Code is |
publicly available at this link. |
1. Introduction |
Consider a large national park that hosts endangered an- |
imals, which are also in high demand on a black market, |
creating a major poaching problem. An important strategy |
in an anti-poaching portfolio is to obtain aerial imagery us- |
ing drones that helps detect poaching activity, either ongo- |
ing, or in the form of traps laid on the ground [1, 2, 3, 9, 8]. |
The quality of the resulting photographs, however, is gen- |
erally somewhat poor, making the detection problem ex- |
tremely difficult. Moreover, park rangers can only inspect |
relatively few small regions to confirm poaching activity, |
doing so sequentially. Crucially, inspecting such regionsyields new ground truth information about poaching activ- |
ity that we can use to decide which regions to inspect in |
the future. We can distill some key generalizable structure |
Figure 1: A comparison of a greedy search policy (dashed line) |
with an active search strategy (solid line) for the small car tar- |
get class. The greedy policy is not able to adapt when a car is |
not found in the starting cell and needlessly searches many similar |
cells. The active strategy adapts and explores regions with differ- |
ent visual characteristics, eventually finding the objects of interest. |
from this scenario: given a broad area image (often with |
a relatively low resolution), sequentially query small areas |
within it (e.g., by sending park rangers to the associated re- |
gions, on the ground), with each query returning the ground |
truth, to identify as many target objects (e.g., poachers or |
traps) as possible. The number of queries we can make is |
typically limited, for example, by budget or resource con- |
straints. Moreover, query results (e.g., detected poaching |
activity in a particular region) are highly informative about |
the locations of target objects in other regions , for example, |
due to spatial correlation. We refer to this general modeling |
framework as visual active search (VAS) . Numerous other |
scenarios share this broad structure, such as identification |
of drug or human trafficking sites, broad area search-and- |
rescue, identifying landmarks, and many others. |
A simple solution to the broad-area search problem is to |
divide the broad area into many grid cells, train a classifier |
1arXiv:2211.15788v3 [cs.CV] 29 Oct 2023 |
to predict existence of a target object in each grid cell, and |
simply explore the top Kcells in terms of predicted likeli- |
hood of the object being in the cell. We call this the greedy |
policy, which essentially reduces geospatial active search to |
the familiar object identification (or detection) problem. In |
Figure 1, we offer some intuition about why this idea fails |
to capture important problem structure. Suppose we look |
for small cars in an image, starting in the grid marked start, |
which we initially think is the best place to look. The greedy |
policy being a one-shot predictor of likely grid cells con- |
taining target, continues to explore similar regions (marked |
as). What this approach ignores, as does framing the |
problem as traditional one-shot object identification, is the |
fact that both success and failure of past queries are infor- |
mative due to complex spatial correlation among objects |
and other patterns in the scene; here, because the car was |
not found, we proceed to instead explore regions that have |
somewhat different characteristics. The key to visual active |
search, therefore, is to learn how to make use of the ground |
truth information obtained over a sequence of queries to de- |
cide where to query next. Additionally, Section 5 of [11] |
provides a rigorous analysis of the general sub-optimality |
of greedy policies in active search settings. |
Relationship to Active Search and Active Learning VAS |
is closely related to active search [11, 10, 15, 13]. Ac- |
tive search is typically concerned with binary classification, |
and aims to maximize the number of discovered positively- |
labeled inputs. It uses a function f, which predicts labels of |
inputs x, as a means to this end, with each query serving the |
dual-purpose of improving fas well as identifying a posi- |
tive instance. A central concern in active search, therefore, |
is achieving a balance of exploration (learning f) and ex- |
ploitation (identifying target inputs). This consideration is |
also the key distinction between active search and active |
learning [25], which is concerned solely with improving |
the predictive quality of f. Thus, if we had only a single |
query, active learning would typically choose xfor which |
prediction is highly uncertain, whereas active search would |
choose xfor which f(x)is most confidently positive. We |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.