weipang142857's picture
Upload folder using huggingface_hub
f6cc031 verified
<Poster Width="1734" Height="1214">
<Panel left="36" right="192" width="398" height="289">
<Text>Task</Text>
<Text></Text>
<Figure left="100" right="244" width="91" height="89" no="1" OriWidth="0" OriHeight="0
" />
<Text> Fully annotated</Text>
<Figure left="264" right="243" width="94" height="91" no="2" OriWidth="0" OriHeight="0" />
<Text> Weakly annotated</Text>
<Text> Many computer vision tasks require fully annotated data, but</Text>
<Text> Time-consuming, Laborious, Human various</Text>
<Text> More and more online media sharing websites (e.g. Flickr) provide</Text>
<Text>weakly annotated data, However,</Text>
<Text> Weaker supervision, Ambiguity (background clutter, occlusion…)</Text>
<Text> Challenge: Weakly Supervised Object Localisation (WSOL).</Text>
</Panel>
<Panel left="36" right="482" width="397" height="696">
<Text>Existing Approaches vs. Ours</Text>
<Text>Three types of cues are exploited in existing WSOL:</Text>
<Text> Object-saliency: A region containing the object should look different</Text>
<Text>from background in general.</Text>
<Text> Intra-class: The region should look similar to other regions containing</Text>
<Text>the object of interest in other training images.</Text>
<Text> Inter-class: The region should look dissimilar to any regions that are</Text>
<Text>known to not contain the object of interest.</Text>
<Text>However, they are independently trained:</Text>
<Figure left="39" right="671" width="396" height="273" no="3" OriWidth="0.390751" OriHeight="0.189455
" />
<Text>Independent learning ignores the fact that:</Text>
<Text> The knowledge that multiple objects co-exist within each image is</Text>
<Text>not exploited.</Text>
<Text> The background is relevant to different foreground object classes.</Text>
<Text>Our contributions:</Text>
<Text> We propose the novel concept of joint modelling of all object</Text>
<Text>classes and backgrounds for weakly supervised object localisation.</Text>
<Text> We formulate a novel Bayesian topic model suitable for localization</Text>
<Text>of objects and utilizing various types of prior knowledge available.</Text>
<Text> We provide a solution for exploiting unlabeled data for semi+weakly</Text>
<Text>supervised learning of object localisation.</Text>
</Panel>
<Panel left="455" right="194" width="820" height="407">
<Text>Methodology</Text>
<Text>Preprocessing and Representation:</Text>
<Text> Regular Grid SIFT Descriptors. Sampled every 5 pixels.</Text>
<Text> Quantising using 𝑁𝑣 = 2000 word codebook.</Text>
<Text> Words and corresponding locations:</Text>
<Text>Our Model:</Text>
<Figure left="460" right="354" width="401" height="239" no="4" OriWidth="0.363584" OriHeight="0.146113
" />
<Text>Observed variables:</Text>
<Text>𝐽 Ο = {𝑥𝑗 , 𝑙𝑗 }</Text>
<Text>𝑗=1Low-level feature words and corresponding location</Text>
<Text>Latent variables:</Text>
<Text> H=For each topic k and image j𝐾,𝐽{{𝜋𝑘 }𝐾</Text>
<Text>𝑘=1 , {𝑦𝑗 , 𝜇𝑘𝑗 , Λ 𝑘𝑗 , 𝜃𝑗 }</Text>
<Text>𝑘=1,𝑗=1 }</Text>
<Text>Given parameters:</Text>
<Text>Label information and prior𝐽, {𝛼𝑗 }</Text>
<Text>𝑗=1𝑘=1𝐾</Text>
<Text>𝜋𝑘0 , 𝜇𝑘0 , Λ0</Text>
<Text>𝑘 , 𝛽𝑘0 , 𝜈𝑘0 Π=</Text>
<Text>Joint distribution:</Text>
<Text>𝑝 𝑥𝑖𝑗 𝑦𝑖𝑗 , 𝜃𝑗 𝑝 𝑦𝑖𝑗 𝜃𝑗</Text>
<Text>𝑘𝑝(𝜋𝑘 |𝜋𝑘0 )</Text>
<Text>𝑖𝑗𝑗</Text>
<Text>𝑝 𝜇𝑗𝑘 , Λ𝑗𝑘 𝜇𝑘0 , Λ0</Text>
<Text>𝑘 , 𝛽𝑘0 , 𝜈𝑘0 )𝑝(𝜃𝑗 |𝛼𝑗 )𝑝 𝑂, 𝐻 Π =𝑝 𝑥𝑖𝑗 𝑦𝑖𝑗 , 𝜃𝑗 𝑝 𝑦𝑖𝑗 𝜃𝑗</Text>
<Text>𝑘𝑝(𝜋𝑘 |𝜋𝑘0 )</Text>
<Text>𝑖𝑗𝑁𝑗</Text>
<Text>𝑝 𝜇𝑗𝑘 , Λ𝑗𝑘 𝜇𝑘0 , Λ0</Text>
<Text>𝑘 , 𝛽𝑘0 , 𝜈𝑘0 )𝑝(𝜃𝑗 |𝛼𝑗 )𝑝 𝑂, 𝐻 Π =𝐾𝐽</Text>
<Text>Prior Knowledge:</Text>
<Text> Human knowledge objects and their relationships with backgrounds</Text>
<Text> Objects are compact whilst background spread across the image.</Text>
<Text> Objects stand out against background.</Text>
<Text> Transferred knowledge</Text>
<Text> Appearance and Geometry information from existing dataset.</Text>
<Text>Object Localisation:</Text>
<Text> Our-Gaussian Aligning a window to the ellipse obtained from q 𝜇, Λ</Text>
<Text> Our-Sampling Non-maximum suppression sampling over heat-map</Text>
</Panel>
<Panel left="455" right="602" width="818" height="203">
<Text>Results</Text>
<Text>Dataset: PASCAL VOC 2007. Three variants are used:</Text>
<Text> VOC07-6×2 : 6 classes with Left and Right poses, 12 classes in total.</Text>
<Text> VOC07-14: 14 classes, other 6 were used as annotated auxiliary data</Text>
<Text> VOC07-20: all 20 classes, each class contain all pose data.</Text>
<Text>PASCAL criterion:</Text>
<Text> intersection-over-union > 0.5 between Ground-Truth and predicted box</Text>
<Text>Comparison with state-of-the-art</Text>
<Text> Initialisation: Localising object of interest in weakly labelled images.</Text>
<Text> Refined by detector: A conventional object detector can be trained</Text>
<Text>using initial annotation. Then it can be used to refine object location.</Text>
<Figure left="874" right="634" width="403" height="175" no="5" OriWidth="0.375145" OriHeight="0.152368
" />
</Panel>
<Panel left="455" right="805" width="819" height="372">
<Text>Example: Foreground Topics</Text>
<Text></Text>
<Figure left="458" right="840" width="818" height="212" no="6" OriWidth="0.775145" OriHeight="0.16622
" />
<Text>Figs. (c) and (d) illustrate that the object of interest “explain away”</Text>
<Text>other objects of no interest.</Text>
<Text> A car is successfully located in Fig. (c) using the heat map of car topic.</Text>
<Text>Fig. (d) shows that the motorbike heat map is quite accurately</Text>
<Text>selective, with minimal response obtained on the other vehicular clutter.</Text>
<Text>Fig. (e) indicates how the Gaussian can sometimes give a better location.</Text>
<Text>Fig. (f) shows that the single Gaussian assumption is not ideal when the</Text>
<Text>foreground topic has a less compact response.</Text>
<Text>A failure case is shown in Fig. (g), where a bridge structure resembles</Text>
<Text>the boat in Fig (a) resulting strong response from the foreground topic,</Text>
<Text>whilst the actual boat topic is small and overwhelmed.</Text>
</Panel>
<Panel left="1294" right="193" width="400" height="512">
<Text>Example: Background Topics</Text>
<Text></Text>
<Figure left="1308" right="229" width="378" height="336" no="7" OriWidth="0.327168" OriHeight="0.257373
" />
<Text>Background non-annotated data has been modelled in our framework.</Text>
<Text>Irrelevant pixels will be explained to reduce confusion with object.</Text>
<Text>Automatically learned background topics have clear semantic meanings,</Text>
<Text>corresponding to common components as shown in the Figure.</Text>
<Text> Some background components are mixed, e.g. the water topic gives</Text>
<Text>strong response to both water and sky. But this is understandable</Text>
<Text>since water and sky are almost visually indistinguishable in the image.</Text>
</Panel>
<Panel left="1295" right="705" width="399" height="292">
<Text>Example: Semi-supervised Learning</Text>
<Figure left="1303" right="745" width="396" height="139" no="8" OriWidth="0.37341" OriHeight="0.101877
" />
<Text>𝑓𝑔 Unknown image can set as 𝛼</Text>
<Text>𝑗 =0.1. (soft constraint)</Text>
<Text> 10% labelled data + 90% unlabeled data (relevant) or unrelated data</Text>
<Text> Evaluating on (1) initially annotated 10% data (standard WSOL).</Text>
<Text>(2) testing part dataset (localize objects in new images</Text>
<Text> The figure clearly shows unlabeled data helps to learn a better object</Text>
<Text>model.</Text>
</Panel>
<Panel left="1296" right="998" width="397" height="180">
<Text>References</Text>
<Text>[1] T. Deselaers, B. Alexe, and V. Ferrari. Weakly supervised localization</Text>
<Text>and learning with generic knowledge. IJCV. 2012.</Text>
<Text>[2] M. Pandey and S. Lazebnik. Scene recognition and weakly supervised</Text>
<Text>object localization with deformable part-based models. In ICCV, 2011</Text>
<Text>[3] P. Siva and T. Xiang. Weakly supervised object detector learning with</Text>
<Text>model drift detection. In ICCV, 2011.</Text>
<Text>[4] P. Siva, C. Russell, and T. Xiang. In defence of negative mining for</Text>
<Text>annotating weakly labelled data. In ECCV, 2012.</Text>
</Panel>
</Poster>