Buckets:
Title: LoCoNet: Long-Short Context Network for Active Speaker Detection
URL Source: https://arxiv.org/html/2301.08237
Published Time: Thu, 02 May 2024 20:06:49 GMT
Markdown Content: Xizi Wang 1 Feng Cheng 2 Gedas Bertasius 2 David Crandall 1
1 Indiana University 2 UNC Chapel Hill
xiziwang@iu.edu {fengchan,gedas}@cs.unc.edu
Abstract
Active Speaker Detection (ASD) aims to identify who is speaking in each frame of a video. Solving ASD involves using audio and visual information in two complementary contexts: long-term intra-speaker context models the temporal dependencies of the same speaker, while short-term inter-speaker context models the interactions of speakers in the same scene. Motivated by these observations, we propose LoCoNet, a simple but effective Long-Short Context Network that leverages Long-term Intra-speaker Modeling (LIM) and Short-term Inter-speaker Modeling (SIM) in an interleaved manner. LIM employs self-attention for long-range temporal dependencies modeling and cross-attention for audio-visual interactions modeling. SIM incorporates convolutional blocks that capture local patterns for short-term inter-speaker context. Experiments show that LoCoNet achieves state-of-the-art performance on multiple datasets, with 95.2% (+0.3%) mAP on AVA-ActiveSpeaker, 97.2% (+2.7%) mAP on Talkies, and 68.4% (+7.7%) mAP on Ego4D. Moreover, in challenging cases where multiple speakers are present, LoCoNet outperforms previous state-of-the-art methods by 3.0% mAP on AVA-ActiveSpeaker. The code is available at https://github.com/SJTUwxz/LoCoNet_ASD.
1 Introduction
Real-world interactive computer vision systems need to recognize not only the physical properties of a scene, such as objects and people, but also the social properties, including how people interact with each other. One fundamental task is identifying, at any moment, who is speaking in a complex scene with multiple interacting individuals. This Active Speaker Detection (ASD) problem[2, 39, 65, 24, 42, 3, 58, 11, 35, 32, 25, 40, 29, 18, 53, 59, 62, 70] is crucial for many real-world applications like human-robot interaction[66, 33, 61], speech diarization[20, 68, 73, 21, 16, 14], video re-targeting[6, 44, 74], multimodal learning[7, 60, 19, 34, 28, 55, 4, 63, 30, 50, 46, 17, 34, 49, 50, 54, 55], etc.
Figure 1: Comparison of ASD methods in terms of mean average precision (mAP) on the AVA-ActiveSpeaker dataset, average-FLOPs, and number of parameters. Note that average-FLOPs is the computation required to predict the speaking activity of one face crop.
How can we tell whether someone is speaking? Visual cues such as movements of the mouth and eyes, when correlated with audio signals, often offer direct and primary evidence. The inter-modality synchronization over a longer audio-visual segment also provides complementary information[65, 2, 48, 39]. The first row of Fig2 shows how Long-term Intra-speaker Modeling helps discern this primary indicator by observing one person for a long time span.
Figure 2: Long-term Intra-speaker Modeling (LIM), Short-term Inter-speaker Modeling (SIM), and comparison of LoCoNet with existing long-term parallel-inference ASD methods. Red boxes show inactive speakers and green boxes show active speakers. LIM uses the features of a single speaker across all frames to capture long-term relationships. SIM models the relationships of speakers within a short m ๐ m italic_m-frame segment to capture the conversation pattern. The speaker context modeling of the existing long-term parallel-inference ASD methods[65, 45, 15] only focuses on LIM, while LoCoNet models LIM and SIM to learn both contexts.
However, in a complex video, a personโs face is often occluded, turned away, off-frame, or very small, posing challenges for directly inferring speaking activity from visual cues. Fortunately, valuable evidence about the target speaker can be gleaned from the behaviors of others in the scene[61]. The second row of Fig2 shows an example: from the left m ๐ m italic_m-frame video segment, even with a partially visible face of the man on the right, it is easy to see that he is speaking at T i+m subscript ๐ ๐ ๐ T_{i+m}italic_T start_POSTSUBSCRIPT italic_i + italic_m end_POSTSUBSCRIPT because the woman in the middle turns her head (T i+1โT i+mโsubscript ๐ ๐ 1 subscript ๐ ๐ ๐ T_{i+1}\to T_{i+m}italic_T start_POSTSUBSCRIPT italic_i + 1 end_POSTSUBSCRIPT โ italic_T start_POSTSUBSCRIPT italic_i + italic_m end_POSTSUBSCRIPT) and neither of the other two people open their mouths at T i+m subscript ๐ ๐ ๐ T_{i+m}italic_T start_POSTSUBSCRIPT italic_i + italic_m end_POSTSUBSCRIPT. Notably, the womanโs gaze towards the man on the right at T i+m subscript ๐ ๐ ๐ T_{i+m}italic_T start_POSTSUBSCRIPT italic_i + italic_m end_POSTSUBSCRIPT does not provide substantial information on whether the man is speaking at a distant time T j+1 subscript ๐ ๐ 1 T_{j+1}italic_T start_POSTSUBSCRIPT italic_j + 1 end_POSTSUBSCRIPT. Therefore, we argue that Inter-speaker Modeling is sufficient in short-term temporal windows, since activities of speakers within a short time range are more correlated than speakers separated farther away in time. Cognitive research[57, 43] also suggests that speaker-listener coupling is more coordinated in nearby frames.
To solve this per-frame video classification task, existing ASD methods can be put into two categories: 1) parallel-inference methods[65, 45, 15, 75, 40] take all frames as input and predict the results for all frames in one pass. 2) sequential-inference methods[39, 2, 3, 48, 42] take a short clip as input and only give prediction result of the center frame. Thus, a sliding window strategy is often adopted to get results for all the frames. As depicted in Fig.1, the average-FLOPs of parallel-inference methods[65, 15, 45] to predict the speaking activity of one face crop are often much lower than sequential-inference methods[39, 2, 42, 3, 48]. However, most parallel-inference methods[65, 15, 45, 72] that take long video clips as input do not consider multi-speaker context, which could cause performance degradation as speakersโ interaction is crucial for ASD task.
With the above issues in mind, we propose LoCoNet, an end-to-end Long-Short Context Network. Long-term Intra-speaker Modeling (LIM) employs a self-attention mechanism[67] for long-range dependencies modeling and a cross-attention mechanism for audio-visual interactions. Short-term Inter-speaker Modeling (SIM) incorporates convolutional blocks to capture local conversational patterns. Moreover, while most ASD methods use vision backbones for audio encoding[2, 65, 42, 3] due to high temporal downsampling in most audio backbones[38, 27, 56, 12, 23], we propose VGGFrame to leverage pretrained AudioSet[22] weights to extract per-frame audio features. We also use a parallel inference strategy for more efficient video processing.
Our extensive experiments validate the effectiveness of our approach. On the AVA-ActiveSpeaker dataset[58], LoCoNet achieves 95.2% mAP, outperforming previous state-of-the-art method SPELL+[48] by 0.3% with 38ร38\times 38 ร less computational cost. Furthermore, LoCoNet achieves 97.2% (+2.7%) mAP on Talkies[42] and 68.4% (+7.7%) mAP on Ego4Dโs Audio-Visual benchmark[24]. LoCoNet works especially well in challenging scenarios such as with multiple speakers or small speaker faces.
2 Related Work
Most recent techniques for ASD can be characterized in terms of three salient dimensions: frame-level processing strategy that determines the inference speed of the method, the extracted context information to enhance the feature representations for prediction, and the training mechanisms.
Frame-level processing strategy. Given a long video, existing ASD methods employ two main strategies for generating per-frame predictions. 1) Parallel-inference[65, 45, 15, 40] takes all frames of the video as input and predict the per-frame results in one pass. Such methods are often fast, as depicted in Fig.1. However, they typically do not consider interactions among multiple speakers. 2) Sequential-inference[39, 2, 42, 3, 48] predicts one frame by sampling a short clip centered around that frame. They need a sliding window strategy to produce predictions for all frames, resulting in slower inference speed. Our proposed LoCoNet adopts a parallel inference strategy, combining fast inference speed with effective consideration of speakersโ interactions.
Context Modeling. ASD benefits from both intra-speaker and inter-speaker contexts. TalkNet[65] models long-term temporal intra-speaker context to distinguish speaking and non-speaking frames. ASC[2] employs self-attention for long-term inter-/intra-speaker modeling, and an LSTM for long-term temporal refinement. ASDNet[39] aggregates short-term features of target speaker and background speakers at nearby frames, and leverages Bidirectional GRU[13] for long-term temporal modeling. Light-ASD[45] also uses Bidirectional GRU for temporal modeling. MAAS[42], EASEE[3], and SPELL[48] use Graph Convolutional Networks[69, 37] to model relationships between the visual nodes and audio nodes of context speakers. TS-TalkNet[31] explores the use of reference speech to assist ASD.
In our proposed LoCoNet, we introduce Long-Short Context Modeling (LSCM), composed of Long-term Intra-speaker Modeling (LIM) and Short-term Inter-speaker Modeling (SIM) in an interleaved manner. LIM captures long-term temporal dependencies with self-attention, and audio-visual interactions with cross-attention. SIM incorporates inter-speaker convolutional blocks to learn speakersโ interactions.
Training mechanisms. Training on long videos can be memory-intensive, prompting some prior work[2, 39, 42, 48] to adopt a multi-stage training mechanism. In this process, a short-term feature extractor is initially trained, and subsequently, a long-term context modeling network is trained on the features extracted by the pre-trained feature extractor. TalkNet[65], EASEE[3], UniCon[75], light-ASD[45], ASD-transformer[15], and ADENet[72] use end-to-end training, fully leveraging the learning capabilities of the model. Similarly, our proposed LoCoNet is also trained end-to-end, enabling joint optimization of audio-visual feature representation learning and context modeling.
3 LoCoNet
Figure 3: An overview of LoCoNet. Given a sequence of face tracks and audio of a target speaker, we sample Sโ1 ๐ 1 S-1 italic_S - 1 speakers from all other people appearing in the scene and stack their face crops as visual input. Our method consists of 3 components: an audio encoder, a visual encoder, and a Long-Short Context Modeling module (LSCM) with N ๐ N italic_N blocks, where each block includes an attention-based Long-term Intra-speaker Model (LIM) and a convolution-based Short-term Inter-speaker Model (SIM) for speaker interaction. LIM involves Audio-Visual Self-Attention for long-term intra-speaker dependencies and Audio-Visual Cross-Attention for audio-visual interaction. The final output is used to classify speaking activity of the target person across all frames.
Given the stacked visual face track Vโโ SรTรHรWร1 ๐ superscript โ ๐ ๐ ๐ป ๐ 1 V\in\mathbb{R}^{S\times T\times H\times W\times 1}italic_V โ blackboard_R start_POSTSUPERSCRIPT italic_S ร italic_T ร italic_H ร italic_W ร 1 end_POSTSUPERSCRIPT and the audio Mel-spectrograms Aโโ 4โขTรM ๐ด superscript โ 4 ๐ ๐ A\in\mathbb{R}^{4T\times M}italic_A โ blackboard_R start_POSTSUPERSCRIPT 4 italic_T ร italic_M end_POSTSUPERSCRIPT, LoCoNet aims to predict the speaking activity R^โโ T^๐ superscript โ ๐\hat{R}\in\mathbb{R}^{T}over^ start_ARG italic_R end_ARG โ blackboard_R start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT of the target person in each frame. S ๐ S italic_S is the number of speakers including the target speaker and Sโ1 ๐ 1 S-1 italic_S - 1 context speakers in the same scene. T ๐ T italic_T is the temporal length of the face track. H ๐ป H italic_H and W ๐ W italic_W are the height and width of each visual face crop. M ๐ M italic_M is the number of frequency bins of the audio Mel-spectrograms.
As shown in Fig3, LoCoNet consists of a visual encoder, an audio encoder, and a Long-Short Context Modeling (LSCM) module with N ๐ N italic_N LSCM blocks. We explain each module in more detail below.
3.1 Encoders
Visual encoder. Given the face crop track V iโโ TรHรWร1 subscript ๐ ๐ superscript โ ๐ ๐ป ๐ 1 V_{i}\in\mathbb{R}^{T\times H\times W\times 1}italic_V start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT โ blackboard_R start_POSTSUPERSCRIPT italic_T ร italic_H ร italic_W ร 1 end_POSTSUPERSCRIPT of speaker v i subscript ๐ฃ ๐ v_{i}italic_v start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT, the visual encoder yields a time sequence of visual embeddings f v iโโ TรC,i=1,โฆ,S formulae-sequence subscript ๐ subscript ๐ฃ ๐ superscript โ ๐ ๐ถ ๐ 1โฆ๐ f_{v_{i}}\in\mathbb{R}^{T\times C},i=1,...,S italic_f start_POSTSUBSCRIPT italic_v start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT โ blackboard_R start_POSTSUPERSCRIPT italic_T ร italic_C end_POSTSUPERSCRIPT , italic_i = 1 , โฆ , italic_S. The stacked visual embeddings of the target speaker and all the sampled context speakers f vโโ SรTรC subscript ๐ ๐ฃ superscript โ ๐ ๐ ๐ถ f_{v}\in\mathbb{R}^{S\times T\times C}italic_f start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT โ blackboard_R start_POSTSUPERSCRIPT italic_S ร italic_T ร italic_C end_POSTSUPERSCRIPT represent temporal context of each speaker independently.
Audio Encoder. The audio encoder takes audio Mel-spectrograms Aโโ 4โขTรM ๐ด superscript โ 4 ๐ ๐ A\in\mathbb{R}^{4T\times M}italic_A โ blackboard_R start_POSTSUPERSCRIPT 4 italic_T ร italic_M end_POSTSUPERSCRIPT as input. We need frame-level audio features for per-frame classification, but most pretrained audio encoders[27, 56, 12, 23] are for audio classification and thus have a high degree of temporal downsampling. To solve this problem, we propose VGGFrame as the audio encoder, which can fully utilize the pretrained VGGish[27]. The architecture of VGGFrame is illustrated in Fig4. We remove the temporal downsampling layer after block-4 and add a deconvolutional layer to upsample the temporal dimension. We concatenate the intermediate features with the upsampled features to extract a hierarchy of representations. VGGFrame outputs the audio embeddings f a iโโ TรC subscript ๐ subscript ๐ ๐ superscript โ ๐ ๐ถ f_{a_{i}}\in\mathbb{R}^{T\times C}italic_f start_POSTSUBSCRIPT italic_a start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT โ blackboard_R start_POSTSUPERSCRIPT italic_T ร italic_C end_POSTSUPERSCRIPT. To align with S ๐ S italic_S speakers, we repeat f a i subscript ๐ subscript ๐ ๐ f_{a_{i}}italic_f start_POSTSUBSCRIPT italic_a start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT S ๐ S italic_S times to produce the audio embedding f aโโ SรTรC subscript ๐ ๐ superscript โ ๐ ๐ ๐ถ f_{a}\in\mathbb{R}^{S\times T\times C}italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT โ blackboard_R start_POSTSUPERSCRIPT italic_S ร italic_T ร italic_C end_POSTSUPERSCRIPT.
Figure 4: An illustration of our proposed audio encoder VGGFrame. We apply a deconvolutional layer to upsample the output feature of block-4. The output features of block-3 (before max pooling) and deconvolutional layer are concatenated and transformed to per-frame output features of shape TรC ๐ ๐ถ T\times C italic_T ร italic_C.
3.2 Long-Short Context Modeling
The visual and audio embeddings, derived independently by the visual and audio encoders, lack consideration of intra/inter-speaker context. Our Long-Short Context Modeling (LSCM) is designed to enhance these embeddings by learning long-term intra-speaker and short-term inter-speaker context in an interleaved manner. As shown in Fig.3, LSCM consists of N ๐ N italic_N blocks, each incorporating a Long-term Intra-speaker Modeling (LIM) module and a Short-term Inter-speaker Modeling (SIM) module consecutively. LIM limits the model to look at the same speaker across all frames, encouraging it to learn speaker-independent patterns from audio and visual interactions. In contrast, SIM constrains the model to examine all speakers in nearby frames and capture local interactions. Such inductive bias is instrumental in enhancing the modelโs capacity to glean valuable insights from these contextual dimensions.
LSCM inputs audio embeddings f a subscript ๐ ๐ f_{a}italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT and visual embeddings f v subscript ๐ ๐ฃ f_{v}italic_f start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT, and produces context-aware embeddings u a N,u v Nโโ SรTรC superscript subscript ๐ข ๐ ๐ superscript subscript ๐ข ๐ฃ ๐ superscript โ ๐ ๐ ๐ถ u_{a}^{N},u_{v}^{N}\in\mathbb{R}^{S\times T\times C}italic_u start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_N end_POSTSUPERSCRIPT , italic_u start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_N end_POSTSUPERSCRIPT โ blackboard_R start_POSTSUPERSCRIPT italic_S ร italic_T ร italic_C end_POSTSUPERSCRIPT. These embeddings are concatenated to yield the final embeddings u N=cโขoโขnโขcโขaโขtโข(u a N,u v N)superscript ๐ข ๐ ๐ ๐ ๐ ๐ ๐ ๐ก superscript subscript ๐ข ๐ ๐ superscript subscript ๐ข ๐ฃ ๐ u^{N}=concat(u_{a}^{N},u_{v}^{N})italic_u start_POSTSUPERSCRIPT italic_N end_POSTSUPERSCRIPT = italic_c italic_o italic_n italic_c italic_a italic_t ( italic_u start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_N end_POSTSUPERSCRIPT , italic_u start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_N end_POSTSUPERSCRIPT ). Via a linear layer, u N superscript ๐ข ๐ u^{N}italic_u start_POSTSUPERSCRIPT italic_N end_POSTSUPERSCRIPT is used to predict the speaking activities R^โโ T^๐ superscript โ ๐\hat{R}\in\mathbb{R}^{T}over^ start_ARG italic_R end_ARG โ blackboard_R start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT of the target person. The computation process at each LSCM block l ๐ l italic_l is detailed below.
3.2.1 Long-term Intra-speaker Modeling (LIM)
LIM consists of two submodules: (i) Audio/Visual Self-Attention models an individual personโs behavior over a longer time period, and (ii) Audio-Visual Cross-Attention learns the interaction between audio and visual embeddings.
Audio/Visual Self-Attention. Since the model needs large capacity and to learn long-term dependencies, we employ the attention mechanism[67] with a Transformer layer applied on the temporal dimension to achieve long-term modeling,
uv l superscript subscript๐ข ๐ฃ ๐\displaystyle\tilde{u}{v}^{l}over~ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT=LNโข(MHAโข(u v lโ1,u v lโ1,u v lโ1)+u v lโ1),absent LN MHA superscript subscript ๐ข ๐ฃ ๐ 1 superscript subscript ๐ข ๐ฃ ๐ 1 superscript subscript ๐ข ๐ฃ ๐ 1 superscript subscript ๐ข ๐ฃ ๐ 1\displaystyle=\text{LN}(\text{MHA}(u{v}^{l-1},u_{v}^{l-1},u_{v}^{l-1})+u_{v}^% {l-1}),= LN ( MHA ( italic_u start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l - 1 end_POSTSUPERSCRIPT , italic_u start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l - 1 end_POSTSUPERSCRIPT , italic_u start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l - 1 end_POSTSUPERSCRIPT ) + italic_u start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l - 1 end_POSTSUPERSCRIPT ) ,(1)
uv l superscript subscript๐ข ๐ฃ ๐\displaystyle\tilde{u}{v}^{l}over~ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT=LNโข(MLPโข(uv l)+uv l),absent LN MLP superscript subscript๐ข ๐ฃ ๐ superscript subscript๐ข ๐ฃ ๐\displaystyle=\text{LN}(\text{MLP}(\tilde{u}{v}^{l})+\tilde{u}_{v}^{l}),= LN ( MLP ( over~ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT ) + over~ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT ) ,(2)
where u v 0=f v superscript subscript ๐ข ๐ฃ 0 subscript ๐ ๐ฃ u_{v}^{0}=f_{v}italic_u start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 0 end_POSTSUPERSCRIPT = italic_f start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT, u a 0=f a superscript subscript ๐ข ๐ 0 subscript ๐ ๐ u_{a}^{0}=f_{a}italic_u start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 0 end_POSTSUPERSCRIPT = italic_f start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT, u v lโ1โโ SรTรC superscript subscript ๐ข ๐ฃ ๐ 1 superscript โ ๐ ๐ ๐ถ u_{v}^{l-1}\in\mathbb{R}^{S\times T\times C}italic_u start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l - 1 end_POSTSUPERSCRIPT โ blackboard_R start_POSTSUPERSCRIPT italic_S ร italic_T ร italic_C end_POSTSUPERSCRIPT are the output visual embeddings of the previous LSCM block, LNโข(โ
)LNโ
\text{LN}(\cdot)LN ( โ
) denotes Layer Normalization[5], MHAโข(q,k,v)MHA ๐ ๐ ๐ฃ\text{MHA}(q,k,v)MHA ( italic_q , italic_k , italic_v ) is multi-head attention[67] with query q ๐ q italic_q, key k ๐ k italic_k, value v ๐ฃ v italic_v, and MLP is a multi-layer perceptron. Audio Self-Attention is applied in the same way to u a lโ1 superscript subscript ๐ข ๐ ๐ 1 u_{a}^{l-1}italic_u start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l - 1 end_POSTSUPERSCRIPT to obtain the audio output ua l superscript subscript๐ข ๐ ๐\tilde{u}_{a}^{l}over~ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT.
Audio-Visual Cross-Attention. The visual and audio streams are so far processed separately. To enhance visual features with audio and vice versa, we use an Audio-Visual dual Cross-Attention,
u^v l superscript subscript^๐ข ๐ฃ ๐\displaystyle\widehat{u}{v}^{l}over^ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT=LNโข(MHAโข(uv l,ua l,ua l)+uv l),absent LN MHA superscript subscript๐ข ๐ฃ ๐ superscript subscript๐ข ๐ ๐ superscript subscript๐ข ๐ ๐ superscript subscript๐ข ๐ฃ ๐\displaystyle=\text{LN}(\text{MHA}(\tilde{u}{v}^{l},\tilde{u}{a}^{l},\tilde{% u}{a}^{l})+\tilde{u}{v}^{l}),= LN ( MHA ( over~ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT , over~ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT , over~ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT ) + over~ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT ) ,(3)
u^v l superscript subscript^๐ข ๐ฃ ๐\displaystyle\widehat{u}{v}^{l}over^ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT=LNโข(MLPโข(u^v l)+u^v l),absent LN MLP superscript subscript^๐ข ๐ฃ ๐ superscript subscript^๐ข ๐ฃ ๐\displaystyle=\text{LN}(\text{MLP}(\widehat{u}{v}^{l})+\widehat{u}{v}^{l}),= LN ( MLP ( over^ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT ) + over^ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT ) ,(4)
where u^v l superscript subscript^๐ข ๐ฃ ๐\widehat{u}{v}^{l}over^ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT are audio-enhanced visual embeddings. Visual-enhanced audio embeddings u^a l superscript subscript^๐ข ๐ ๐\widehat{u}{a}^{l}over^ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT are obtained in the same way via Audio-Visual Cross-Attention given ua l superscript subscript๐ข ๐ ๐\tilde{u}{a}^{l}over~ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT and uv l superscript subscript๐ข ๐ฃ ๐\tilde{u}{v}^{l}over~ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT.
3.2.2 Short-term Inter-speaker Modeling (SIM)
For a given moment in the video, the speaking activity of a target person is more coordinated with other speakers in closer frames[57], so the model should capture local temporal inter-speaker relationships. To do this, we employ a small Inter-speaker Convolutional Network,
u v l superscript subscript ๐ข ๐ฃ ๐\displaystyle u_{v}^{l}italic_u start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT=MLPโข(LNโข(Conv sรkโข(u^v l)))+u^v l,absent MLP LN subscript Conv ๐ ๐ superscript subscript^๐ข ๐ฃ ๐ superscript subscript^๐ข ๐ฃ ๐\displaystyle=\text{MLP}(\text{LN}(\text{Conv}{s\times k}(\widehat{u}{v}^{l}% )))+\widehat{u}{v}^{l},= MLP ( LN ( Conv start_POSTSUBSCRIPT italic_s ร italic_k end_POSTSUBSCRIPT ( over^ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT ) ) ) + over^ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT ,(5) u a l superscript subscript ๐ข ๐ ๐\displaystyle u{a}^{l}italic_u start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT=MLPโข(LNโข(Conv sรkโข(u^a l)))+u^a l,absent MLP LN subscript Conv ๐ ๐ superscript subscript^๐ข ๐ ๐ superscript subscript^๐ข ๐ ๐\displaystyle=\text{MLP}(\text{LN}(\text{Conv}{s\times k}(\widehat{u}{a}^{l}% )))+\widehat{u}_{a}^{l},= MLP ( LN ( Conv start_POSTSUBSCRIPT italic_s ร italic_k end_POSTSUBSCRIPT ( over^ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT ) ) ) + over^ start_ARG italic_u end_ARG start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT ,(6)
where visual embeddings u v lโโ SรTรC superscript subscript ๐ข ๐ฃ ๐ superscript โ ๐ ๐ ๐ถ u_{v}^{l}\in\mathbb{R}^{S\times T\times C}italic_u start_POSTSUBSCRIPT italic_v end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT โ blackboard_R start_POSTSUPERSCRIPT italic_S ร italic_T ร italic_C end_POSTSUPERSCRIPT and audio embeddings u a lโโ SรTรC superscript subscript ๐ข ๐ ๐ superscript โ ๐ ๐ ๐ถ u_{a}^{l}\in\mathbb{R}^{S\times T\times C}italic_u start_POSTSUBSCRIPT italic_a end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l end_POSTSUPERSCRIPT โ blackboard_R start_POSTSUPERSCRIPT italic_S ร italic_T ร italic_C end_POSTSUPERSCRIPT are the output of the l ๐ l italic_l-th LSCM block that will be passed to the next block. k ๐ k italic_k is the temporal length of the receptive field and s ๐ s italic_s is the number of speakers considered. Explicitly modeling inter-speaker context in nearby frames enables cross-frame inter-speaker information exchange. Our SIM module with a short temporal receptive field can help capture local dynamic patterns in interactions.
3.3 Training and Inference
Following[10, 1, 64], we train our model with multiple supervisions utilizing u Nโโ SรTร2โขC superscript ๐ข ๐ superscript โ ๐ ๐ 2 ๐ถ u^{N}\in\mathbb{R}^{S\times T\times 2C}italic_u start_POSTSUPERSCRIPT italic_N end_POSTSUPERSCRIPT โ blackboard_R start_POSTSUPERSCRIPT italic_S ร italic_T ร 2 italic_C end_POSTSUPERSCRIPT (Sec.3.2) and u i superscript ๐ข ๐ u^{i}italic_u start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT derived from each intermediate block of LSCM. For each u i superscript ๐ข ๐ u^{i}italic_u start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT, a fully-connected (FC) layer is applied, yielding prediction results R^iโโ T superscript^๐ ๐ superscript โ ๐\hat{R}^{i}\in\mathbb{R}^{T}over^ start_ARG italic_R end_ARG start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT โ blackboard_R start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT corresponding to the target speaker for each frame. All FC layers share their parameters. The overall loss function is L=โi=1 N CrossEntropyโข(R^i,R)๐ฟ superscript subscript ๐ 1 ๐ CrossEntropy superscript^๐ ๐ ๐ L=\sum_{i=1}^{N}\text{CrossEntropy}(\hat{R}^{i},R)italic_L = โ start_POSTSUBSCRIPT italic_i = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_N end_POSTSUPERSCRIPT CrossEntropy ( over^ start_ARG italic_R end_ARG start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT , italic_R ) where R ๐ R italic_R is the ground-truth. During inference, we can reduce the computation by a factor of S ๐ S italic_S by reusing speakersโ features extracted by the visual encoder, which uses the most FLOPs.
3.4 Implementation details
Following[65], our visual encoder consists of a 3D convolutional layer, ResNet-18[26], and a visual temporal convolution network (V-TCN)[41].
Our audio encoder is the proposed VGGFrame initialized with VGGish[12] weights pretrained on AudioSet[22].
We sample S=3 ๐ 3 S=3 italic_S = 3 speakers and T=200 ๐ 200 T=200 italic_T = 200 frames. In SIM, s ๐ s italic_s is set to 3, which is the same as S ๐ S italic_S, and k ๐ k italic_k is set to 7 frames. The face crops are resized to 112ร112 112 112 112\times 112 112 ร 112. Visual augmentation includes randomly resized cropping, horizontal flipping, and rotation. For audio augmentation, another audio signal is randomly chosen from the rest of the training set and added as noise to the target audio. We train LoCoNet with Adam[36] for 25 epochs on 4 RTX6000 GPUs with batch size 4 using PyTorch[51]. The learning rate is 5ร10โ5 5 superscript 10 5 5\times 10^{-5}5 ร 10 start_POSTSUPERSCRIPT - 5 end_POSTSUPERSCRIPT and reduced by 5% each epoch.
4 Experimental Setup
4.1 Datasets
AVA-ActiveSpeaker[58] is a standard benchmark for ASD, consisting of 262 videos from Hollywood movies with 3.65 million frames and 5.3 million face crops. Following[45, 48, 47, 15, 3, 31], we evaluate on the validation set.
Talkies[42] is an in-the-wild ASD dataset with 23,507 face tracks extracted from 421,997 labeled frames. It focuses on challenging cases with more speakers, diverse actors and scenes, and more off-screen speech.
Ego4D[24]โs Audio-Visual benchmark has 3,670 hours of egocentric video of unscripted daily activities in many environments. It includes many challenges that complement exocentric benchmarks, including unusual viewpoints and speakers who are off-screen.
4.2 Evaluation Metric
Following[65, 2, 39, 42, 3], we use the official ActivityNet[9] evaluation tool to compute mean average precision (mAP) and evaluate on the AVA-ActiveSpeaker[58]. We also compute AUC[8] as another evaluation metric using Sklearn[52]. We use mAP to evaluate on Talkies[42] and Ego4D[24].
5 Results and Analysis
We first compare the proposed method LoCoNet with previous state-of-the-art methods on multiple datasets and challenging scenarios. Then we validate our hypotheses of long-term intra-speaker and short-term inter-speaker modeling. Finally, following [39, 2, 65, 45, 75], we conduct extensive ablations on each component of LoCoNet(on AVA-ActiveSpeaker[58] unless otherwise noted).
5.1 Comparison with State-of-the-Art
In this section, we compare our approach with state-of-the-art methods on the three datasets.
AVA-ActiveSpeaker. From Table1, end-to-end methods exhibit fewer FLOPs while maintaining competitive mAP compared to multi-stage methods. The higher FLOPs of multi-stage methods stem from their sequential-inference processing strategy (Sec. 2) of stacking multiple neighboring frames centered at time t ๐ก t italic_t. LoCoNet achieves a 95.2% mAP, surpassing the best-performing end-to-end ASD method Light-ASD[45] by 1.1% while using modest average-FLOPs. Moreover, LoCoNet outperforms previous state-of-the-art multi-stage method SPELL+[48] by 0.3% with about 32% fewer parameters and over 38ร\timesร fewer average-FLOPs.
Method Vid. Enc.Params average-mAP AUC FLOPs Multi-Stage ASC [2]R-18[26]23.0โขM 23.0 ๐ 23.0M 23.0 italic_M 1.0โขG 1.0 ๐บ 1.0G 1.0 italic_G 87.1- MAAS [42]R-18 23.0โขM 23.0 ๐ 23.0M 23.0 italic_M 1.6โขG 1.6 ๐บ 1.6G 1.6 italic_G 88.8- ASDNet [39]3DRNext-101 51.0โขMโ 51.0 superscript ๐โ 51.0M^{\dagger}51.0 italic_M start_POSTSUPERSCRIPT โ end_POSTSUPERSCRIPT 13.2โขGโ 13.2 superscript ๐บโ 13.2G^{\dagger}13.2 italic_G start_POSTSUPERSCRIPT โ end_POSTSUPERSCRIPT 93.5- SPELL [48]R-18+TSM 23.5โขMโ 23.5 superscript ๐โ 23.5M^{{\dagger}}23.5 italic_M start_POSTSUPERSCRIPT โ end_POSTSUPERSCRIPT 8.7โขGโ 8.7 superscript ๐บโ 8.7G^{{\dagger}}8.7 italic_G start_POSTSUPERSCRIPT โ end_POSTSUPERSCRIPT 94.2- SPELL+ [48]R-50+TSM 51.23โขMโ 51.23 superscript ๐โ 51.23M^{{\dagger}}51.23 italic_M start_POSTSUPERSCRIPT โ end_POSTSUPERSCRIPT 19.6โขGโ 19.6 superscript ๐บโ 19.6G^{{\dagger}}19.6 italic_G start_POSTSUPERSCRIPT โ end_POSTSUPERSCRIPT 94.9- End-to-End UniCon [75]R-18 23.8โขM 23.8 ๐ 23.8M 23.8 italic_M 3.0โขGโ 3.0 superscript ๐บโ 3.0G^{{\dagger}}3.0 italic_G start_POSTSUPERSCRIPT โ end_POSTSUPERSCRIPT 92.2 97.0 TalkNet [65]R-18+VTCN 15.7โขM 15.7 ๐ 15.7M 15.7 italic_M 0.51โขG 0.51 ๐บ 0.51G 0.51 italic_G 92.3 96.8 ASD-Trans [15]R-18+VTCN 15.0โขM 15.0 ๐ 15.0M 15.0 italic_M 0.55โขG 0.55 ๐บ 0.55G 0.55 italic_G 93.0- EASEE [3]3D R-50 26.8โขMโ 26.8 superscript ๐โ 26.8M^{{\dagger}}26.8 italic_M start_POSTSUPERSCRIPT โ end_POSTSUPERSCRIPT 4.3โขGโ 4.3 superscript ๐บโ 4.3G^{{\dagger}}4.3 italic_G start_POSTSUPERSCRIPT โ end_POSTSUPERSCRIPT 94.1- Light-ASD 45D Conv 1.02โขM 1.02 ๐ 1.02M 1.02 italic_M 0.2โขG 0.2 ๐บ 0.2G 0.2 italic_G 94.1- TS-TalkNet [31]R-18+VTCN 36.8โขM 36.8 ๐ 36.8M 36.8 italic_M 2.3โขG 2.3 ๐บ 2.3G 2.3 italic_G 93.9- LoCoNet(Ours)R-18+VTCN 34.3โขM 34.3 ๐ 34.3M 34.3 italic_M 0.51โขG 0.51 ๐บ 0.51G 0.51 italic_G 95.2 98.0
Table 1: Comparison with SOTAs on AVA-ActiveSpeaker. 3DRNext denotes 3D ResNext[71]. R denotes 2D ResNet[26]. average-FLOPs represents the averaged FLOPs needed to process a single face crop. โ denotes our estimates based on their visual encoders. Most methods incur higher costs by extracting features for each frame through stacking multiple adjacent frames (i.e., 11 in SPELL). LoCoNet achieves the highest mAP with modest FLOPs.
Method Train Set mAP AVA Talkies(%) MAAS[42]โโ79.7 EASEE[3]โโ86.7 LoCoNetโโ88.4 EASEE[3]โโ93.6 light-ASD[45]โโ93.9 LoCoNetโโ96.1 EASEE[3]โโ94.5 LoCoNetโโ97.2
Table 2: Comparison on Talkies dataset under three training settings: train on AVA-ActiveSpeaker alone, train on Talkies alone, or train on AVA-ActiveSpeaker and finetune on Talkies.
Talkies set. We evaluate LoCoNet with other methods under three training settings: (i) AVA-ActiveSpeaker, (ii) Talkies, and (iii) pretrained on AVA-ActiveSpeaker and finetuned on Talkies. As shown in Table2, LoCoNet outperforms EASEE by 1.7%, 2.5%, and 2.7% in these three settings, respectively. It also outperforms light-ASD by 2.2% when both models are trained on Talkies only.
Ego4D dataset. We evaluate our method on Ego4D Audio-Visual benchmark[24]. LoCoNet achieves 68.4% mAP, outperforming TalkNet and Challenge Winner[47] (SPELL[48]-based) by 16.7% and 7.7% respectively. Egocentric videos, characterized by constant camera motion, lower clarity, and more complex scenes compared to exocentric videos, demonstrate the potential of the proposed Long-Short Context Modeling in real-life scenarios. Our approach benefits from capturing both multi-speaker interaction and single-speaker behavior.
Method mAP (%) TalkNet[65]51.7 Challenge Winner[47]60.7 LoCoNet 68.4
Table 3: Comparison on Ego4D dataset. The challenge winner is not specifically optimized for ASD but the large improvement (+7.7%percent 7.7+7.7%+ 7.7 %) still shows the strong generalizability of LoCoNet. The result of TalkNet was obtained by us with their released code.
5.2 Challenging Scenario Evaluation
Quantitative analysis.
We report the performance of LoCoNet on AVA-ActiveSpeaker under different face sizes: (i) Small: faces with width less than 64 pixels; (ii) Medium: faces with width between 64 and 128 pixels; (iii) Large: faces with width larger than 128 pixels. We also study the effect of the number of visible faces in a video frame (1, 2, or 3).
The results, along with the portions of each category, are shown in Tables4 and 5 respectively. LoCoNet consistently performs the best across all scenarios, exhibiting the most significant improvement in the challenging multi-speaker case: +3.0% for 3 faces. This suggests that our method more
effectively models both the target speakerโs speaking pattern and the interactions of context speakers, allowing accurate inference of the speaking activity of the target person.
Method Face Size Small(18%)Medium(30%)Large(52%) ASC[2]56.2 79.0 92.2 MAAS[42]55.2 79.4 93.0 TalkNet[65]63.7 85.9 95.3 ASDNet[39]74.3 89.8 96.3 EASEE[3]75.9 90.6 96.7 light-ASD[45]77.5 91.2 96.5 LoCoNet 77.8 93.0 97.3
Table 4: Results as a function of face size. LoCoNet achieves the highest mAP among all face sizes.
Method# Faces 1(45%)2(33%)3(11%) ASC[2]91.8 83.8 67.6 MAAS[42]93.3 85.8 68.2 TalkNet[65]95.4 89.6 80.3 ASDNet[39]95.7 92.4 83.7 EASEE[3]96.5 92.4 83.9 light-ASD[45]96.2 92.6 84.4 LoCoNet 97.0 94.6 87.4
Table 5: Results as a function of visible faces in the scene. Larger improvements are observed on more challenging cases (i.e., 3 faces).
Figure 5: Results comparison of LoCoNet and TalkNet on challenging scenarios of AVA-ActiveSpeaker.Red boxes denote not-active speaker. Green box denote active speaker. Orange circles refer to false predictions. The video on the left shows a multi-people conversation with four speakers, and separate conversation of two. The video on the right shows an active speaker with a small face. Both scenes are challenging, and LoCoNet predicts accurately in most cases.
Qualitative analysis. Fig5 visualizes the results of LoCoNet and TalkNet[65] on AVA-ActiveSpeaker[58] with the groundtruth labels. The video on the left shows four visible speakers talking, posing a challenge in distinguishing the active speaker amid multiple discussions in the same scene. LoCoNet accurately locates the active speakers in this case, whereas TalkNet fails to recognize some of them. The first two columns of the video on the right shows a woman with a very small visible face as the active speaker, while the man with a large visible face is not speaking. TalkNet fails to locate the active speaker while LoCoNet succeeds. By combining long-term intra-speaker context to compare the speaking pattern of each individual and short-term inter-speaker context to examine the conversations, our approach better overcomes this challenging speaking scenarios. However, in the last column, both methods fail to recognize the active speaker at the back. This scenario is especially challenging as the two active speakers are in separate conversations with one being far less salient than the other, making it difficult to infer the speaking activity of the less salient speaker.
5.3 Attention visualizations of LIM and SIM
We next visualize the effectiveness of Long-term Intra-speaker Modeling (LIM) and Short-term Inter-speaker Modeling (SIM). The left part of Fig.6 visualizes the attention weights across different frames of a single speaker in LIM. It is evident that speaking and non-speaking activities are distinctly separated, with clear boundaries when speaking activities change. This verifies that LIM contributes to accurate speaking activity detection. The right part of Fig.6 shows the portions of information (measured by L2-norm, as convolutions are used for SIM) drawn by the target speaker at target frame from all speakers at nearby frames (and information flows with very small portions are not shown). The woman on the left gradually turns her face to the target speaker, which is the most indicative sign that the target speaker has started speaking. The distribution of information flow reveals that SIM infers from the behaviors of context speakers, assigning more attention to the woman on the left.
Figure 6: Visualization of LIM and SIM. Left: attention weights of one speaker in LIM, showing the ability to capture long-term intra-speaker context. Right: to predict the speaking activity of the target speaker at target frame, we show the information drawn from not-speaking target and speaking target and context speakers at nearby frames.
5.4 Context Modeling Analysis
Does Intra-speaker Modeling require long-term? We train LoCoNet by keeping the number of speakers S ๐ S italic_S as 1, and varying the temporal length of input frames T ๐ T italic_T from 20 to 400. Table6 indicates that the network performs worst when trained with the shortest video segments of 20 frames (0.8 sec). Performance improves as video segments become longer, with a 5.0% mAP increase at 100 frames (4 sec) and an additional 0.4% at 200 frames. This underscores the importance of long-term temporal context in intra-speaker modeling, aligning with findings from TalkNet[65] and ASC[2] that long-term temporal context provides better evidence of a speaking episode.
We found 200 frames to be a good balance between performance and memory cost.
Frames 20 100 200 300 400
mAP (%)87.8 92.8 93.2 93.2 OOM
Table 6: Temporal Length in Long-term Intra-speaker Modeling. We set S=1 ๐ 1 S=1 italic_S = 1 so the context is intra-speaker only.
Speakers 1 2 3 mAP (%)93.2 94.6 95.2
Table 7: Number of Speakers S ๐ S italic_S in Short-term Inter-speaker Modeling. We set the number of frames T ๐ T italic_T to 200.
Receptive Field 1 7 15 31 101 GFLOPs 0.28 0.98 1.93 3.82 12.1 mAP(%)94.5 95.2 95.2 95.0 94.9
Table 8: Temporal Receptive Field k ๐ k italic_k in SIM, indicating how many neighboring frames each speaker can explicitly query on in the SIM module. S ๐ S italic_S is set as 3 and T ๐ T italic_T as 200.
Is Short-term Inter-speaker Modeling Sufficient? We first validate the importance of inter-speaker context. Keeping the temporal length at 200 frames, we vary the number of speakers S ๐ S italic_S from 1 to 3. Larger S ๐ S italic_S values were not considered because >>> 99% of videos in AVA-ActiveSpeaker[58] have at most 3 speakers in the same scene. Table7 demonstrates that performance increases with more speakers included in training. This supports our hypothesis that modeling multiple speakers in the inter-speaker context is necessary for ASD.
Next, with S=3 ๐ 3 S=3 italic_S = 3, we vary the temporal receptive field k ๐ k italic_k of SIM from 1 (40 msec) to 101 (4 sec). Table8 shows a 0.7% performance increase when increasing k ๐ k italic_k from 1 to 7, confirming our assumption that motion in short-term inter-speaker context is more valuable than in-frame context. Performance saturates with further increases in the receptive field, and the computation cost increases drastically. This reinforces our hypothesis in Sec.1 that short-term inter-speaker modeling is sufficient.
5.5 Other Ablations
Does each component of LoCoNet help? In Table9, replacing ResNet-34[26] with VGGFrame as the audio encoder enhances mAP by 1.6%, showing the effectiveness of an audio encoder pretrained on audio datasets compared to a common vision encoder. With VGGFrame as audio encoder, adding 3 Long-term Intra-speaker Modeling (LIM) modules increases the performance by 1.3%, while 3 Short-term Inter-speaker Modeling (SIM) modules adds 1.2%. Both together improve mAP by 2.4%.
Ablation Settings mAP R-34 VGGFrame LIM SIM(%) โ91.2 โ92.8 โโ94.1 โโ94.0 โโโ95.2
Table 9: Ablations on components of LoCoNet. We study the efficacy of audio encoder (ResNet-34 or VGGFrame), Long-term Intra-speaker Modeling (LIM) and Short-term Inter-speaker Modeling (SIM). LIM and SIM are stacked 3 times if applied.
Convolution versus Window Self-Attention in SIM.
Design mAP(%) Convolution 95.2 Window Self-Attention 94.4
Table 10: Ablation on designs of Short-term Inter-speaker Modeling. We implement SIM using Convolution or Window Self-Attention (Eqn.5). We keep the reception field the same (i.e., 7 frames) and stack it three times (N=3 ๐ 3 N=3 italic_N = 3).
Besides Convolution, we also try Window Self-Attention to capture local patterns for SIM (Eqn.5). In Table10, Convolution outperforms Window Self-Attention by 0.8% mAP, highlighting the superiority of Convolution in modeling local patterns of speakersโ interaction.
N ๐ N italic_N 0 1 2 3 4 5 mAP(%)92.8 94.3 95.0 95.2 95.0 95.0
Table 11: Ablation on the number of blocks in LSCM (N ๐ N italic_N).N=0 ๐ 0 N=0 italic_N = 0 refers to LoCoNet with no context modeling.
Number of blocks in LSCM. We vary the number of blocks N ๐ N italic_N in Long-Short Context Modeling (Sec.3.2). Adding 1 block of LSCM yields a performance gain of 1.5%, showing the effectiveness of LIM and SIM. Adding two more blocks further increases by 0.7%, but results saturate at N=3 ๐ 3 N=3 italic_N = 3.
6 Conclusion
In this work, we observe that speaker activity can be more efficiently inferred from long-term intra-speaker context and short-term inter-speaker context. We thus design an end-to-end long-short context ASD framework that uses self-attention and cross-attention mechanisms to model long-term intra-speaker context and a convolutional network to model short-term inter-speaker context. With a simple backbone network, our method achieves state-of-the-art performance on 3 mainstream ASD benchmarks and significantly outperforms previous state of the art methods by 7.7% on Ego4D. We also show that in challenging scenarios where multiple speakers are in the same scene or speakers have small faces, our proposed method also outperforms previous methods. All of these results show the robustness and effectiveness of our method. Similar to existing long-term ASD methods, our method utilizes 8 sec of context. Future work should study how to implement larger contexts. Additionally, enhancing ASD in egocentric datasets could include adding other modalities, such as eye gaze.
Acknowledgements
The authors gratefully acknowledge Prof.David Crandallโs guidance and feedback on earlier versions of this work. This work was supported in part by the National Science Foundation under award DRL-2112635 to the AI Institute for Engaged Learning, Sony Faculty Innovation Award, Laboratory for Analytic Sciences via NC State University, and ONR Award N00014-23-1-2356. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.
References
- Al-Rfou et al. [2019] Rami Al-Rfou, Dokook Choe, Noah Constant, Mandy Guo, and Llion Jones. Character-level language modeling with deeper self-attention. In Proceedings of the AAAI conference on artificial intelligence, pages 3159โ3166, 2019.
- Alcรกzar et al. [2020] Juan Leรณn Alcรกzar, Fabian Caba, Long Mai, Federico Perazzi, Joon-Young Lee, Pablo Arbelรกez, and Bernard Ghanem. Active speakers in context. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12465โ12474, 2020.
- Alcazar et al. [2022] Juan Leon Alcazar, Moritz Cordes, Chen Zhao, and Bernard Ghanem. End-to-end active speaker detection. arXiv preprint arXiv:2203.14250, 2022.
- Arandjelovic and Zisserman [2018] Relja Arandjelovic and Andrew Zisserman. Objects that sound. In Proceedings of the European conference on computer vision (ECCV), pages 435โ451, 2018.
- Ba et al. [2016] Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton. Layer normalization. arXiv preprint arXiv:1607.06450, 2016.
- Bansal et al. [2018] Aayush Bansal, Shugao Ma, Deva Ramanan, and Yaser Sheikh. Recycle-gan: Unsupervised video retargeting. In Proceedings of the European conference on computer vision (ECCV), pages 119โ135, 2018.
- Beyan et al. [2020] Cigdem Beyan, Muhammad Shahid, and Vittorio Murino. Realvad: A real-world dataset and a method for voice activity detection by body motion analysis. IEEE Transactions on Multimedia, 23:2071โ2085, 2020.
- Bradley [1997] Andrew P Bradley. The use of the area under the roc curve in the evaluation of machine learning algorithms. Pattern recognition, 30(7):1145โ1159, 1997.
- Caba Heilbron et al. [2015] Fabian Caba Heilbron, Victor Escorcia, Bernard Ghanem, and Juan Carlos Niebles. Activitynet: A large-scale video benchmark for human activity understanding. In Proceedings of the ieee conference on computer vision and pattern recognition, pages 961โ970, 2015.
- Carion et al. [2020] Nicolas Carion, Francisco Massa, Gabriel Synnaeve, Nicolas Usunier, Alexander Kirillov, and Sergey Zagoruyko. End-to-end object detection with transformers. In European conference on computer vision, pages 213โ229. Springer, 2020.
- Chakravarty and Tuytelaars [2016] Punarjay Chakravarty and Tinne Tuytelaars. Cross-modal supervision for learning active speaker detection in video. In European Conference on Computer Vision, pages 285โ301. Springer, 2016.
- Chen et al. [2020] Honglie Chen, Weidi Xie, Andrea Vedaldi, and Andrew Zisserman. Vggsound: A large-scale audio-visual dataset. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 721โ725. IEEE, 2020.
- Cho et al. [2014] Kyunghyun Cho, Bart Van Merriรซnboer, Dzmitry Bahdanau, and Yoshua Bengio. On the properties of neural machine translation: Encoder-decoder approaches. arXiv preprint arXiv:1409.1259, 2014.
- Chung et al. [2020] Joon Son Chung, Jaesung Huh, Arsha Nagrani, Triantafyllos Afouras, and Andrew Zisserman. Spot the conversation: speaker diarisation in the wild. arXiv preprint arXiv:2007.01216, 2020.
- Datta et al. [2022] Gourav Datta, Tyler Etchart, Vivek Yadav, Varsha Hedau, Pradeep Natarajan, and Shih-Fu Chang. Asd-transformer: Efficient active speaker detection using self and multimodal transformers. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 4568โ4572. IEEE, 2022.
- Ding et al. [2020] Yifan Ding, Yong Xu, Shi-Xiong Zhang, Yahuan Cong, and Liqiang Wang. Self-supervised learning for audio-visual speaker diarization. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 4367โ4371. IEEE, 2020.
- Duan et al. [2022] Haihan Duan, Junhua Liao, Lehao Lin, and Wei Cai. Flad: a human-centered video content flaw detection system for meeting recordings. In Proceedings of the 32nd Workshop on Network and Operating Systems Support for Digital Audio and Video, pages 43โ49, 2022.
- Everingham et al. [2009] Mark Everingham, Josef Sivic, and Andrew Zisserman. Taking the bite out of automated naming of characters in tv video. Image and Vision Computing, 27(5):545โ559, 2009.
- Gao et al. [2020] Ruohan Gao, Tae-Hyun Oh, Kristen Grauman, and Lorenzo Torresani. Listen to look: Action recognition by previewing audio. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10457โ10467, 2020.
- Garcia-Romero et al. [2017] Daniel Garcia-Romero, David Snyder, Gregory Sell, Daniel Povey, and Alan McCree. Speaker diarization using deep neural network embeddings. In 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 4930โ4934. IEEE, 2017.
- Gebru et al. [2017] Israel D Gebru, Sileye Ba, Xiaofei Li, and Radu Horaud. Audio-visual speaker diarization based on spatiotemporal bayesian fusion. IEEE transactions on pattern analysis and machine intelligence, 40(5):1086โ1099, 2017.
- Gemmeke et al. [2017] Jort F Gemmeke, Daniel PW Ellis, Dylan Freedman, Aren Jansen, Wade Lawrence, R Channing Moore, Manoj Plakal, and Marvin Ritter. Audio set: An ontology and human-labeled dataset for audio events. In 2017 IEEE international conference on acoustics, speech and signal processing (ICASSP), pages 776โ780. IEEE, 2017.
- Gong et al. [2021] Yuan Gong, Yu-An Chung, and James Glass. Ast: Audio spectrogram transformer. arXiv preprint arXiv:2104.01778, 2021.
- Grauman et al. [2022] Kristen Grauman, Andrew Westbury, Eugene Byrne, Zachary Chavis, Antonino Furnari, Rohit Girdhar, Jackson Hamburger, Hao Jiang, Miao Liu, Xingyu Liu, et al. Ego4d: Around the world in 3,000 hours of egocentric video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18995โ19012, 2022.
- Gurvich et al. [2023] Ilya Gurvich, Ido Leichter, Dharmendar Reddy Palle, Yossi Asher, Alon Vinnikov, Igor Abramovski, Vishak Gopal, Ross Cutler, and Eyal Krupka. A real-time active speaker detection system integrating an audio-visual signal with a spatial querying mechanism. arXiv preprint arXiv:2309.08295, 2023.
- He et al. [2016] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770โ778, 2016.
- Hershey et al. [2017] Shawn Hershey, Sourish Chaudhuri, Daniel PW Ellis, Jort F Gemmeke, Aren Jansen, R Channing Moore, Manoj Plakal, Devin Platt, Rif A Saurous, Bryan Seybold, et al. Cnn architectures for large-scale audio classification. In 2017 ieee international conference on acoustics, speech and signal processing (icassp), pages 131โ135. IEEE, 2017.
- Hu et al. [2021] Di Hu, Yake Wei, Rui Qian, Weiyao Lin, Ruihua Song, and Ji-Rong Wen. Class-aware sounding objects localization via audiovisual correspondence. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(12):9844โ9859, 2021.
- Huang and Koishida [2020] Chong Huang and Kazuhito Koishida. Improved active speaker detection based on optical flow. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pages 950โ951, 2020.
- Jati and Georgiou [2019] Arindam Jati and Panayiotis Georgiou. Neural predictive coding using convolutional neural networks toward unsupervised learning of speaker characteristics. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 27(10):1577โ1589, 2019.
- Jiang et al. [2023] Yidi Jiang, Ruijie Tao, Zexu Pan, and Haizhou Li. Target active speaker detection with audio-visual cues. arXiv preprint arXiv:2305.12831, 2023.
- Jung et al. [2023] Chaeyoung Jung, Suyeon Lee, Kihyun Nam, Kyeongha Rho, You Jin Kim, Youngjoon Jang, and Joon Son Chung. Talknce: Improving active speaker detection with talk-aware contrastive learning. arXiv preprint arXiv:2309.12306, 2023.
- Kang and Han [2021] Soo-Han Kang and Ji-Hyeong Han. Video captioning based on both egocentric and exocentric views of robot vision for human-robot interaction. International Journal of Social Robotics, pages 1โ11, 2021.
- Kazakos et al. [2019] Evangelos Kazakos, Arsha Nagrani, Andrew Zisserman, and Dima Damen. Epic-fusion: Audio-visual temporal binding for egocentric action recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5492โ5501, 2019.
- Kim et al. [2021] You Jin Kim, Hee-Soo Heo, Soyeon Choe, Soo-Whan Chung, Yoohwan Kwon, Bong-Jin Lee, Youngki Kwon, and Joon Son Chung. Look whoโs talking: Active speaker detection in the wild. arXiv preprint arXiv:2108.07640, 2021.
- Kingma and Ba [2014] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Kipf and Welling [2016] Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016.
- Kong et al. [2020] Qiuqiang Kong, Yin Cao, Turab Iqbal, Yuxuan Wang, Wenwu Wang, and Mark D Plumbley. Panns: Large-scale pretrained audio neural networks for audio pattern recognition. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 28:2880โ2894, 2020.
- Kรถpรผklรผ et al. [2021] Okan Kรถpรผklรผ, Maja Taseska, and Gerhard Rigoll. How to design a three-stage architecture for audio-visual active speaker detection in the wild. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1193โ1203, 2021.
- Kyoung and Song [2023] Minyoung Kyoung and Hwa Jeon Song. Modeling long-term multimodal representations for active speaker detection with spatio-positional encoder. IEEE Access, 2023.
- Lea et al. [2016] Colin Lea, Rene Vidal, Austin Reiter, and Gregory D Hager. Temporal convolutional networks: A unified approach to action segmentation. In European conference on computer vision, pages 47โ54. Springer, 2016.
- Leรณn-Alcรกzar et al. [2021] Juan Leรณn-Alcรกzar, Fabian Caba Heilbron, Ali Thabet, and Bernard Ghanem. Maas: Multi-modal assignation for active speaker detection. arXiv preprint arXiv:2101.03682, 2021.
- Levinson and Torreira [2015] Stephen C Levinson and Francisco Torreira. Timing in turn-taking and its implications for processing models of language. Frontiers in psychology, 6:731, 2015.
- Li et al. [2018] Bing Li, Chia-Wen Lin, Boxin Shi, Tiejun Huang, Wen Gao, and C-C Jay Kuo. Depth-aware stereo video retargeting. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 6517โ6525, 2018.
- Liao et al. [2023] Junhua Liao, Haihan Duan, Kanghui Feng, Wanbing Zhao, Yanbing Yang, and Liangyin Chen. A light weight model for active speaker detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 22932โ22941, 2023.
- Michelsanti et al. [2021] Daniel Michelsanti, Zheng-Hua Tan, Shi-Xiong Zhang, Yong Xu, Meng Yu, Dong Yu, and Jesper Jensen. An overview of deep-learning-based audio-visual speech enhancement and separation. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 29:1368โ1396, 2021.
- Min [2022] Kyle Min. Intel labs at ego4d challenge 2022: A better baseline for audio-visual diarization. arXiv preprint arXiv:2210.07764, 2022.
- Min et al. [2022] Kyle Min, Sourya Roy, Subarna Tripathi, Tanaya Guha, and Somdeb Majumdar. Learning long-term spatial-temporal graphs for active speaker detection. In Computer VisionโECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23โ27, 2022, Proceedings, Part XXXV, pages 371โ387. Springer, 2022.
- Ngiam et al. [2011] Jiquan Ngiam, Aditya Khosla, Mingyu Kim, Juhan Nam, Honglak Lee, and Andrew Y Ng. Multimodal deep learning. In Proceedings of the 28th international conference on machine learning (ICML-11), pages 689โ696, 2011.
- Owens and Efros [2018] Andrew Owens and Alexei A Efros. Audio-visual scene analysis with self-supervised multisensory features. In Proceedings of the European conference on computer vision (ECCV), pages 631โ648, 2018.
- Paszke et al. [2019] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
- Pedregosa et al. [2011] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825โ2830, 2011.
- Qian et al. [2021a] Xinyuan Qian, Alessio Brutti, Oswald Lanz, Maurizio Omologo, and Andrea Cavallaro. Audio-visual tracking of concurrent speakers. IEEE Transactions on Multimedia, 24:942โ954, 2021a.
- Qian et al. [2021b] Xinyuan Qian, Maulik Madhavi, Zexu Pan, Jiadong Wang, and Haizhou Li. Multi-target doa estimation with an audio-visual fusion mechanism. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 4280โ4284. IEEE, 2021b.
- Rao et al. [2022] Varshanth R Rao, Md Ibrahim Khalil, Haoda Li, Peng Dai, and Juwei Lu. Decompose the sounds and pixels, recompose the events. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 2144โ2152, 2022.
- Ravanelli and Bengio [2018] Mirco Ravanelli and Yoshua Bengio. Speaker recognition from raw waveform with sincnet. In 2018 IEEE Spoken Language Technology Workshop (SLT), pages 1021โ1028. IEEE, 2018.
- Richardson et al. [2008] Daniel Richardson, Rick Dale, and Kevin Shockley. Synchrony and swing in conversation: Coordination, temporal dynamics, and communication. Embodied communication in humans and machines, pages 75โ94, 2008.
- Roth et al. [2020] Joseph Roth, Sourish Chaudhuri, Ondrej Klejch, Radhika Marvin, Andrew Gallagher, Liat Kaver, Sharadh Ramaswamy, Arkadiusz Stopczynski, Cordelia Schmid, Zhonghua Xi, et al. Ava active speaker: An audio-visual dataset for active speaker detection. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 4492โ4496. IEEE, 2020.
- Saenko et al. [2005] Kate Saenko, Karen Livescu, Michael Siracusa, Kevin Wilson, James Glass, and Trevor Darrell. Visual speech recognition with loosely synchronized feature streams. In Tenth IEEE International Conference on Computer Vision (ICCVโ05) Volume 1, pages 1424โ1431. IEEE, 2005.
- Shahid et al. [2021] Muhammad Shahid, Cigdem Beyan, and Vittorio Murino. S-vvad: Visual voice activity detection by motion segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 2332โ2341, 2021.
- Skantze [2021] Gabriel Skantze. Turn-taking in conversational systems and human-robot interaction: a review. Computer Speech & Language, 67:101178, 2021.
- Slaney and Covell [2000] Malcolm Slaney and Michele Covell. Facesync: A linear operator for measuring synchronization of video facial images and audio tracks. Advances in neural information processing systems, 13, 2000.
- Son Chung et al. [2017] Joon Son Chung, Andrew Senior, Oriol Vinyals, and Andrew Zisserman. Lip reading sentences in the wild. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 6447โ6456, 2017.
- Szegedy et al. [2015] Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. Going deeper with convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1โ9, 2015.
- Tao et al. [2021] Ruijie Tao, Zexu Pan, Rohan Kumar Das, Xinyuan Qian, Mike Zheng Shou, and Haizhou Li. Is someone speaking? exploring long-term temporal features for audio-visual active speaker detection. In Proceedings of the 29th ACM International Conference on Multimedia, pages 3927โ3935, 2021.
- Thomaz et al. [2016] Andrea Thomaz, Guy Hoffman, Maya Cakmak, et al. Computational human-robot interaction. Foundations and Trendsยฎ in Robotics, 4(2-3):105โ223, 2016.
- Vaswani et al. [2017] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, ลukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017.
- Wang et al. [2018] Quan Wang, Carlton Downey, Li Wan, Philip Andrew Mansfield, and Ignacio Lopz Moreno. Speaker diarization with lstm. In 2018 IEEE International conference on acoustics, speech and signal processing (ICASSP), pages 5239โ5243. IEEE, 2018.
- Wang et al. [2019] Yue Wang, Yongbin Sun, Ziwei Liu, Sanjay E Sarma, Michael M Bronstein, and Justin M Solomon. Dynamic graph cnn for learning on point clouds. Acm Transactions On Graphics (tog), 38(5):1โ12, 2019.
- Wuerkaixi et al. [2022] Abudukelimu Wuerkaixi, You Zhang, Zhiyao Duan, and Changshui Zhang. Rethinking audio-visual synchronization for active speaker detection. In 2022 IEEE 32nd International Workshop on Machine Learning for Signal Processing (MLSP), pages 01โ06. IEEE, 2022.
- Xie et al. [2017] Saining Xie, Ross Girshick, Piotr Dollรกr, Zhuowen Tu, and Kaiming He. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1492โ1500, 2017.
- Xiong et al. [2022] Junwen Xiong, Yu Zhou, Peng Zhang, Lei Xie, Wei Huang, and Yufei Zha. Look&listen: Multi-modal correlation learning for active speaker detection and speech enhancement. IEEE Transactions on Multimedia, 2022.
- Xu et al. [2021] Eric Zhongcong Xu, Zeyang Song, Chao Feng, Mang Ye, and Mike Zheng Shou. Ava-avd: Audio-visual speaker diarization in the wild. arXiv preprint arXiv:2111.14448, 2021.
- Yang et al. [2020] Zhuoqian Yang, Wentao Zhu, Wayne Wu, Chen Qian, Qiang Zhou, Bolei Zhou, and Chen Change Loy. Transmomo: Invariance-driven unsupervised video motion retargeting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5306โ5315, 2020.
- Zhang et al. [2021] Yuanhang Zhang, Susan Liang, Shuang Yang, Xiao Liu, Zhongqin Wu, Shiguang Shan, and Xilin Chen. Unicon: Unified context network for robust active speaker detection. In Proceedings of the 29th ACM International Conference on Multimedia, pages 3964โ3972, 2021.
Xet Storage Details
- Size:
- 82.2 kB
- Xet hash:
- 1688d98d4cc82628e61581fdb22e655f04134cd154a75a11e707756aab35bd76
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.
