This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Spatial-Temporal Knowledge-Embedded Transformer for Video Scene Graph Generation

Tao Pu, Tianshui Chen, Hefeng Wu, Yongyi Lu, Liang Lin This work was supported in part by National Key R&D Program of China under Grant No. 2021ZD0111601, National Natural Science Foundation of China (NSFC) under Grant No. 62206060, 62272494, and 61836012, GuangDong Basic and Applied Basic Research Foundation under Grant No. SL2022A04J01626, 2023A1515012845, and 2023A1515011374, and Fundamental Research Funds for the Central Universities, Sun Yat-sen University, under Grant No. 23ptpy111, GuangDong Province Key Laboratory of Information Security Technology. (Corresponding author: Tianshui Chen)Tao Pu, Hefeng Wu, and Liang Lin are with the School of Computer Science and Engineering, Sun Yat-Sen University, Guangzhou 510006, China. (e-mail: putao3@mail2.sysu.edu.cn, wuhefeng@mail.sysu.edu.cn, linliang@ieee.org)Tianshui Chen and Yongyi Lu are with the School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, China. (e-mail: tianshuichen@gmail.com, yylu1989@gmail.com).
Abstract

Video scene graph generation (VidSGG) aims to identify objects in visual scenes and infer their relationships for a given video. It requires not only a comprehensive understanding of each object scattered on the whole scene but also a deep dive into their temporal motions and interactions. Inherently, object pairs and their relationships enjoy spatial co-occurrence correlations within each image and temporal consistency/transition correlations across different images, which can serve as prior knowledge to facilitate VidSGG model learning and inference. In this work, we propose a spatial-temporal knowledge-embedded transformer (STKET) that incorporates the prior spatial-temporal knowledge into the multi-head cross-attention mechanism to learn more representative relationship representations. Specifically, we first learn spatial co-occurrence and temporal transition correlations in a statistical manner. Then, we design spatial and temporal knowledge-embedded layers that introduce the multi-head cross-attention mechanism to fully explore the interaction between visual representation and the knowledge to generate spatial- and temporal-embedded representations, respectively. Finally, we aggregate these representations for each subject-object pair to predict the final semantic labels and their relationships. Extensive experiments show that STKET outperforms current competing algorithms by a large margin, e.g., improving the mR@50 by 8.1%, 4.7%, and 2.1% on different settings over current algorithms.

Index Terms:
Video Scene Graph Generation, Spatial-Temporal Knowledge Learning, Vision and Language
Refer to caption
Figure 1: Qualitative results of our proposed STKET. The green color indicates visual relationships that emerge in the current image. The gray color indicates visual relationships that disappeared in the current image.

I Introduction

Scene graph generation (SGG) [1, 2, 3] aims to depict a visual scene as a structured graph, where nodes correspond to semantic objects and edges refer to the corresponding relationships. It is treated as a promising approach to bridge the significant gap between vision and natural language domains, due to its capacity to represent accurately the semantics of visual contents in the holistic scene. Recently, lots of efforts have been dedicated to demonstrating its efficacy in numerous visual reasoning tasks, including video action recognition [4, 5, 6], video segmentation [7, 8, 9], video captioning [10, 11, 12, 13], and video question answering [14, 15, 16, 17]. However, the intricate spatial interactions among semantic objects pose a significant challenge to learning visual relationships in static images [18, 19, 20]. An even more substantial challenge exists in the task of generation scene graphs in videos (VidSGG) [21, 22, 23], as it further requires exploring temporal motions and interactions, hence making VidSGG a formidable yet unresolved task.

Current works [24, 25] primarily focus on aggregating object-level visual information from spatial and temporal perspectives to learn relationship representation for VidSGG. In contrast, humans rely on not only visual cues but also accumulated prior knowledge of spatial-temporal correlations to discern ambiguous visual relationships. As illustrated in Figure 1, prior knowledge consists of two aspects. 1) Spatial co-occurrence correlations: the relationship between certain object categories tends toward specific interactions. Given the subject and object of person and sandwich, their relationship tends to be eating or holding instead of wiping. 2) Temporal consistency/transition correlations: the relationship of a given pair tend to be consistent across continuous video clip or have a high probability of transit to another specific relationship. Given the subject and object of person and cup with the relationship of holding in the current frame, it is likely to keep the holding relationship or transit to the relationship of drinking. Consequently, integrating these correlations can effectively regularize spatial prediction space within each image and sequential variation space across temporal frames, thereby reducing ambiguous predictions.

In this work, we find that initializing spatial-temporal knowledge embeddings with statistical correlations can better guide model learning spatial-temporal correlations, and the multi-head cross-attention mechanisms can better integrate spatial-temporal correlations with visual information. To this end, we propose a novel spatial-temporal knowledge-embedded transformer (STKET), which introduces multi-head cross-attention layers to incorporate the prior spatial-temporal knowledge for aggregating spatial-temporal contextual information. Specifically, we first initialize the spatial co-occurrence and temporal transition correlations via statistical matrices from the training set and then embed these correlations into learnable spatial-temporal knowledge representations. Then, we design a spatial-knowledge embedded layer to exploit the within-image co-occurrence correlation to guide aggregating spatial contextual information and a temporal-knowledge embedded layer to incorporate cross-image transition correlations to help extract temporal contextual information, which can generate spatial- and temporal-embedded relationship representations, respectively. Finally, we aggregate both spatial- and temporal-embedded relationship representations of each object pair to predict the semantic labels and relationships. Compared with current leading competitors, STKET enjoys two appealing advantages: 1) Integrating these correlations can help to better aggregate spatial and temporal contextual information and thus learn more representative relationship representation to facilitate VidSGG. 2) Incorporating these correlations can effectively regularize relationship prediction, which can evidently reduce the dependencies on training samples and thus dramatically improve the performance, especially for the relationships with limited samples.

The contributions of this work can be summarized in three folds. First, we propose a spatial-temporal knowledge-embedded transformer (STKET) that incorporates spatial co-occurrence and temporal transition correlations to guide aggregating spatial and temporal contextual information, which, on the one hand, facilitates learning more representative relationship representation and, on the other hand, regularizes predication space to reduce the dependencies on training samples. To our knowledge, this is the first attempt to explicitly integrate spatial and temporal knowledge to promote VidSGG. Second, we introduce unified multi-head cross-attention mechanisms to integrate the spatial and temporal correlations via the spatial and temporal knowledge-embedded layers, respectively. Finally, we conduct experiments on the Action Genome dataset [26] to demonstrate the superiority of STKET. It obtains an obvious performance improvement over current state-of-the-art algorithms, especially for the relationships with limited samples, e.g., with the 9.98%-23.98% R50 improvement for the top-10 least frequency relationships compared with the previous best-performing algorithm. Codes are available at https://github.com/HCPLab-SYSU/STKET.

II Related Work

II-A Image Scene Graph Generation

Over the past decade, scene graph generation (SGG) [27, 28, 29, 30] has attracted considerable interest across various communities due to its ability to precisely present the semantics of visual contents of complex visual scenes. It aims to identify all objects and their visual relationships within an image, necessitating visual perception and natural language understanding. Hence, much effort has been invested in aligning visual and semantic spaces for relationship representation learning [31, 32, 33]. Further research has highlighted the importance of each subject-object pair for inferring ambiguous relationships. Xu et al. [1] propose to iteratively refine predictions by passing contextual messages, and Zellers et al. [34] capture a global context using a bidirectional LSTM. Their impressive performance demonstrates that spatial context is crucial for recognizing visual relationships. Therefore, subsequent research underlines the role of spatial context in generating scene graphs. To this end, many works adopt graph convolutional networks [35] or similar architectures to pass messages among different objects. Chen et al. [36] built a graph that associates detected objects according to these statistical correlations and employs it to learn the context among different objects to regularize prediction. Tang et al. [37] design a dynamic tree structure to encode the context among different object regions efficiently. With the impressive progress of transformer [38], more and more works propose utilizing this kind of model to learn more representative features from spatial context. Cong et al. [39] introduce an encoder-decoder architecture to reason about the visual feature context and visual relationships, using different types of attention mechanisms with coupled subject and object queries. Kundu et al. [40] propose contextualized relational reasoning using a two-stage transformer-based architecture for effective reasoning over cluttered, complex semantic structures. Despite impressive progress on static images, leading algorithms may suffer significant performance drops when applied to recognize dynamic visual relationships in videos because it requires an in-depth exploration of temporal consistency/transition correlations across different images.

II-B Video Scene Graph Generation

Building on successfully exploring the spatial context within images, researchers have explored the spatial context and temporal correlation simultaneously in video scene graph generation. With the advent of ImageNet-VidVRD [21], a benchmark of video visual relation detection, numerous approaches [22, 41, 42, 43, 23] have been proposed to employ object-tracking mechanisms to dive into temporal correlations among different image frames. Qian et al. [22] propose to use a graph convolution network to pass messages and conduct reasoning in the fully-connected spatial-temporal graphs. Similarly, Tsai et al. [41] design a gated spatiotemporal energy graph that exploits the statistical dependency between relational entities spatially and temporally. Recently, Teng et al. [43] propose a new detect-to-track paradigm by decoupling the context modeling for relation prediction from the complicated low-level entity tracking. Zheng et al. [23] propose a unified one-stage model that exploits static queries and recurrent queries to enable efficient object pair tracking with spatiotemporal contexts. However, introducing object-tracking models results in high computational cost and memory consumption and quickly overwhelms the valuable information due to many irrelevant frames, leading to sub-optimal performance.

An alternative line of research tries to address the task of video scene graph generation based on detected object proposals. Compared with tracking-based VidSGG algorithms, this kind of method focuses on modeling the temporal context in the narrow sliding window, avoiding the prediction shift resulting from the inconsistency among tracking proposals. Recently, Cong et al. [25] propose a strong baseline that adopts a spatial encoder and a temporal decoder to extract implicitly spatial-temporal contexts. This work demonstrates that the temporal correlation is crucial for inferring dynamic visual relationships. Li et al. [24] propose a novel anticipatory pre-training paradigm to model the temporal correlation implicitly. Wang et al. [44] propose to explore temporal continuity by extracting the entire co-occurrence patterns. Kumar et al. [45] further propose to spatially and temporally localize subjects and objects connected via an unseen predicate with the help of only a few support set videos sharing the common predicate. These works underscore the critical role of temporal correlation in inferring dynamic visual relationships but overlook the essential prior knowledge of spatial-temporal correlations. Differently, we propose to incorporate spatial-temporal knowledge with multi-head cross-attention layers to learn more representative and discriminative feature representation to facilitate VidSGG.

Notably, exploiting an object detector to facilitate video understanding is common. Recently, Wang et al. [46] propose a novel SAOA framework to introduce the spatial location provided by the detector for egocentric action recognition, which aims to reason the interaction between humans and objects from an egocentric perspective. Although aligning local object features and location proposals to capture the spatial context, the SAOA framework ignores the crucial temporal correlation across continuous frames, which is the main distinction between it and our proposed STKET.

Refer to caption
Figure 2: An overall illustration of the proposed STKET framework (left) and its corresponding outputs (right). It first exploits spatial and temporal knowledge-embedded layers to incorporate the spatial co-occurrence and temporal transition correlations into the multi-head cross-attention mechanism to learn spatial- and temporal-embedded representations. Then, it employs a spatial-temporal aggregation module to aggregate these representations for each object pair to predict the final semantic labels and relationships.

II-C Knowledge Representation Learning

Recent advances in deep learning have allowed neural networks to learn potent representations from raw training data for various tasks [47, 38, 48, 49]. However, solely using these vanilla networks may achieve poor performance, especially in weakly supervised learning [50, 51] and domain adaption [52]. To address this challenge, lots of efforts have been made to integrate domain prior knowledge into deep representation learning, resulting in remarkable progress in numerous computer vision tasks, such as few-shot image recognition, facial expression recognition, visual navigation, scene graph generation [53, 54, 55, 56, 57, 36]. Specifically, Peng et al. [53] propose a novel knowledge transfer network that jointly incorporates visual feature learning, knowledge inferring, and classifier learning to fully explore prior knowledge in the few-shot task. Similarly, Chen et al. [54] propose a knowledge-guided graph routing framework, which unifies prior knowledge of statistical label correlations with deep neural networks for multi-label few-shot learning. Pu et al. [55] introduce the prior knowledge between action unit and facial expression to facilitate facial expression recognition. Yang et al. [57] incorporate the prior semantic knowledge into a deep reinforcement learning framework to address the semantic navigation task. Chen et al. [36] incorporate statistical correlations into deep neural networks to facilitate scene graph generation, in which these statistical correlations between object pairs and their relationships can effectively regularize semantic space and make prediction less ambiguous. However, most of these works merely consider prior spatial knowledge of statistic images, and VidSGG involves spatial-temporal contextual information. In this work, we propose to learn spatial-temporal knowledge and incorporate it into the multi-head cross-attention mechanism to learn more representative relationship representations to facilitate VidSGG.

III Method

Overview. In this section, we first introduce the preliminary of video scene graph generation and then describe the process of extracting spatial co-occurrence and temporal transition correlations from the dataset. Finally, the details of our proposed framework, Spatial-Temporal Knowledge-Embedded Transformer (STKET), are given.

III-A Preliminary

Notation. Given a video V={I1,I2,,IT}V=\{I_{1},I_{2},...,I_{T}\}, VidSGG aims to detect all visual relationships between objects, presented as a triplet <subject, predicate, object >, in each frame to generate a scene graph sequence G={G1,G2,,GT}G=\{G_{1},G_{2},...,G_{T}\}, where GtG_{t} is the corresponding scene graph of the frame ItI_{t}. Specifically, define Gt={Bt,Ot,Rt}G_{t}=\{B_{t},O_{t},R_{t}\}, where Bt={bt1,bt2,,btN(t)}B_{t}=\{b_{t}^{1},b_{t}^{2},...,b_{t}^{N(t)}\}, Ot={ot1,ot2,,otN(t)}O_{t}=\{o_{t}^{1},o_{t}^{2},...,o_{t}^{N(t)}\} and Rt={rt1,rt2,,rtK(t)}R_{t}=\{r_{t}^{1},r_{t}^{2},...,r_{t}^{K(t)}\} indicate the bounding box set, the object set and the predicate set, respectively. In the frame ItI_{t}, N(t)N(t) is the number of objects, and K(t)K(t) is the number of relationships between all objects.

Relationship Representation. For the frame ItI_{t}, we employ Faster R-CNN [58] to provides visual feature representation {𝐯t1,,𝐯tN(t)}\{\mathbf{v}^{1}_{t},...,\mathbf{v}^{N(t)}_{t}\}, bounding boxes {𝐛t1,,𝐛tN(t)}\{\mathbf{b}^{1}_{t},...,\mathbf{b}^{N(t)}_{t}\} and object category distribution {𝐝t1,,𝐝tN(t)}\{\mathbf{d}^{1}_{t},...,\mathbf{d}^{N(t)}_{t}\} of object proposals. The relationship representation 𝐱tk\mathbf{x}^{k}_{t} of the relationship between the ii-th and jj-th object proposals contains visual appearances, spatial information, and semantic embeddings, which can be formulated as

𝐱tk=<fs(𝐯ti),fo(𝐯tj),fu(φ(𝐮ti,jfbox(𝐛ti,𝒃tj))),𝐬ji,𝐬tj>,\mathbf{x}^{k}_{t}=<f_{s}(\mathbf{v}^{i}_{t}),f_{o}(\mathbf{v}^{j}_{t}),f_{u}(\varphi(\mathbf{u}^{i,j}_{t}\oplus f_{box}(\mathbf{b}^{i}_{t},\bm{b}^{j}_{t}))),\mathbf{s}^{i}_{j},\mathbf{s}^{j}_{t}>,

where <,><,> is concatenation operation, φ\varphi is flattening operation and \oplus is element-wise addition. fs,fof_{s},f_{o} are implemented respectively by one fully-connected layer which maps a 2048-dimension vector to a 512-dimension vector, and fuf_{u} is implemented by one fully-connected layer which maps a 12544-dimension vector to a 512-dimension vector. 𝐮ti,j256×7×7\mathbf{u}^{i,j}_{t}\in\mathbb{R}^{256\times 7\times 7} is the feature map of the corresponding union box generated by RoIAlign [59] while fboxf_{box} is the function transforming the bounding boxes of subject and object to an entire feature with the same shape as 𝐮ti,j\mathbf{u}^{i,j}_{t}. The semantic embedding vectors 𝐬ti,𝐬tj200\mathbf{s}^{i}_{t},\mathbf{s}^{j}_{t}\in\mathbb{R}^{200} are determined by the object category distribution of subject and object. For brevity, we denote all relationship representations in the frame ItI_{t} by 𝐗t={𝐱t1,,𝐱tK(t)}\mathbf{X}_{t}=\{\mathbf{x}^{1}_{t},...,\mathbf{x}^{K(t)}_{t}\}.

III-B Spatial-Temporal Knowledge Representation

When inferring visual relationships, humans leverage not only visual cues but also accumulated prior knowledge [60]. This approach has been validated on various vision tasks [61, 54, 56]. Inspired by this, we propose to distill the prior spatial-temporal knowledge directly from the training set for facilitating the VidSGG task. Specifically, we learn spatial co-occurrence and temporal transition correlations in a statistical manner, as shown in Figure 4.

Refer to caption
(a)  
Refer to caption
(b)  
Figure 3: Illustration of learning (a) spatial prior knowledge representation and (b) temporal prior knowledge representation. For better readability, we take the object category pair of person and cup as an example.

Spatial Prior Knowledge. Between different object category pairs, a large difference exists in the spatial co-occurrence correlation of its relationships. To account for this, we construct a spatial co-occurrence matrix Ei,jCE^{i,j}\in\mathbb{R}^{C} for ii-th object category and jj-th object category by measuring the frequency of each predicate in the relation set of this pair. CC is the total number of types of relationships. For example, let us consider the object category pair of person and cup (assume their indices are ii and jj, respectively). At each frame, if both objects are present, the total number of co-occurrences of this pair Ni,jN^{i,j} is incremented by 1. Next, if their predicates contain holding (assume the index is xx), then exi,je^{i,j}_{x} is incremented by 1. Finally, between person and cup, the spatial co-occurrence probability of each predicate is calculated as:

exi,j=exi,j/Ni,j,x{1,,C}.e^{i,j}_{x}=e^{i,j}_{x}/N^{i,j},\quad x\in\{1,...,C\}. (1)

To distill spatial co-occurrence correlations, the model learns the spatial knowledge embedding based on the spatial co-occurrence matrix. Specifically, the model generates the corresponding spatial knowledge embedding 𝐬tk\mathbf{s}^{k}_{t} for kk-th relationship representation in the frame ItI_{t}, the generating process can be formulated as

𝐬tk=fspa(Es(k,t),o(k,t)),\mathbf{s}^{k}_{t}=f_{spa}(E^{s(k,t),o(k,t)}), (2)

where s(k,t)s(k,t) and o(k,t)o(k,t) denote the category of subject and object of kk-th relationship representation in the frame ItI_{t}, fspa()f_{spa}(\cdot) is implemented by four fully-connected layers which map a CC-dimension vector to a 1936-dimension vector. For brevity, we denote the corresponding spatial knowledge embeddings for all object pairs in the frame ItI_{t} by 𝐒t={𝐬t1,,𝐬tK(t)}\mathbf{S}_{t}=\{\mathbf{s}^{1}_{t},...,\mathbf{s}^{K(t)}_{t}\}.

Since the distribution of real-world relationships is seriously unbalanced, directly utilizing these spatial knowledge embeddings may degrade the model performance on the less frequent relationships. Thus, we train these embeddings to predict predicates by using the binary cross-entropy loss as the objective function:

spk=t=1Tk=1K(t)(f(𝐬tk),𝐲tk),\mathcal{L}_{spk}=\sum^{T}_{t=1}\sum^{K(t)}_{k=1}\ell(f(\mathbf{s}^{k}_{t}),\mathbf{y}^{k}_{t}), (3)

where (,)\ell(\cdot,\cdot) denote the binary cross-entropy loss function, f()f(\cdot) denote the predicate classifier, 𝐲tk\mathbf{y}^{k}_{t} denote ground-truth predicate labels of kk-th relationships in the frame ItI_{t}.

Refer to caption
Figure 4: An example of building spatial co-occurrence and temporal transition matrix. Given the video frames and their corresponding relation annotations (left), we first increase the spatial co-occurrence probability of the corresponding predicates (i.e., “hold” and “drink from”). Then, for any two consecutive frames, we take the predicate in the previous and current frames as the row and column index separately and increase the temporal transition probability of the corresponding transition pair (i.e., “hold \rightarrow drink from”).

Temporal Prior Knowledge. In daily life, the interaction between people and objects is characterized by temporal transitions. To identify relationships at different stages, we construct a temporal transition matrix E^i,jC×C\hat{E}^{i,j}\in\mathbb{R}^{C\times C} for ii-th object category and jj-th object category. CC is the total number of types of relationships. For instance, let us consider the object category pair of person and cup as an example (assume the indices are ii and jj, respectively). If the contacting predicate between person and cup in the previous frame is holding (assume the index is xx) and in the current frame is drinking (assume the index is yy), then E^x,yi,j\hat{E}^{i,j}_{x,y} is incremented by 1. Finally, between person and cup, the temporal transition probability of predicates is calculated as:

e^x,yi,j=e^x,yi,j/exi,j,x,y{1,,C}.\hat{e}^{i,j}_{x,y}=\hat{e}^{i,j}_{x,y}/e^{i,j}_{x},\quad x,y\in\{1,...,C\}. (4)

To explore temporal transition correlations, the model learns the corresponding temporal knowledge embedding based on the temporal transition matrix. Specifically, given the predicted object labels and relation labels, the model generates the corresponding temporal knowledge embedding 𝐭tk\mathbf{t}^{k}_{t} for kk-th relationship representation in the frame ItI_{t}, the generating process can be formulated as

𝐭tk=ftem(E^r(k,t)s(k,t),o(k,t)),\mathbf{t}^{k}_{t}=f_{tem}(\hat{E}^{s(k,t),o(k,t)}_{r(k,t)}), (5)

where s(k,t)s(k,t) and o(k,t)o(k,t) denote the category of subject and object of kk-th relationship representation in the frame ItI_{t}, r(k,t)r(k,t) denotes the predicate of kk-th relationship representation in the frame ItI_{t}, ftem()f_{tem}(\cdot) is implemented by four fully-connected layers which map a CC-dimension vector to a 1936-dimension vector. Specifically, we utilize the spatial contextualized representation to coarsely predict the relation labels. For brevity, we denote the corresponding temporal knowledge embeddings for all object pairs in the frame ItI_{t} by 𝐓t={𝐭t1,,𝐭tK(t)}\mathbf{T}_{t}=\{\mathbf{t}^{1}_{t},...,\mathbf{t}^{K(t)}_{t}\}.

Since complicated predicate transition exists for different object category pairs, the temporal knowledge embedding may contain inaccurate temporal correlation, resulting in sub-optimal performance. Therefore, we train these embeddings to predict predicates at the next frame by using the binary cross-entropy loss as the objective function:

tpk=t=1T1i=1K(t)(f(𝐭tk),𝐲t+1k),\mathcal{L}_{tpk}=\sum^{T-1}_{t=1}\sum^{K(t)}_{i=1}\ell(f(\mathbf{t}^{k}_{t}),\mathbf{y}^{k}_{t+1}), (6)

where we assume kk-th relationship representation in the frame ItI_{t} and frame It+1I_{t+1} belong the same subject-object pair for easily understanding.

III-C Knowledge-Embedded Attention Layer

Between object pairs and their relationships, there are apparent spatial co-occurrence correlations within each image and strong temporal transition correlations across different images. Thus, we propose incorporating spatial-temporal knowledge into the multi-head cross-attention mechanism to learn spatial- and temporal-embedded representations.

Spatial knowledge often encapsulates information about positions, distances, and relationships among entities. On the other hand, temporal knowledge concerns the sequence, duration, and intervals between actions. Given their unique properties, treating them separately allows specialized modeling to capture the inherent patterns more accurately. Therefore, we design spatial and temporal knowledge-embedded layers that thoroughly explore the interaction between visual representation and spatial-temporal knowledge.

Exploring Spatial Context. As shown in Figure 2, we first employ the spatial knowledge-embedded layers (SKEL) to explore spatial co-occurrence correlations within each image. Specifically, we take all relationship representations of the current frame ItI_{t} as the input, i.e., 𝐅0,tS=𝐗t={𝐱t1,,𝐱tK(t)}\mathbf{F}^{S}_{0,t}=\mathbf{X}_{t}=\{\mathbf{x}^{1}_{t},...,\mathbf{x}^{K(t)}_{t}\}. Then, the SKEL incorporates corresponding spatial knowledge embeddings in the keys and queries to fuse the information from the relationship representation and its corresponding spatial prior. Since we stack NsN_{s} identical spatial knowledge-embedded layers, the input of the nn-th layer is the output of (n1)(n-1)-th layer. Thus, the queries 𝐐\mathbf{Q}, keys 𝐊\mathbf{K}, values 𝐕\mathbf{V} and output of the nn-th SKEL is presented as:

𝐐=𝐖Q𝐅(n1),tS+𝐒t,\displaystyle\mathbf{Q}=\mathbf{W}_{Q}\mathbf{F}^{S}_{(n-1),t}+\mathbf{S}_{t}, (7)
𝐊=𝐖K𝐅(n1),tS+𝐒t,\displaystyle\mathbf{K}=\mathbf{W}_{K}\mathbf{F}^{S}_{(n-1),t}+\mathbf{S}_{t}, (8)
𝐕=𝐖V𝐅(n1),tS,\displaystyle\mathbf{V}=\mathbf{W}_{V}\mathbf{F}^{S}_{(n-1),t}, (9)
𝐅n,tS=Attspa.(𝐐,𝐊,𝐕),\displaystyle\mathbf{F}^{S}_{n,t}=Att_{spa.}(\mathbf{Q},\mathbf{K},\mathbf{V}), (10)

where the 𝐖Q\mathbf{W}_{Q}, 𝐖K\mathbf{W}_{K} and 𝐖V\mathbf{W}_{V} denote the linear transformations, the Attspa.()Att_{spa.}(\cdot) denote cross-attention layer in the SKEL. For simplicity, we remove the subscript nn and denote the final output of the SKEL as 𝐅tS\mathbf{F}^{S}_{t}.

Parsing Temporal Correlation. As claimed in prior works [25, 24], the dynamic visual relation can be easily recognized with the given temporal information in the previous frame. Thus, we design the temporal knowledge-embedded layer (TKEL) to explore the temporal correlation in the current frame ItI_{t} and the previous frame It1I_{t-1}. At first, we adopt a sliding window over the sequence of spatial contextualized representation [𝐅1S,,𝐅TS][\mathbf{F}^{S}_{1},...,\mathbf{F}^{S}_{T}], and the input of the frame ItI_{t} in TKEL is presented as:

𝐅0,tT=[𝐅t1S,𝐅tS],t{2,,T}.\mathbf{F}^{T}_{0,t}=[\mathbf{F}^{S}_{t-1},\mathbf{F}^{S}_{t}],\quad t\in\{2,...,T\}. (11)

Then, TKEL incorporates corresponding spatial and temporal knowledge embeddings in the keys and queries to fuse the information from the relationship representation and its corresponding spatial and temporal prior. Since we stack NtN_{t} identical temporal knowledge-embedded layers, the input of the nn-th layer is the output of (n1)(n-1)-th layer. Thus, the queries 𝐐\mathbf{Q}, keys 𝐊\mathbf{K}, values 𝐕\mathbf{V} and output of the nn-th TKEL is presented as:

𝐐=𝐖Q𝐅(n1),tT+[𝐓t1,𝐒t]+𝐄f,\displaystyle\mathbf{Q}=\mathbf{W}_{Q}\mathbf{F}^{T}_{(n-1),t}+[\mathbf{T}_{t-1},\mathbf{S}_{t}]+\mathbf{E}_{f}, (12)
𝐊=𝐖K𝐅(n1),tT+[𝐓t1,𝐒t]+𝐄f,\displaystyle\mathbf{K}=\mathbf{W}_{K}\mathbf{F}^{T}_{(n-1),t}+[\mathbf{T}_{t-1},\mathbf{S}_{t}]+\mathbf{E}_{f}, (13)
𝐕=𝐖V𝐅(n1),tT,\displaystyle\mathbf{V}=\mathbf{W}_{V}\mathbf{F}^{T}_{(n-1),t}, (14)
𝐅n,tT=Atttem.(𝐐,𝐊,𝐕),\displaystyle\mathbf{F}^{T}_{n,t}=Att_{tem.}(\mathbf{Q},\mathbf{K},\mathbf{V}), (15)

where 𝐄f\mathbf{E}_{f} is the learned frame encoding vector, the Atttem.()Att_{tem.}(\cdot) denote cross-attention layer in TKEL and the 𝐖Q\mathbf{W}_{Q}, 𝐖K\mathbf{W}_{K} and 𝐖V\mathbf{W}_{V} denote the linear transformations. The intention of adding [𝐓t1,𝐒t][\mathbf{T}_{t-1},\mathbf{S}_{t}] into the queries and keys is to incorporate the temporal prior about the previous frame and the spatial prior about the current frame. In this way, the TKEL can effectively capture spatiotemporal context in the sliding window.

Considering that the relationships in a frame have various representations in different batches, we choose the earliest representation appearing in the sliding window. For simplicity, we remove the subscript nn and denote the final output of TKEL as 𝐅tT\mathbf{F}^{T}_{t}.

III-D Spatial-Temporal Aggregation Module

As aforementioned, SKEL explores the spatial co-occurrence correlations within each image, and TKEL explores the temporal transition correlations across different images. Though fully exploring the interaction between visual representation and spatial-temporal knowledge, these two layers generate spatial- and temporal-embedded representations, respectively. To explore the long-term context information, we further design the spatial-temporal aggregation (STA) module to aggregate these representations for each object pair to predict the final semantic labels and their relationships. It takes the spatial- and temporal-embedded relationship representations of the identical subject-object pair in different frames as the input. Specifically, we concatenate these representations of the same object pair to generate context representation:

𝐜tk=cat(𝐟k,tS,𝐟k,tT),\mathbf{c}^{k}_{t}=cat(\mathbf{f}^{S}_{k,t},\mathbf{f}^{T}_{k,t}), (16)

where 𝐟k,tS\mathbf{f}^{S}_{k,t} and 𝐟k,tT\mathbf{f}^{T}_{k,t} denote the spatial- and temporal-embedded representation for the kk-th relationship in the frame ItI_{t}. And then, to find the same subject-object pair in different frames, we adopt the predicted object label and the IoU (i.e., intersection over union) to match the same subject-object pairs detected in frames {Itτ+1,,It}\{I_{t-\tau+1},...,I_{t}\} (more details in the supplementary material). Thus, the input of kk-th relationship in the frame ItI_{t} in the spatial-temporal aggregation module is presented as

𝐟k,tC=[𝐜tτ+1k,,𝐜tk],\mathbf{f}^{C}_{k,t}=[\mathbf{c}^{k}_{t-\tau+1},...,\mathbf{c}^{k}_{t}], (17)

where we assume kk-th relationship representation in frame {Itτ+1,,It}\{I_{t-\tau+1},...,I_{t}\} belong same subject-object pair for easily understanding. And its corresponding output of the spatial-temporal aggregation module is presented as:

𝐐=𝐖Q𝐟k,tC+𝐄f,\displaystyle\mathbf{Q}=\mathbf{W}_{Q}\mathbf{f}^{C}_{k,t}+\mathbf{E}^{{}^{\prime}}_{f}, (18)
𝐊=𝐖K𝐟k,tC+𝐄f,\displaystyle\mathbf{K}=\mathbf{W}_{K}\mathbf{f}^{C}_{k,t}+\mathbf{E}^{{}^{\prime}}_{f}, (19)
𝐕=𝐖V𝐟k,tC,\displaystyle\mathbf{V}=\mathbf{W}_{V}\mathbf{f}^{C}_{k,t}, (20)
𝐟^k,tC=Att(𝐐,𝐊,𝐕),\displaystyle\hat{\mathbf{f}}^{C}_{k,t}=Att(\mathbf{Q},\mathbf{K},\mathbf{V}), (21)

where 𝐄f\mathbf{E}^{{}^{\prime}}_{f} is the learned frame encoding vector, the Att()Att(\cdot) denote the self-attention layer in the spatial-temporal aggregation module, and the 𝐖Q\mathbf{W}_{Q}, 𝐖K\mathbf{W}_{K} and 𝐖V\mathbf{W}_{V} denote the linear transformations. Considering the relationships in a frame have various representations in different batches, we choose the earliest representation appearing in the sliding window.

III-E Loss Function

In the real world, there exist different kinds of relationships between two objects at the same time. Thus, we introduce the binary cross-entropy loss as the objective function for predicate classification as follows:

(r)=p𝒫+log(ϕ(r,p))+q𝒫log(1ϕ(r,q)).\ell(r)=\sum_{p\in\mathcal{P}^{+}}log(\phi(r,p))+\sum_{q\in\mathcal{P}^{-}}log(1-\phi(r,q)). (22)

For a subject-object pair rr, 𝒫+\mathcal{P}^{+} are the annotated predicates, while 𝒫\mathcal{P}^{-} is the set of the predicates not in the annotation. ϕ(r,p)\phi(r,p) indicates the computed confidence score of the pp-th predicate.

During training, we adopt the binary cross-entropy loss as the objective function for supervising SKEL, TKEL, and STA and denote the corresponding loss by s(r)\ell^{s}(r), t(r)\ell^{t}(r) and c(r)\ell^{c}(r) respectively. Therefore, the final classification loss is defined as summing the three losses over all samples, formulated as

cls\displaystyle\mathcal{L}_{cls} =t=1Tr=1K(t)[s(r)+t(r)+c(r)].\displaystyle=\sum^{T}_{t=1}\sum^{K(t)}_{r=1}{[\ell^{s}(r)+\ell^{t}(r)+\ell^{c}(r)]}. (23)

Therefore, the total objective is formulated as:

=cls+spk+tpk.\mathcal{L}=\mathcal{L}_{cls}+\mathcal{L}_{spk}+\mathcal{L}_{tpk}. (24)

In all experiments, we set the loss weights of the three losses to be equal, primarily to ensure that each loss component has an equal contribution to the overall optimization objective.

TABLE I: Comparison with state-of-the-art scene graph generation methods on Action Genome [26].
Method Pred Cls SG Cls SG Gen Average Recall
R@10 R@20 R@50 R@10 R@20 R@50 R@10 R@20 R@50
KERN [36] 73.7 92.3 98.8 50.5 62.4 66.3 23.3 35.2 47.0 61.1
VCTREE [37] 75.5 92.9 99.3 52.4 62.0 65.1 23.9 35.3 46.8 61.5
ReIDN [27] 75.7 93.0 99.0 52.9 62.4 65.1 24.1 35.4 46.8 61.6
GPS-Net [28] 76.2 93.6 99.5 53.6 63.3 66.0 24.4 35.7 47.3 62.2
STTran [25] 77.9 94.2 99.1 54.0 63.7 66.4 24.6 36.2 48.8 62.8
TPI [44] 78.2 94.6 99.1 54.4 64.3 67.1 25.2 36.9 49.3 63.2
APT [24] 78.5 95.1 99.2 55.1 65.1 68.7 25.7 37.9 50.1 63.9
Ours 82.6 96.3 99.9 57.1 65.3 67.1 27.9 38.8 50.6 65.1
TABLE II: Comparison with state-of-the-art scene graph generation methods on Action Genome [26].
Method Pred Cls SG Cls SG Gen Mean
mR@10 mR@20 mR@50 mR@10 mR@20 mR@50 mR@10 mR@20 mR@50
KERN [36] 12.3 14.7 18.4 8.1 9.3 10.3 4.4 5.3 8.2 10.1
STTran [25] 15.4 19.6 21.2 10.7 13.1 14.0 5.0 7.3 10.2 12.9
TPI [44] 16.3 20.8 22.4 11.1 13.5 14.2 5.3 6.5 9.7 13.3
APT [24] 17.2 22.5 25.6 11.5 13.8 14.6 5.8 6.9 10.1 14.2
Ours 20.1 27.3 33.7 15.4 17.2 19.3 7.4 9.7 12.2 18.0

IV Experiments

IV-A Experiment Setting

Dataset As the most widely used dataset for evaluating video scene graph generation, the Action Genome [26] contains 476,229 bounding boxes of 35 object classes (without the person) and 1,715,568 instances of 26 relationship classes annotated for 234, 253 frames. These 26 relationships are subdivided into three different types: (1) attention, (2) spatial, and (3) contact whose number of categories are 3, 6, and 17, respectively. In all experiments, we use the same training and testing split in previous works [25, 24].

Task Following prior arts [36, 43, 25], we evaluate our proposed method and other state-of-the-art methods under three kinds of experiment setups: Predicate Classification (Pred Cls): predict the predicates of object pairs with given ground truth bounding boxes and category labels. Scene Graph Classification (SG Cls): predict both the predicates and the category labels of objects with given ground-truth bounding boxes. Scene Graph Generation (SG Gen): simultaneously detects objects appearing in the image and predicts the predicates of each object pair. In the SG Gen, an object bounding box is considered to be correctly detected only if the predicted object bounding box has at least 0.5 IoU (Intersection over Union) overlap with the ground-truth object bounding box. We use No Constraint strategy of generating a scene graph to evaluate the models. This strategy allows each subject-object pair to have multiple predicates simultaneously.

Evaluation Metric All tasks are evaluated with the Recall@K (short as R@KK) metric (KK = [10, 20, 50]), which measures the ratio of correct instances among the top-KK predicted instances with the highest confidence. We also report the results by using the mean Recall@KK (short as mR@KK) metric that averages R@KK over all relationships.

Training Details Following previous works [25, 24], we adopt the Faster RCNN [58] with a ResNet-101 [47] backbone as the object detector. We first train the detector on the training set of Action Genome [26] and get 24.6 mAP at 0.5 IoU with COCO metrics. The detector is applied to all baselines for fair comparisons. The parameters of the object detector (the object classifier excluded) are fixed when training scene graph generation models. Per-class non-maximal suppression at 0.4 IoU (Intersection over Union) is applied to reduce region proposals provided by RPN.

We use an AdamW [62] optimizer with initial learning rate 2e52e^{-5} and batch size 1 to train our model. Moreover, gradient clipping is applied with a maximal norm of 5. All experiments are implemented by PyTorch [63]. In the spatial-temporal aggregation module, we set the size of the sliding window τ\tau to 4. In KEAL, we stack two identical spatial knowledge-embedded layers to explore spatial co-occurrence correlations and then stack two identical temporal knowledge-embedded layers to temporal transition correlations. The cross-attention and self-attention layers in our proposed framework have 8 heads with d=1936d=1936 and dropout=0.1dropout=0.1. In SKEL and TKEL, the 19361936-dimension input is projected to 20482048-dimension by the feed-forward network, then projected to 19361936-dimension again after ReLU activation. In STA, the 38723872-dimension input is projected to 19361936-dimension.

Frame Encoding Unlike the SKEL, the TKEL adopts a sliding window over the sequence of spatial contextualized representation as its input. Therefore, we need to introduce the learned frame encoding EfE_{f} into the TKLE to help it fully understand the temporal dependencies among different frames. Specifically, we construct the frame encodings EfE_{f} using learned embedding parameters. Specifically, Ef=[e1,e2]E_{f}=[e_{1},e_{2}], where e1e_{1} and e21936e_{2}\in\mathbb{R}^{1936} are the learned vectors. Similarly, we also introduce frame encodings EfE^{{}^{\prime}}_{f} in STA, where Ef=[e1,,eτ]E^{{}^{\prime}}_{f}=[e^{{}^{\prime}}_{1},...,e^{{}^{\prime}}_{\tau}], and each encoding is the learned vectors with a length of 38723872.

Refer to caption
Figure 5: Qualitative results in SG Gen task with top-10 confident predictions. The blue and red colors indicate correct relationships and objects, respectively. The gray colors indicate wrong relationships and objects. For fairness, the same object detector is used for all methods.

Pair Tracking As described in the manuscript, STA takes the spatial- and temporal-embedded representation of the same subject-object pair in different frames as input. We first use the predicted object labels to distinguish different pairs to match the subject-object pairs detected in different frames. If multiple entities of the same category exist, we calculate the intersection over union (IoU) between the two objects across different images to match the subject-object pair. Specifically, we compute the IoU between the bounding box of the target object in the previous frame and that of each object with the same category label in the current frame. If the IoU is higher than 0.8, we consider them to be the same object. We choose the one with the highest IoU if there are multiple candidates.

IV-B Comparison with State-Of-The-Art Methods

To evaluate the effectiveness of STKET, we compare it with existing state-of-the-art VidSGG algorithms, including STTran [25], TPI [44], and APT [24]. Previous works [25] also adapt image-based SGG algorithms to address the VidSGG task by applying the inference process to each frame. We also follow these works to include the image-based SGG algorithms for more comprehensive evaluations and comparisons, including VCTREE [37], KERN [36], ReIDN [27], GPS-Net [28].

We first present the comparison on R@K in Table I. As shown, recent video-based algorithms (e.g., STTran, TPI, and APT) obtain quite a marginal improvement over the image-based SGG algorithms as they further introduce temporal contextual information. By introducing spatial-temporal prior knowledge to guide aggregating spatial and temporal contextual information, the proposed STKET framework consistently outperforms in nearly all settings. For example, it improves the R@10 from 78.5% to 82.6%, 55.1% to 57.1%, and 25.7% to 27.9% on the three tasks, with the improvement of 4.1%, 2.0% and 2.2%, respectively. It also obtains similar improvement on R@20 and R@50 metrics.

Refer to caption
(a)  
Refer to caption
(b)  
Figure 6: (a) The distribution of different relationships on Action Genome [26]. (b) The R@50 results in Pred Cls task of our method, STTran and APT on Action Genome [26].

Considering the mR@K metric offers a better performance measure under uneven distribution [36], we also present the comparison results on this metric. As presented in Table II, STKET obtains even more significant improvement, enhancing mR@50 by 8.1%, 4.7%, and 2.1% compared with the current best-performing APT algorithm. To highlight the necessity of introducing the mR@K metric, we present the distribution across different relationships on the AG dataset in Figure 6a. As depicted, the distribution of relationships is exceedingly long-tailed, in which the top-10 most frequent relationships occupy 90.9% samples while the top-10 least frequency relationships merely occupy 2.1%. In Figure 6b, we further provide the detailed performance of each relationship to understand the performance variance across different relationships better. Evidently, current algorithms like STTran and APT deliver competitive performance for relationships with abundant training samples (e.g., “in front of”, “not looking at”) but falter significantly for relationships with limited training samples (e.g., “wiping”, “twisting”). In contrast, for the top-10 least frequent relationships, STKET improves the R@50 from 9.98% to 23.98% compared to the second-best APT and from 12.10% to 30.65% compared to the baseline STTran. These comparisons demonstrate STKET can effectively regularize VidSGG training and thus reduce the dependencies on training samples by explicitly incorporating spatial-temporal prior knowledge.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Figure 7: Top: Relationship transition matrix of “man-tv” (left) and “man-cup” (right). For brevity, we print the relation index instead of the relation label (see the corresponding labels below figure, i.e., 0 denotes “looking at” and 25 denotes “writing on”). Bottom: The entropy of spatial-temporal knowledge across different predicates.

In Figure 5, we visualize the qualitative results of our method and current leading VidSGG methods (i.e., STTran and APT). The results show that existing state-of-the-art VidSGG algorithms tend to predict numerous false-positive relationships, while our method successfully predicts nearly all relationships with high accuracy. This highlights the strength of integrating spatial-temporal knowledge in discerning dynamic relationships, especially within intricate interactions. For instance, across all frames, our STKET accurately identifies the attentional relationship among person, cup, and laptop, where current leading VidSGG algorithms stumble. It is also noteworthy that our method not only performs well on high-frequency relationships but also on low-frequency relationships. For instance, in the last frame, our method correctly predicts the predicate of “drink from” whose samples merely occupy 0.355% of the total samples, while other methods miss it. This success is largely due to our method’s explicit incorporation of spatial-temporal correlations, which helps to lessen the dependency on training samples significantly.

IV-C Ablative Study

In this section, we conduct comprehensive experiments to analyze the actual contribution of each crucial component. Here, we mainly present the R@10, R@20, mR@10, and mR@20 on Pred Cls and SG Cls as they can better describe the performance.

IV-C1 Analysis of STKET

As aforementioned, STKET integrates spatial-temporal knowledge into the multi-head cross-attention mechanism to aggregate spatial and temporal contextual information. In this way, it effectively regularizes spatial prediction space within each image and sequential variation space across temporal frames, thereby reducing ambiguous predictions. To verify the effectiveness of exploring these spatial-temporal correlations, we implement the baseline STTran method for comparison purposes, setting the layers of its spatial encoder and temporal decoder to two for fair comparisons. As shown in Table III, the baseline STTran method obtains the R@10 and R@20 values of 77.8% and 94.1% on Pred Cls and the R@10 and R@20 values of 53.8% and 63.6% on SG Cls. By incorporating spatial-temporal correlations to regularize training and inference, the STKET boosts the R@10 and R@20 values to 82.6% and 96.3% on Pred Cls and the R@10 and R@20 values to 57.1% and 65.3% on SG Cls. Similarly, it consistently outperforms the baseline method on other metrics(i.e., the mR@10 and mR@20 metrics), as shown in Table III.

TABLE III: Comparison of the baseline STTran method (STTran), the baseline STTran method merely using the spatial encoder (STTran SE), the baseline STTran method merely using the temporal decoder (STTran TD), our framework merely using SKEL (Our SKEL), our framework merely using SKEL without the loss spk\mathcal{L}_{spk} (Ours SKEL w/o spk\mathcal{L}_{spk}), our framework merely using TKEL (Our TKEL), our framework merely using TKEL without the loss tpk\mathcal{L}_{tpk} (Ours TKEL w/o tpk\mathcal{L}_{tpk}), our framework removing STA (Ours w/o STA) and our framework (Ours).
Methods Pred Cls SG Cls
R@10 R@20 mR@10 mR@20 R@10 R@20 mR@10 mR@20
STTran 77.8 94.1 15.4 19.6 53.8 63.6 10.7 13.1
STTran SE 75.3 91.1 13.8 18.9 51.6 62.1 9.5 12.2
STTran TD 76.7 93.2 14.5 19.2 53.1 63.0 10.0 12.6
Ours SKEL 78.0 94.3 15.5 19.8 54.2 64.1 11.3 13.5
Ours SKEL w/o spk\mathcal{L}_{spk} 77.3 93.9 14.9 19.3 53.6 63.7 10.5 12.8
Ours TKEL 81.2 95.5 18.7 25.1 56.0 64.4 14.3 16.2
Ours TKEL w/o tpk\mathcal{L}_{tpk} 80.7 95.1 18.2 24.4 55.3 63.7 13.6 15.8
Ours w/o STA 81.9 95.8 19.4 26.5 56.5 64.9 14.8 16.7
Ours 82.6 96.3 20.1 27.3 57.1 65.3 15.4 17.2

Further, to emphasize the effectiveness of spatial-temporal knowledge, we conduct qualitative experiments that visualize two relationship transition matrices of different subject-object pairs (i.e., “man-tv” and “man-cup”) in Figure 7, where a lighter color indicates a higher probability of relationship transition. It’s worth noting that transition probabilities vary widely depending on the predicates and object categories involved, e.g., the transition likelihood for common predicates like “looking at” is much higher than for rare predicates such as “writing on”. This variance within the transition matrix highlights the importance of incorporating spatial-temporal knowledge for effective relationship prediction regularization. Additionally, we evaluate the entropy of spatial-temporal knowledge embeddings generated by our STKET model. Lower entropy values in these embeddings, particularly those for infrequent relationship categories like “eating”, “writing on”, and “twisting”, indicate a significant amount of prior information. This suggests that these embeddings offer substantial prediction regularization, potentially explaining why our STKET model excels over existing leading algorithms in predicting less common relationships.

Since the STKET framework consists of three complementary modules, i.e., the SKEL module, the TKEL module, and the STA module, in the following, we further conduct more ablation experiments to analyze the actual contribution of each module for a more in-depth understanding.

IV-C2 Analysis of SKEL

To evaluate the actual contribution of the SKEL module, we compare the performance of our STKEL framework merely using this module (namely, “Ours SKEL”) with the performance of the baseline STTran method merely using its spatial encoder (namely, “STTran SE”). As shown in Table III, the SKEL module improves the R@10 from 75.3% to 78.0% and the R@20 from 91.1% to 94.3% on Pred Cls, with improvements of 2.7% and 3.2%, respectively. Similarly, it improves the R@10 from 51.6% to 54.2% and the R@20 from 62.1% to 64.1% on SG Cls, with improvements of 2.6% and 2.0%, respectively. It is worth noting that the SKEL module achieves not only performance improvement on the R@10 and R@20 metrics but also on the mR@10 and mR@20 metrics. Specifically, this module obtains an mR@10 improvement of 1.7% and 1.8% and an mR@20 improvement of 0.9% and 1.3% on Pred Cls and SG Cls, respectively. These results demonstrate that the spatial co-occurrence correlation within each image can help to regularize spatial prediction space within each image effectively.

In the SKEL module, the loss spk\mathcal{L}_{spk} helps to learn accurate spatial knowledge embedding, thereby effectively regularizing spatial prediction space within each image. To evaluate the contribution of this loss, we conduct experiments that only utilize the SKEL module without the loss (namely, “Ours SKEL w/o spk\mathcal{L}_{spk}”) for comparison purposes. As presented in Table III, it decreases the performance by 0.7%/0.4% on Pred Cls with R@10/20 and 0.6%/0.4% on SG Cls with R@10/20. Similarly, it degrades performance by 0.6%/0.5% on Pred Cls with mR@10/20 and 0.8%/0.7% on SG Cls with mR@10/20.

IV-C3 Analysis of TKEL

To evaluate the actual contribution of the TKEL module, we conduct experiments that remove the SKEL and STA module in STKET while only using this module (namely, “Our TKEL”) and compare it with the baseline STTran method merely using its temporal decoder (namely, “STTran TD”) for comparison purposes. Different from “STTran SE” which explores the spatial context within the single frame, “STTran TD” introduces a sliding window to capture the temporal dependencies between frames and thus achieves obvious performance improvement, as presented in Table III. Specifically, it improves the performance by 1.4%/2.1% on Pred Cls with R@10/20 and 1.5%/0.9% on SG Cls with R@10/20, which demonstrates the importance of temporal correlations in recognizing dynamic visual relationships. However, “Our TKEL” performs better than this baseline “STTran TD”, with an R@10/20 improvement of 4.5% and 4.4% on Pred Cls and an R@10/20 improvement of 2.9% and 1.4% on SG Cls. Furthermore, “Our TKEL” consistently outperforms the baseline “STTran TD” in terms of the mR@10 and mR@20 metrics, i.e., it provides an R@10/20 improvement of 4.2% and 5.9% on Pred Clsand an mR@10/20 improvement of 4.3% and 3.6% on SG Cls. These results demonstrate that integrating temporal knowledge can be helpful in regularizing model learning correct temporal correlations, thereby effectively reducing ambiguous predictions.

In the TKEL module, the loss tpk\mathcal{L}_{tpk}helps to learn accurate temporal knowledge embedding, thereby effectively regularizing sequential variation space across temporal frames. To evaluate the contribution of this loss, we conduct experiments that merely use TKEL without the loss (namely, “Ours TKEL w/o tpk\mathcal{L}_{tpk}”) for comparison purposes. As presented, it decreases the performance by 0.5%/0.4% on Pred Cls with R@10/20 and 0.7%/0.7% on SG Cls with R@10/20. Similarly, it degrades performance by 0.5%/0.7% on Pred Cls with mR@10/20 and 0.7%/0.4% on SG Cls with mR@10/20.

Refer to caption
(a) Impact of sliding window in TKEL
Refer to caption
(b) Impact of sliding window in STA
Figure 8: Ablation analysis of the frame number of sliding window in (a) TKEL and (b) STA modules. Evaluated on Action Genome [26].

IV-C4 Analysis of STA

Considering the sliding window may result in many irrelevant representations, which may easily overwhelm the valuable information, it is difficult for the TKEL module to capture the long-term context of each subject-object pair. Thus, we design the STA module that aggregates spatial- and temporal-embedded representations of the identical subject-object pair across different frames to explore the long-term context. To evaluate its actual contribution, we conduct experiments that remove STA (namely, “Ours w/o STA”). As shown in Table III, it decreases the performance by 0.7%/0.5% on Pred Cls with R@10/20 and 0.6%/0.4% on SG Cls with R@10/20. It also decreases performance by 0.7%/0.8% on Pred Cls with mR@10/20 and 0.6%/0.5% on SG Cls with mR@10/20.

IV-C5 Analysis of Sliding Window

In the TKEL and STA modules, we introduce the sliding window to explore the temporal context contained in previous frames, in which the frame number is a crucial threshold that controls the sequence length. Setting it to a small value may miss temporal correlations, while setting it to a large value may introduce much irrelevant information and result in high computation costs. To figure out the optimal settings, we conduct experiments with different frame numbers. As shown in Figure 8, introducing more previous frames can significantly improve performance when there is only the current frame (i.e., the frame number is one). Besides, it is worth noting that a large frame number does not always mean better performance because long-term frame sequences contain complicated temporal contexts, which may mislead the model.

IV-D Limitation

As aforementioned, our STKET framework leverages spatial-temporal prior knowledge to effectively regularize relationship predictions, thereby lessening reliance on training samples and mitigating the imbalance problem in the Action Genome. The significant performance boost within tailed relationship categories demonstrates it, as shown in Figure 6b. Nevertheless, in extreme scenarios where subject-predicate-object triplets are exceedingly rare, the STKET framework struggles to provide accurate statistical regularization and thus may predict false predicates. As illustrated in 9, the bias from skewed long-tail distribution leads to a common but incorrect predicate “sit on” instead of the rare but correct predicate “lean on”, in the relationship of “person-lean on-sofa”. The same type of misclassification also occur with ”person-carry-cloth”. For these challenges, we conjecture that integrating spatial-temporal knowledge from various distributions (e.g., ImageNet-VidVRD [21], Home Action Genome [64]) can provide more accurate and robust regularization. Furthermore, we argue that leveraging the abundant contextual information in web text data could enhance the regularization of such rare relationships.

Refer to caption
Figure 9: Instances of failure cases resulted from relationship imbalance. For rare subject-predicate-object triplets, the model may predict similar predicates that have higher frequency instead of the ground truth predicate, which occurs less frequently in the training set.

V Conclusion

In this work, we propose to explore spatial-temporal prior knowledge to facilitate VidSGG via a spatial-temporal knowledge-embedded transformer (STKET) framework. It contains the spatial and temporal knowledge embedded layers that integrate spatial co-occurrence correlations to guide aggregating spatial contextual information and temporal transition correlations to help extract temporal contextual information. In this way, it can, on the one hand, aggregate spatial and temporal information to learn more representative relationship representation and, on the other hand, effectively regularize relationship prediction and thus reduce the dependencies on training samples. Extensive experiments illustrate its superiority over current state-of-the-art algorithms.

References

  • [1] D. Xu, Y. Zhu, C. B. Choy, and L. Fei-Fei, “Scene graph generation by iterative message passing,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 5410–5419.
  • [2] G. Ren, L. Ren, Y. Liao, S. Liu, B. Li, J. Han, and S. Yan, “Scene graph generation with hierarchical context,” IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 2, pp. 909–915, 2020.
  • [3] L. Tao, L. Mi, N. Li, X. Cheng, Y. Hu, and Z. Chen, “Predicate correlation learning for scene graph generation,” IEEE Transactions on Image Processing, vol. 31, pp. 4173–4185, 2022.
  • [4] Z. Tu, H. Li, D. Zhang, J. Dauwels, B. Li, and J. Yuan, “Action-stage emphasized spatiotemporal vlad for video action recognition,” IEEE Transactions on Image Processing, vol. 28, no. 6, pp. 2799–2812, 2019.
  • [5] T. Han, W. Xie, and A. Zisserman, “Temporal alignment networks for long-term video,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 2906–2916.
  • [6] Z. Zhou, C. Ding, J. Li, E. Mohammadi, G. Liu, Y. Yang, and Q. M. J. Wu, “Sequential order-aware coding-based robust subspace clustering for human action recognition in untrimmed videos,” IEEE Transactions on Image Processing, vol. 32, pp. 13–28, 2023.
  • [7] Z. Tang, Y. Liao, S. Liu, G. Li, X. Jin, H. Jiang, Q. Yu, and D. Xu, “Human-centric spatio-temporal video grounding with visual transformers,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 32, no. 12, pp. 8238–8249, 2021.
  • [8] Z. Ding, T. Hui, J. Huang, X. Wei, J. Han, and S. Liu, “Language-bridged spatial-temporal interaction for referring video object segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 4964–4973.
  • [9] T. Hui, S. Liu, Z. Ding, S. Huang, G. Li, W. Wang, L. Liu, and J. Han, “Language-aware spatial-temporal collaboration for referring video segmentation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  • [10] T. Nishimura, A. Hashimoto, Y. Ushiku, H. Kameko, and S. Mori, “State-aware video procedural captioning,” in Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 1766–1774.
  • [11] Y. Huang, H. Xue, J. Chen, H. Ma, and H. Ma, “Semantic tag augmented xlanv model for video captioning,” in Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 4818–4822.
  • [12] H. Wang, G. Lin, S. C. H. Hoi, and C. Miao, “Cross-modal graph with meta concepts for video captioning,” IEEE Transactions on Image Processing, vol. 31, pp. 5150–5162, 2022.
  • [13] X. Hua, X. Wang, T. Rui, F. Shao, and D. Wang, “Adversarial reinforcement learning with object-scene relational graph for video captioning,” IEEE Transactions on Image Processing, vol. 31, pp. 2004–2016, 2022.
  • [14] Z. Yang, N. Garcia, C. Chu, M. Otani, Y. Nakashima, and H. Takemura, “Bert representations for video question answering,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2020, pp. 1556–1565.
  • [15] P. Zeng, H. Zhang, L. Gao, J. Song, and H. T. Shen, “Video question answering with prior knowledge and object-sensitive learning,” IEEE Transactions on Image Processing, vol. 31, pp. 5936–5948, 2022.
  • [16] L. Gao, Y. Lei, P. Zeng, J. Song, M. Wang, and H. T. Shen, “Hierarchical representation network with auxiliary tasks for video captioning and video question answering,” IEEE Transactions on Image Processing, vol. 31, pp. 202–215, 2022.
  • [17] Y. Liu, X. Zhang, F. Huang, B. Zhang, and Z. Li, “Cross-attentional spatio-temporal semantic graph networks for video question answering,” IEEE Transactions on Image Processing, vol. 31, pp. 1684–1696, 2022.
  • [18] D. Liu, M. Bober, and J. Kittler, “Visual semantic information pursuit: A survey,” IEEE transactions on pattern analysis and machine intelligence, vol. 43, no. 4, pp. 1404–1422, 2019.
  • [19] P. Xu, X. Chang, L. Guo, P.-Y. Huang, X. Chen, and A. G. Hauptmann, “A survey of scene graph: Generation and application,” IEEE Trans. Neural Netw. Learn. Syst, vol. 1, 2020.
  • [20] X. Chang, P. Ren, P. Xu, Z. Li, X. Chen, and A. Hauptmann, “A comprehensive survey of scene graphs: Generation and application,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 1, pp. 1–26, 2021.
  • [21] X. Shang, T. Ren, J. Guo, H. Zhang, and T.-S. Chua, “Video visual relation detection,” in Proceedings of the 25th ACM international conference on Multimedia, 2017, pp. 1300–1308.
  • [22] X. Qian, Y. Zhuang, Y. Li, S. Xiao, S. Pu, and J. Xiao, “Video relation detection with spatio-temporal graph,” in Proceedings of the 27th ACM International Conference on Multimedia, 2019, pp. 84–93.
  • [23] S. Zheng, S. Chen, and Q. Jin, “Vrdformer: End-to-end video visual relation detection with transformers,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 18 836–18 846.
  • [24] Y. Li, X. Yang, and C. Xu, “Dynamic scene graph generation via anticipatory pre-training,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 13 874–13 883.
  • [25] Y. Cong, W. Liao, H. Ackermann, B. Rosenhahn, and M. Y. Yang, “Spatial-temporal transformer for dynamic scene graph generation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 16 372–16 382.
  • [26] J. Ji, R. Krishna, L. Fei-Fei, and J. C. Niebles, “Action genome: Actions as compositions of spatio-temporal scene graphs,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 10 236–10 247.
  • [27] J. Zhang, K. J. Shih, A. Elgammal, A. Tao, and B. Catanzaro, “Graphical contrastive losses for scene graph parsing,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 11 535–11 543.
  • [28] X. Lin, C. Ding, J. Zeng, and D. Tao, “Gps-net: Graph property sensing network for scene graph generation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 3746–3753.
  • [29] G. Ren, L. Ren, Y. Liao, S. Liu, B. Li, J. Han, and S. Yan, “Scene graph generation with hierarchical context,” IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 2, pp. 909–915, 2020.
  • [30] Y. Lu, H. Rai, J. Chang, B. Knyazev, G. Yu, S. Shekhar, G. W. Taylor, and M. Volkovs, “Context-aware scene graph generation with seq2seq transformers,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 15 931–15 941.
  • [31] C. Lu, R. Krishna, M. Bernstein, and L. Fei-Fei, “Visual relationship detection with language priors,” in European conference on computer vision.   Springer, 2016, pp. 852–869.
  • [32] Y. Li, W. Ouyang, B. Zhou, K. Wang, and X. Wang, “Scene graph generation from objects, phrases and region captions,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 1261–1270.
  • [33] J. Yang, J. Lu, S. Lee, D. Batra, and D. Parikh, “Graph r-cnn for scene graph generation,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 670–685.
  • [34] R. Zellers, M. Yatskar, S. Thomson, and Y. Choi, “Neural motifs: Scene graph parsing with global context,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 5831–5840.
  • [35] T. N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” arXiv preprint arXiv:1609.02907, 2016.
  • [36] T. Chen, W. Yu, R. Chen, and L. Lin, “Knowledge-embedded routing network for scene graph generation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 6163–6171.
  • [37] K. Tang, H. Zhang, B. Wu, W. Luo, and W. Liu, “Learning to compose dynamic tree structures for visual contexts,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 6619–6628.
  • [38] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
  • [39] Y. Cong, M. Y. Yang, and B. Rosenhahn, “Reltr: Relation transformer for scene graph generation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  • [40] S. Kundu and S. N. Aakur, “Is-ggt: Iterative scene graph generation with generative transformers,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 6292–6301.
  • [41] Y.-H. H. Tsai, S. Divvala, L.-P. Morency, R. Salakhutdinov, and A. Farhadi, “Video relationship reasoning using gated spatio-temporal energy graph,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 10 424–10 433.
  • [42] C. Liu, Y. Jin, K. Xu, G. Gong, and Y. Mu, “Beyond short-term snippet: Video relation detection with spatio-temporal global context,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 10 840–10 849.
  • [43] Y. Teng, L. Wang, Z. Li, and G. Wu, “Target adaptive context aggregation for video scene graph generation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 13 688–13 697.
  • [44] S. Wang, L. Gao, X. Lyu, Y. Guo, P. Zeng, and J. Song, “Dynamic scene graph generation via temporal prior inference,” in Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 5793–5801.
  • [45] Y. Kumar and A. Mishra, “Few-shot referring relationships in videos,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 2289–2298.
  • [46] X. Wang, L. Zhu, Y. Wu, and Y. Yang, “Symbiotic attention for egocentric action recognition with object-centric alignment,” IEEE transactions on pattern analysis and machine intelligence, 2020.
  • [47] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
  • [48] H. Li, C. Li, A. Zheng, J. Tang, and B. Luo, “Mskat: Multi-scale knowledge-aware transformer for vehicle re-identification,” IEEE Transactions on Intelligent Transportation Systems, vol. 23, no. 10, pp. 19 557–19 568, 2022.
  • [49] Z. Wang, J. Zhang, T. Chen, W. Wang, and P. Luo, “Restoreformer++: Towards real-world blind face restoration from undegraded key-value pairs,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  • [50] T. Chen, T. Pu, H. Wu, Y. Xie, and L. Lin, “Structured semantic transfer for multi-label recognition with partial labels,” in Proceedings of the AAAI conference on artificial intelligence, vol. 36, no. 1, 2022, pp. 339–346.
  • [51] T. Pu, T. Chen, H. Wu, and L. Lin, “Semantic-aware representation blending for multi-label image recognition with partial labels,” in Proceedings of the AAAI conference on artificial intelligence, vol. 36, no. 2, 2022, pp. 2091–2098.
  • [52] Y. Xie, T. Chen, T. Pu, H. Wu, and L. Lin, “Adversarial graph representation adaptation for cross-domain facial expression recognition,” in Proceedings of the 28th ACM international conference on Multimedia, 2020.
  • [53] Z. Peng, Z. Li, J. Zhang, Y. Li, G.-J. Qi, and J. Tang, “Few-shot image recognition with knowledge transfer,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2019.
  • [54] T. Chen, L. Lin, R. Chen, X. Hui, and H. Wu, “Knowledge-guided multi-label few-shot learning for general image recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 3, pp. 1371–1384, 2022.
  • [55] T. Pu, T. Chen, Y. Xie, H. Wu, and L. Lin, “Au-expression knowledge constrained representation learning for facial expression recognition,” in 2021 IEEE international conference on robotics and automation (ICRA).   IEEE, 2021, pp. 11 154–11 161.
  • [56] T. Chen, T. Pu, H. Wu, Y. Xie, L. Liu, and L. Lin, “Cross-domain facial expression recognition: A unified evaluation benchmark and adversarial graph learning,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 12, pp. 9887–9903, 2022.
  • [57] W. Yang, X. Wang, A. Farhadi, A. Gupta, and R. Mottaghi, “Visual semantic navigation using scene priors,” in Proceedings of International Conference on Learning Representations (ICLR), 2019.
  • [58] S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: Towards real-time object detection with region proposal networks,” Advances in neural information processing systems, vol. 28, 2015.
  • [59] K. He, G. Gkioxari, P. Dollár, and R. Girshick, “Mask r-cnn,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 2961–2969.
  • [60] A. R. Vandenbroucke, J. Fahrenfort, J. Meuwese, H. Scholte, and V. Lamme, “Prior knowledge about objects determines neural color representation in human visual cortex,” Cerebral cortex, vol. 26, no. 4, pp. 1401–1408, 2016.
  • [61] C.-W. Lee, W. Fang, C.-K. Yeh, and Y.-C. F. Wang, “Multi-label zero-shot learning with structured knowledge graphs,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 1576–1585.
  • [62] I. Loshchilov and F. Hutter, “Decoupled weight decay regularization,” arXiv preprint arXiv:1711.05101, 2017.
  • [63] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga et al., “Pytorch: An imperative style, high-performance deep learning library,” Advances in neural information processing systems, vol. 32, 2019.
  • [64] N. Rai, H. Chen, J. Ji, R. Desai, K. Kozuka, S. Ishizaka, E. Adeli, and J. C. Niebles, “Home action genome: Cooperative compositional action understanding,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 11 184–11 193.
[Uncaptioned image] Tao Pu received a B.E. degree from the School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, China, in 2020, where he is currently pursuing a Ph.D. degree in computer science. He has authored and coauthored approximately 10 papers published in top-tier academic journals and conferences, including T-PAMI, AAAI, ACM MM, etc.
[Uncaptioned image] Tianshui Chen received a Ph.D. degree in computer science at the School of Data and Computer Science Sun Yat-sen University, Guangzhou, China, in 2018. Prior to earning his Ph.D, he received a B.E. degree from the School of Information and Science Technology in 2013. He is currently an associated professor in the Guangdong University of Technology. His current research interests include computer vision and machine learning. He has authored and coauthored approximately 40 papers published in top-tier academic journals and conferences, including T-PAMI, T-NNLS, T-IP, T-MM, CVPR, ICCV, AAAI, IJCAI, ACM MM, etc. He has served as a reviewer for numerous academic journals and conferences. He was the recipient of the Best Paper Diamond Award at IEEE ICME 2017.
[Uncaptioned image] Hefeng Wu received a B.S. degree in computer science and technology and a Ph.D. degree in computer application technology from Sun Yat-sen University, Guangzhou, China. He is currently a research professor at the School of Computer Science and Engineering, Sun Yat-sen University, China. His research interests include computer vision, multimedia, and machine learning. He has published works in and served as reviewers for many top-tier academic journals and conferences, including T-PAMI, T-IP, T-MM, CVPR, ICCV, AAAI, ACM MM, etc.
[Uncaptioned image] Yongyi Lu is currently an Associate Professor at Guangdong University of Technology (GDUT). He received his Ph.D. in Computer Science and Engineering at Hong Kong University of Science and Technology (HKUST) in 2018. From 2019-2022, he was a Postdoctoral Fellow in Prof. Alan Yuille’s group at Johns Hopkins University. He received his B.E. and Master both at Sun Yat-Sen University, Guangzhou, China in 2011 and 2014. His research focuses on computer vision, medical image analysis and the interdisciplinary fields of them. He has published over 30 peer-reviewed conference/journal articles such as CVPR, ICCV, ECCV, NeurIPS, MICCAI, BMVC and T-IP, with a total citation of over 2500. He has won the 4th place in ImageNet DET challenge in 2014 and 5th place in ImageNet VID challenge in 2015. He serves as a reviewer of conferences/journals including IJCV, T-IP, T-MM, T-CSVT, Nature Machine Intelligence, Neurocomputing, CVPR, ICCV, ECCV, NeurIPS, ICLR and BMVC. He is a member of IEEE.
[Uncaptioned image] Liang Lin (Fellow, IEEE) is a full professor at Sun Yat-sen University. From 2008 to 2010, he was a postdoctoral fellow at the University of California, Los Angeles. From 2016–2018, he led the SenseTime R&D teams to develop cutting-edge and deliverable solutions for computer vision, data analysis and mining, and intelligent robotic systems. He has authored and coauthored more than 100 papers in top-tier academic journals and conferences (e.g., 15 papers in TPAMI and IJCV and 60+ papers in CVPR, ICCV, NIPS, and IJCAI). He has served as an associate editor of IEEE Trans. Human-Machine Systems, The Visual Computer, and Neurocomputing and as an area/session chair for numerous conferences, such as CVPR, ICME, ACCV, and ICMR. He was the recipient of the Annual Best Paper Award by Pattern Recognition (Elsevier) in 2018, the Best Paper Diamond Award at IEEE ICME 2017, the Best Paper Runner-Up Award at ACM NPAR 2010, Google Faculty Award in 2012, the Best Student Paper Award at IEEE ICME 2014, and the Hong Kong Scholars Award in 2014. He is a Fellow of IEEE, IAPR, and IET.