Skip to yearly menu bar Skip to main content


Relational Space-Time Query in Long-Form Videos

Xitong Yang · Fu-Jen Chu · Matt Feiszli · Raghav Goyal · Lorenzo Torresani · Du Tran

West Building Exhibit Halls ABC 217
award Highlight
[ ]


Egocentric videos are often available in the form of uninterrupted, uncurated long videos capturing the camera wearers’ daily life activities.Understanding these videos requires models to be able to reason about activities, objects, and their interactions. However, current video benchmarks study these problems independently and under short, curated clips. In contrast, real-world applications, e.g., AR assistants, require bundling these problems for both model development and evaluation. In this paper, we propose to study these problems in a joint framework for long video understanding. Our contributions are three-fold. First, we propose an integrated framework, namely Relational Space-Time Query (ReST), for evaluating video understanding models via templated spatiotemporal queries. Second, we introduce two new benchmarks, ReST-ADL and ReST-Ego4D, which augment the existing egocentric video datasets with abundant query annotations generated by the ReST framework. Finally, we present a set of baselines and in-depth analysis on the two benchmarks and provide insights about the query tasks. We view our integrated framework and benchmarks as a step towards comprehensive, multi-step reasoning in long videos, and believe it will facilitate the development of next generations of video understanding models.

Chat is not available.