Evaluating Temporal Queries over Videos

dc.contributor.advisorYu, Xiaohui
dc.contributor.authorChen, Yueting
dc.date.accessioned2023-12-08T14:32:52Z
dc.date.available2023-12-08T14:32:52Z
dc.date.issued2023-12-08
dc.date.updated2023-12-08T14:32:51Z
dc.degree.disciplineComputer Science
dc.degree.levelDoctoral
dc.degree.namePhD - Doctor of Philosophy
dc.description.abstractVideos have been an important part of people's daily lives and are continuously growing in terms of volume, size, and variety of content. Recent advances in Computer Vision (CV) algorithms have improved accuracy and efficiency, making video annotations possible with high accuracy. In this work, we follow a general framework to first obtain annotations utilizing state-of-the-art CV algorithms, and then consider three research problems on evaluating temporal queries with such annotations. Specifically, we first investigate the temporal queries that consider only co-occurrence relationships between objects on video feeds, where we take the first step and define such queries in a way that they incorporate certain physical aspects of video capture such as object occlusion. We propose two techniques, Marked Frame Set (MFS) and Sparse State Graph (SSG), to organize all detected objects in the intermediate data generation layer, which effectively, given the queries, minimizes the number of objects and frames that have to be considered during query evaluation. Then, we consider the query with a ranking mechanism that aims to retrieve clips from large video repositories in which objects co-occur in a query-specified fashion. We propose a two-phased approach, where we build indexes during the Ingestion Phase, and then answer queries during the Query Phase using the Partition-Based Query Processing (PBQP) algorithm, which efficiently produces the desired (query-specified) number of results with the highest scores. Finally, we further consider both spatial and temporal information with graph representations and define the problem of Spatial and Temporal Constrained Ranked Retrieval (STAR Retrieval) over videos. Based on the graph representation, we propose a two-phase approach, consisting of the ingestion phase, where we construct and materialize the Graph Index (GI), and the query phase, where we compute the top-ranked windows (video clips) according to the window matching score efficiently. We propose two algorithms to perform Spatial Matching (SMA) and Temporal Matching (TM) separately with an early-stopping mechanism. We present the details of the above three research problems and our proposed methods. Via experiments conducted on various datasets, we show the effectiveness of our proposed methods.
dc.identifier.urihttps://hdl.handle.net/10315/41670
dc.languageen
dc.rightsAuthor owns copyright, except where explicitly noted. Please contact the author directly with licensing requests.
dc.subjectComputer science
dc.subject.keywordsVideo query processing
dc.subject.keywordsVideo database
dc.subject.keywordsVideo annotations
dc.subject.keywordsSpatial query
dc.subject.keywordsTemporal query
dc.subject.keywordsRanked window retrieval
dc.titleEvaluating Temporal Queries over Videos
dc.typeElectronic Thesis or Dissertation

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Chen_Yueting_2023_PhD.pdf
Size:
7.52 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 2 of 2
No Thumbnail Available
Name:
license.txt
Size:
1.87 KB
Format:
Plain Text
Description:
No Thumbnail Available
Name:
YorkU_ETDlicense.txt
Size:
3.39 KB
Format:
Plain Text
Description:

Collections