Skip to Main content Skip to Navigation
Theses

Structured deep learning for video analysis

Abstract : With the massive increase of video content on Internet and beyond, the automatic understanding of visual content could impact many different application fields such as robotics, health care, content search or filtering. The goal of this thesis is to provide methodological contributions in Computer Vision and Machine Learning for automatic content understanding from videos. We emphasis on problems, namely fine-grained human action recognition and visual reasoning from object-level interactions. In the first part of this manuscript, we tackle the problem of fine-grained human action recognition. We introduce two different trained attention mechanisms on the visual content from articulated human pose. The first method is able to automatically draw attention to important pre-selected points of the video conditioned on learned features extracted from the articulated human pose. We show that such mechanism improves performance on the final task and provides a good way to visualize the most discriminative parts of the visual content. The second method goes beyond pose-based human action recognition. We develop a method able to automatically identify unstructured feature clouds of interest in the video using contextual information. Furthermore, we introduce a learned distributed system for aggregating the features in a recurrent manner and taking decisions in a distributed way. We demonstrate that we can achieve a better performance than obtained previously, without using articulated pose information at test time. In the second part of this thesis, we investigate video representations from an object-level perspective. Given a set of detected persons and objects in the scene, we develop a method which learns to infer the important object interactions through space and time using the video-level annotation only. That allows to identify important objects and object interactions for a given action, as well as potential dataset bias. Finally, in a third part, we go beyond the task of classification and supervised learning from visual content by tackling causality in interactions, in particular the problem of counterfactual learning. We introduce a new benchmark, namely CoPhy, where, after watching a video, the task is to predict the outcome after modifying the initial stage of the video. We develop a method based on object- level interactions able to infer object properties without supervision as well as future object locations after the intervention.
Document type :
Theses
Complete list of metadata

https://hal.archives-ouvertes.fr/tel-03117344
Contributor : Abes Star :  Contact
Submitted on : Tuesday, March 2, 2021 - 5:03:32 PM
Last modification on : Wednesday, March 3, 2021 - 3:21:29 AM

File

these.pdf
Version validated by the jury (STAR)

Identifiers

  • HAL Id : tel-03117344, version 2

Citation

Fabien Baradel. Structured deep learning for video analysis. Artificial Intelligence [cs.AI]. Université de Lyon, 2020. English. ⟨NNT : 2020LYSEI045⟩. ⟨tel-03117344v2⟩

Share

Metrics

Record views

68

Files downloads

89