Skip to Main content Skip to Navigation
Theses

Deep Inside Visual-Semantic Embeddings

Abstract : Nowadays Artificial Intelligence (AI) is omnipresent in our society. The recentdevelopment of learning methods based on deep neural networks alsocalled "Deep Learning" has led to a significant improvement in visual representation models.and textual.In this thesis, we aim to further advance image representation and understanding.Revolving around Visual Semantic Embedding (VSE) approaches, we explore different directions: We present relevant background covering images and textual representation and existing multimodal approaches. We propose novel architectures further improving retrieval capability of VSE and we extend VSE models to novel applications and leverage embedding models to visually ground semantic concept. Finally, we delve into the learning process andin particular the loss function by learning differentiable approximation of ranking based metric.
Complete list of metadata

https://tel.archives-ouvertes.fr/tel-03402492
Contributor : Abes Star :  Contact
Submitted on : Monday, October 25, 2021 - 4:29:26 PM
Last modification on : Tuesday, November 16, 2021 - 4:53:03 AM

File

ENGILBERGE_Martin_2020.pdf
Version validated by the jury (STAR)

Identifiers

  • HAL Id : tel-03402492, version 1

Citation

Martin Engilberge. Deep Inside Visual-Semantic Embeddings. Machine Learning [cs.LG]. Sorbonne Université, 2020. English. ⟨NNT : 2020SORUS150⟩. ⟨tel-03402492⟩

Share

Metrics

Record views

49

Files downloads

22