Almut Sophia Koepke
Research Fellow
a-sophia.koepke (at) uni-tuebingen.de Google Scholar

Profile

Almut Sophia Koepke is a post-doctoral research fellow in the EML group where she works with Zeynep Akata. Prior to that, she completed her DPhil in the Visual Geometry Group (VGG) at the University of Oxford, supervised by Andrew Zisserman.

Her research focusses on multi-modal learning problems with sound, vision, and language.

Publications

Exposing and Mitigating Spurious Correlations for Cross-Modal Retrieval
Jae Myung Kim, A. Sophia Koepke, Cordelia Schmid, Zeynep Akata
Multimodal Learning and Applications Workshop at the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPRW), 2023
Paper | Code

 

Temporal and cross-modal attention for audio-visual zero-shot learning
Otniel-Bogdan Mercea*, Thomas Hummel*, A. Sophia Koepke, Zeynep Akata
European Conference on Computer Vision (ECCV), 2022
Paper | Code

 

Audio-visual generalised zero-shot learning with cross-modal attention and language
Otniel-Bogdan Mercea, Lukas Riesch, A. Sophia Koepke, Zeynep Akata
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022
Paper | Code

 

CLEVR-X: A visual reasoning dataset for natural language explanations
Leonard Salewski, A. Sophia Koepke, Hendrik Lensch, Zeynep Akata
Springer Lecture Notes on Artificial Intelligence, 2022
Paper | Project page | Code
This was also presented at the CVPR 2022 Workshop on Explainable AI for Computer Vision (XAI4CV).

 

Audio retrieval with natural language queries: A benchmark study
A. Sophia Koepke*, Andreea-Maria Oncescu*, Joao F. Henriques, Zeynep Akata, Samuel Albanie
Transactions on Multimedia, 2022
Paper | Project page | Code
Extension of the INTERSPEECH paper with a new dataset and new results.

 

Audio retrieval with natural language queries
Andreea-Maria Oncescu*, A. Sophia Koepke*, Joao F. Henriques, Zeynep Akata, Samuel Albanie
INTERSPEECH, 2021
Paper | Project page | Code
Shortlisted for best student paper award.

 

Distilling audio-visual knowledge by compositional contrastive learning
Yanbei Chen, Yongqin Xian, A. Sophia Koepke, Ying Shan, Zeynep Akata
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021
Paper | Code

 

Sight to sound: An end-to-end approach for visual piano transcription
A. Sophia Koepke, Olivia Wiles, Yael Moses, Andrew Zisserman
The International Conference on Acoustics, Speech, & Signal Processing (ICASSP), 2020
Paper | Project page
Oral presentation.

 

Self-supervised learning of class embeddings from video
Olivia Wiles, A. Sophia Koepke, Andrew Zisserman
IEEE/CVF International Conference on Computer Vision Workshop (ICCV Workshop), 2019
Paper

 

Visual pitch estimation
A. Sophia Koepke, Olivia Wiles, Andrew Zisserman
Sound and Music Computation Conference (SMC), 2019
Paper | Project page

 

Self-supervised learning of a facial attribute embedding from video
Olivia Wiles*, A. Sophia Koepke*, Andrew Zisserman
British Machine Vision Conference (BMVC), 2018
Paper | Supplementary material | Project page | Code
Oral presentation.

 

X2Face: A network for controlling face generation by using images, audio, and pose codes
Olivia Wiles*, A. Sophia Koepke*, Andrew Zisserman
European Conference on Computer Vision (ECCV), 2018
Paper | Project page | Code

 

* denotes equal contribution