Otniel-Bogdan Mercea

otniel_bogdan_mercea.jpg

I am on the job market looking for research scientist roles. If you have a project that fits with my research skills, do not hesitate to contact me.

I am a final-year PhD candidate jointly affiliated with the Max Planck Institute for Intelligent Systems and the University of Tubingen, part of the IMPRS-IS doctoral program. My research is supervised by Prof. Zeynep Akata and Prof. Andreas Geiger. During my PhD, I was also a guest PhD student at Helmholtz Munich and Technical University of Munich, supervised by Prof. Zeynep Akata.

I earned my MSc in Artificial Intelligence from the University of Edinburgh in 2020, completing my thesis under the supervision of Prof. Amos Storkey. Prior to that, I obtained my BEng in Computers and Information technology from Politehnica University of Timisoara in 2019, with my undergraduate thesis supervised by Prof. Calin-Adrian Popa.

I have had the privilege of applying my research expertise in industry through several impactful opportunities. In the autumn of 2024, I interned at Google DeepMind, where I worked under the supervision of Stefano Pellegrini, Jasper Uijlings, and Cordelia Schmid on enhancing SAM 2 for scenarios involving significant occlusion or movement. Previously, from 2023 to 2024, I spent eight months at Google Research, collaborating with Anurag Arnab, Alexey Gritsenko, and Cordelia Schmid on the efficient adaptation of large-scale models. During this time, I also worked with Aleksandra Nowak, Utku Evci, and Yann Dauphin from Google DeepMind on a related project in efficient adaptation strategies. Prior to these roles, I was a Machine Learning Researcher at Everseen, where I focused on real-time multi-camera tracking systems.

Most of my research focuses on improving deep learning efficiency through data-efficient (low-shot) and model-efficient (adaptation) methods for large-scale models. Furthermore, I have extensive experience in multimodal learning, including audio-visual learning and multi-modal large language models, as well as in video semantic segmentation and tracking.


News

Dec 27, 2024 I have finished my 4 month internship at Google DeepMind. I have worked on enhancing SAM 2 for scenarios involving significant occlusion or movement, under the supervision of Stefano Pellegrini, Jasper Uijlings and Cordelia Schmid.
Oct 21, 2024 New preprint is available on optimal adapter placement for efficient transfer learning.
Apr 04, 2024 Our work on audio-visual GZSL using large multi-modal models was accepted at CVPR 2024 workshops (L3D-IVU).
Mar 22, 2024 After 8 months, I finished my internship at Google Research. I have worked on efficient adaptation under the supervision of Anurag Arnab, Alexey Gritsenko and Cordelia Schmid.
Feb 27, 2024 LoSA was accepted as a SPOTLIGHT at CVPR 2024.
Aug 21, 2023 ReGAdA was accepted as an ORAL to BMVC 2023.
Aug 11, 2023 AV-Diff was accepted at DAGM GCPR 2023

Selected publications

  1. losa.png
    Time-Memory-and Parameter-Efficient Visual Adaptation
    Otniel-Bogdan Mercea, Alexey Gritsenko, Cordelia Schmid, and 1 more author
    2024
    SPOTLIGHT @ CVPR
  2. avdiff.jpeg
    Text-to-feature diffusion for audio-visual few-shot learning
    Otniel-Bogdan Mercea, Thomas Hummel, A Sophia Koepke, and 1 more author
    2023
    DAGM GCPR
  3. avca.png
    Audio-visual generalised zero-shot learning with cross-modal attention and language
    Otniel-Bogdan Mercea, Lukas Riesch, A Koepke, and 1 more author
    2022
    CVPR
  4. tcaf.png
    Temporal and cross-modal attention for audio-visual zero-shot learning
    Otniel-Bogdan Mercea*, Thomas Hummel*, A Sophia Koepke, and 1 more author
    2022
    ECCV

Selected patents

  1. patent_tracking.png
    System and method for tracking and identifying moving objects
    Ana Cristina Todoran, and Otniel-Bogdan Mercea
    2023
    US Patent App. 17/562,364
  2. patent_adjusting.png
    System and method for adjusting a position of an order taking device
    Ana Cristina Todoran, Otniel-Bogdan Mercea, and Razvan-Dorel Cioarga
    2023
    US Patent App. 17/562,365