Research papers and contributions
IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2025
We propose a content-agnostic deepfake speech detection model trained on a curated dataset of 1M+ real-fake speech pairs. By using transcripts from real speech to generate synthetic audio using cutting-edge TTS systems (YourTTS, XTTS-v2, OpenVoice), we eliminate linguistic cues and force the model to rely solely on acoustic artifacts. Our model achieves superior performance compared to prior work and outperforms human listeners across all speech lengths.
Analyzed failure cases of MASt3R for multi-view reconstruction under challenging capture conditions. Collected a small scene dataset and documented limitations of current state-of-the-art methods.
I'm always open to discussing new research ideas, potential collaborations, or interesting projects. Feel free to reach out!
Get in Touch