Multimodal AI PhD Intern (Spring 2026)

AI overview

Collaborate on cutting-edge research in multi-modal deepfake detection and publish peer-reviewed papers, leveraging skills in Python, PyTorch, and modern deep learning stacks.

Who we are.

Reality Defender is an award-winning cybersecurity company helping enterprises and governments detect deepfakes and AI-generated media. Utilizing a patented multi-model approach, Reality Defender is robust against the bleeding edge of generative platforms producing video, audio, imagery, and text media. Reality Defender's API-first deepfake detection platform empowers teams and developers alike to identify fraud, disinformation campaigns, and harmful deepfakes in real time.

Backed by world class investors including DCVC, Illuminate Financial, Y Combinator, Booz Allen Hamilton, IBM, Accenture, Rackhouse, and Argon VC, Reality Defender works with leading enterprise clients, financial institutions, and governments in order to ensure AI-generated media is not used for malicious purposes.

Youtube: Reality Defender Wins RSA Most Innovative Startup

The Multimodal AI Internship.

The 4-month internship is designed for current PhD students and candidates to partner with Reality Defender's AI team to conduct cutting-edge research and publish peer-reviewed papers. Your primary collaborators will be Surya Koppisetti and Yi Zhu, who will guide and advise your efforts within multi-modal deepfake detection. This internship can be performed remotely, although you're welcome to work from our HQ in New York City.

What you'll do.

  • Investigate and propose new methods for detecting generative multi-modal content, spanning audio and vision modalities.

  • Perform research on multi-modal deepfake detection and reasoning tasks.

  • Collaborate with researchers in the team.

  • Write up results of research for internal reports and submission to academic journals/workshops.

  • Independently implement and evaluate ideas on modern deep learning stack - Python, PyTorch, and GPU-enabled cloud compute, like AWS/GCP.

Who you are.

  • PhD student in a relevant technical field, preferably three or more years into the program

  • Experience in multi-modal learning, such as in audio-visual classification and audio-language reasoning.

  • Proficient in Python and in building deep learning models with PyTorch.

  • Published peer-reviewed research papers in reputable AI and speech venues, e.g. CVPR, NeurIPS, ACL, Interspeech.

  • Excited about Reality Defender's mission to build a best-in-class and comprehensive deepfake and AI-generated content detection platform.

  • Available to start in Spring 2026, for a minimum duration of 4 months.

Reality Defender builds advanced cybersecurity solutions that empower enterprises and governments to detect deepfakes and AI-generated media. With a patented multi-model approach and an API-first platform, we provide robust tools to combat fraud and disinformation in real time, making us a key player in the fight against the misuse of AI technologies.

View all jobs
Salary
$5,000 – $7,000 per month
Report this job
Apply for this job