Skip to yearly menu bar Skip to main content


Poster+Demo Session
in
Workshop: Audio Imagination: NeurIPS 2024 Workshop AI-Driven Speech, Music, and Sound Generation

MLADDC: Multi-Lingual Audio Deepfake Detection Corpus

ARTH SHAH · Ravindrakumar M. Purohit · Dharmendra Vaghera · Hemant Patil

[ ] [ Project Page ]
Sat 14 Dec 4:15 p.m. PST — 5:30 p.m. PST

Abstract:

This study develop Multi-Lingual Audio Deepfake Detection Corpus (MLADDC) to boost the ADD research. Existing datasets suffer from several limitations, in particular, they are limited to one or two languages. Proposed dataset contains 20 languages, which have been released in 4 Tracks (6 - Indian languages, 14 - International languages, 20 languages half-truth data, and combined data). Moreover, the proposed dataset has 400K files (1,125+ hours) of data, which makes it one of the largest datasets. Deepfakes in MLADDC have been produced using advanced DL methods, such as HiFiGAN and BigVGAN. Another novelty lies in its sub-dataset, that has partial deepfakes (Half-Truth). We compared our dataset with various existing datasets, using cross-database method. For comparison, we also proposed baseline accuracy of 68.44 %, and EER of 40.9 % with MFCC features and CNN classifier (14 languages track only) indicating technological challenges associated with ADD task on proposed dataset.

Chat is not available.