An Improved System for Partially Fake Audio Detection Using Pre-trained Model

Jianqian Zhang, Hanyue Liu, Mengyuan Deng, Jing Wang*, Yi Sun, Liang Xu, Jiahao Li

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

The technology of speech synthesis and conversion has made good progress with the development of deep learning. However, such technology can also do harm to information security and may be applied for illegal uses. Therefore, researchers have conducted a lot of research on the task of speech deep forgery detection recently. Corresponding to this, the Audio Deep Synthesis Detection Challenge 2023 (ADD 2023) is held. In this paper, we propose a fake audio detecting system using a pre-trained model, focusing on partially fake audio detection tasks. We have presented our models to the ADD 2023 challenge. In the final competition, our system got a score of 0.4855 in the manipulation region location track.

Original languageEnglish
Title of host publicationMan-Machine Speech Communication - 18th National Conference, NCMMSC 2023, Proceedings
EditorsJia Jia, Zhenhua Ling, Xie Chen, Ya Li, Zixing Zhang
PublisherSpringer Science and Business Media Deutschland GmbH
Pages346-353
Number of pages8
ISBN (Print)9789819706006
DOIs
Publication statusPublished - 2024
Event18th National Conference on Man-Machine Speech Communication, NCMMSC 2023 - Suzhou, China
Duration: 8 Dec 202311 Dec 2023

Publication series

NameCommunications in Computer and Information Science
Volume2006
ISSN (Print)1865-0929
ISSN (Electronic)1865-0937

Conference

Conference18th National Conference on Man-Machine Speech Communication, NCMMSC 2023
Country/TerritoryChina
CitySuzhou
Period8/12/2311/12/23

Keywords

  • audio deepfake detection
  • deep-learning
  • fake region location
  • manipulation

Fingerprint

Dive into the research topics of 'An Improved System for Partially Fake Audio Detection Using Pre-trained Model'. Together they form a unique fingerprint.

Cite this