Conference Paper

Detecting the Undetectable: Human Judgments and the Challenge of Synthetic Voices

Loading...
Thumbnail Image

Fulltext URI

Document type

Text/Conference Paper

Additional Information

Date

Journal Title

Journal ISSN

Volume Title

Publisher

European Society for Socially Embedded Technologies (EUSSET)

Abstract

Synthetic voices generated using artificial intelligence (AI) are becoming increasingly indistinguishable from human voices, raising important concerns about trust, deception, and detection in digital communication. This preliminary work synthesizes the current landscape of research on human perception in detecting synthetic voices. We reviewed 13 papers from databases including ACM, IEEE, Springer, and MDPI, and identified five main types of perceptual cues that users rely on to detect voice synthesis: Intuition/Gut Feeling, Liveliness, Emotions, Linguistic Features, and Acoustic and Environmental Features. Our findings highlight the need for further empirical user studies to better understand how individuals perceive and assess the risks posed by synthetic voices. Such research can inform both educational and regulatory strategies aimed at increasing awareness and mitigating the potential harms of synthetic voice technologies.

Description

Amirkhani, Sima; Stevens, Gunnar; Shajalal, MD; Boden, Alexander (2025): Detecting the Undetectable: Human Judgments and the Challenge of Synthetic Voices. Proceedings of the 12th International Conference on Communities & Technologies (C&T 2025). DOI: 10.48340/ct2025-1030. European Society for Socially Embedded Technologies (EUSSET). Posters and Demos. Siegen, Germany. 21–23 July, 2025

Keywords

Synthetic voices, voice deepfakes, human perception, AI-generated speech, detection cues, deception, user studies

Citation

URI

Endorsement

Review

Supplemented By

Referenced By

Creative Commons license

Except where otherwised noted, this item's license is described as https://creativecommons.org/licenses/by/4.0/