[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3321408.3326664acmotherconferencesArticle/Chapter ViewAbstractPublication Pagesacm-turcConference Proceedingsconference-collections
research-article

Cochleagram-based identification of electronic disguised voice with pitch scaling in the noisy environment

Published: 17 May 2019 Publication History

Abstract

Audio editing software makes voice camouflage easily. That threats to the security and authenticity of audio. Whether the audio forensics can identify voice disguised by software has become an important issue. At the same time, since the audio used in daily life always contains noise, the other key point is improving the anti-noise performance. This paper proposed an algorithm on identification of electronic disguised voice with pitch scaling, which has high anti-noise performance. The algorithm is based on Least Mean Square (LMS) filter and cochleagram, an acoustic characteristic which could reflects the auditory features of human ear. In the algorithm, the noisy voice is sent to the LMS filter for noise reduction. Then cochleagram is extracted from the output signal of LMS filter. The cochleagram is handled at different resolution to construct the Least Mean Square-Multi Resolution Cochleagram (LMS-MRCG) feature. the Gaussian Mixture Model-Universe Background Model (GMM-UBM) is used as detection classifier to identify disguised voice. The pitch scaling type contains 5 different pitch for each speaker's voice. In the end the algorithm needs to identify the pitch type of each speaker. The results show that the algorithm has high detection rate Voice with different genders and languages both can be identified. Under the influence of various environmental noises such as Gaussian white noise, pink noise, factory noise, vehicle noise, etc. the algorithm maintains stable identification performance. Especially in low SNR environment, algorithm can maintain high accuracy of forensic classification. In the environment of noise-free, overall identification rate can reach 97.50%. In the low SNR environment as low as -5dB, identification rate can still remain above 85.83%.

References

[1]
H. J. Wu, Y. Wang and J. W. Huang (2018), Identification of Electronic Disguised Voices, IEEE Transactions on Information Forensics and Security, 9(3), 489--500.
[2]
H. Kaur (2017), Speaker Identification of Disguised Voices Using MFCC Statistical Moment and SVM Classifier. Ph.D. Dissertation. Thapar Institute of Engineering & Technology, 66--79.
[3]
W. C. Cao, H. X. Wang, H. Zhao, etc. (2016), Identification of Electronic Disguised Voices in the Noisy Environment, International Workshop on Digital-Forensics and Watermarking, 75--87.
[4]
D. Wang and G. J. Brown (2016), Fundamentals of computational auditory scene analysis, Computational auditory scene analysis: Principles, algorithms and applications, 41--44.
[5]
R. D. Patterson, K. Robinson, J. Holdsworth, D. McKeown, C. Zhang, and M. Allerhand (1992), Complex sounds and auditory images, Auditory physiology and perception. 83(4), 429--446.
[6]
O. Cheng, W. Abdulla, and Z. Salcic (2005), Performance evaluation of front-end processing for speech recognition systems, The University of Auckland, New Zealand, 621.
[7]
X. Valero and F. Alias (2012), Gammatone cepstral coefficients: Biologically inspired features for non-speech audio classification, IEEE Trans-actions on Multimedia, 14(6), 1684--1689.
[8]
R. V. Sharan and T. J. Moir (2015), Cochleagram Image Feature for Improved Robustness in Sound Recognition, IEEE International Conference on Digital Signal Processing (DSP), 441--444.
[9]
Q. Wang, J. Du, L. R. Dai and C. H. Lee (2018), A Multi objective Learning and Ensembling Approach to High-Performance Speech Enhancement With Compact Neural Network Architectures, IEEE/ACM Transactions on Audio, Speech, and Language Processing, 26(7), 1185--1197.
[10]
J. Chen, Y. Wang, and D. L. Wang (2014), A Feature Study for Classification-Based Speech Separation at Low Signal-to-Noise Ratios, IEEE/ACM Transaction on Audio, Speech, and Language Processing, 22(12), 1993--2002.
[11]
B. Widrow, J. M. McCool, M. G. Larimore, and C. R. Johnson, Jr. (1976), Stationary and nonstationary learning characteristics of the LMS adaptive filter, Proc. IEEE, 64(8), 1151--1162.
[12]
H. K. Raymond and W. J. Edward (1992), A Variable Step Size LMS Algorithm, IEEE Transactions on Signal Processing, 40(7), 1663--1641.
[13]
D. A. Reynolds (1997), Comparison of background normalization methods for text-independent speaker verification, Proceedings of the European Conference on Speech Communication and Technology, 963--966.
[14]
N. Chauhan, M. Chandra (2017), Speaker Recognition and Verification Using Artificial Neural Network. 2017 2nd IEEE International Conference on Wireless Communications, Signal Processing and Networking. Chennai, India, 1147--1149.
[15]
D. Wang, X. W. Zhang, CSLT TRP 20150016: THCHS-30: A Free Chinese Speech Corpus.

Cited By

View all
  • (2024)Cochleagram to Recognize Dysphonia: Auditory Perceptual Analysis for Health InformaticsIEEE Access10.1109/ACCESS.2024.339280812(59198-59210)Online publication date: 2024
  1. Cochleagram-based identification of electronic disguised voice with pitch scaling in the noisy environment

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ACM TURC '19: Proceedings of the ACM Turing Celebration Conference - China
    May 2019
    963 pages
    ISBN:9781450371582
    DOI:10.1145/3321408
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 17 May 2019

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. GMM-UBM
    2. anti-noise
    3. audio forensic
    4. cochleagram
    5. pitch scaling

    Qualifiers

    • Research-article

    Conference

    ACM TURC 2019

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)1
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 14 Jan 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Cochleagram to Recognize Dysphonia: Auditory Perceptual Analysis for Health InformaticsIEEE Access10.1109/ACCESS.2024.339280812(59198-59210)Online publication date: 2024

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media