{"title":"Performance of Phase-Space Voicing-State Classification for Co-Channel Speech","authors":"Y. Mahgoub, R. Dansereau","doi":"10.1109/IMTC.2005.1604216","DOIUrl":null,"url":null,"abstract":"This paper discusses the performance of a classification algorithm that is capable of determining the voicing-state of co-channel speech. The algorithm uses some features of the reconstructed phase-space of the speech data as a measure to identify the three voicing-states of co-channel speech; unvoiced/unvoiced (U/U), voiced/unvoiced (V/U), and voiced/voiced (V/V). The proposed method requires neither a priori information nor speech training data. Nonetheless, simulation results show enhanced performance in identifying the three voicing-states compared to other existing techniques. The algorithm also shows a reliable performance for different SIR values as well as different levels of background noise","PeriodicalId":244878,"journal":{"name":"2005 IEEE Instrumentationand Measurement Technology Conference Proceedings","volume":"87 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2005-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2005 IEEE Instrumentationand Measurement Technology Conference Proceedings","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IMTC.2005.1604216","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
This paper discusses the performance of a classification algorithm that is capable of determining the voicing-state of co-channel speech. The algorithm uses some features of the reconstructed phase-space of the speech data as a measure to identify the three voicing-states of co-channel speech; unvoiced/unvoiced (U/U), voiced/unvoiced (V/U), and voiced/voiced (V/V). The proposed method requires neither a priori information nor speech training data. Nonetheless, simulation results show enhanced performance in identifying the three voicing-states compared to other existing techniques. The algorithm also shows a reliable performance for different SIR values as well as different levels of background noise