首页 > 最新文献

Journal of the Audio Engineering Society最新文献

英文 中文
HRTF Clustering for Robust Training of a DNN for Sound Source Localization 用于声源定位的DNN鲁棒训练的HRTF聚类
IF 1.4 4区 工程技术 Q1 Arts and Humanities Pub Date : 2022-12-12 DOI: 10.17743/jaes.2022.0051
Hugh O’Dwyer, F. Boland
This study shows how spherical sound source localization of binaural audio signals in the mismatchedhead-relatedtransferfunction(HRTF)conditioncanbeimprovedbyimplementing HRTF clustering when using machine learning. A new feature set of cross-correlation function, interaural level difference, and Gammatone cepstral coefficients is introduced and shown to outperform state-of-the-art methods in vertical localization in the mismatched HRTF condition by up to 5%. By examining the performance of Deep Neural Networks trained on single HRTF sets from the CIPIC database on other HRTFs, it is shown that HRTF sets can be clustered into groups of similar HRTFs. This results in the formulation of central HRTF sets representativeoftheirspecificcluster.BytrainingamachinelearningalgorithmonthesecentralHRTFs,itisshownthatamorerobustalgorithmcanbetrainedcapableofimprovingsound sourcelocalizationaccuracybyupto13%inthemismatchedHRTFcondition.Concurrently,localizationaccuracyisdecreasedbyapproximately6%inthematchedHRTFcondition,which accountsforlessthan9%ofalltestconditions.ResultsdemonstratethatHRTFclusteringcanvastlyimprovetherobustnessofbinauralsoundsourcelocalizationtounseenHRTFconditions.
本研究表明,在使用机器学习时,如何通过实施HRTF聚类来改善在不匹配的头部相关传递函数(HRTF)条件下双耳音频信号的球形声源定位。引入了一个新的互相关函数、耳间水平差和伽玛酮倒谱系数的特征集,并表明在不匹配的HRTF条件下,该特征集在垂直定位方面优于最先进的方法高达5%。通过检查在来自CIPIC数据库的单个HRTF集上训练的深度神经网络在其他HRTF上的性能,表明HRTF集可以聚类为相似的HRTF组。这导致了具有特定聚类代表性的中心HRTF集合的公式化。通过对这些中心HRTF的机器学习算法进行训练,可以得出结论,在匹配的HRTF条件下,可以训练出一种更完善的算法,能够将声源定位精度提高13%。同时,在预定的HRTF情况下,定位精度降低约6%,结果表明,HRTF聚类可以极大地提高声源定位在HRTF条件下的可信度。
{"title":"HRTF Clustering for Robust Training of a DNN for Sound Source Localization","authors":"Hugh O’Dwyer, F. Boland","doi":"10.17743/jaes.2022.0051","DOIUrl":"https://doi.org/10.17743/jaes.2022.0051","url":null,"abstract":"This study shows how spherical sound source localization of binaural audio signals in the mismatchedhead-relatedtransferfunction(HRTF)conditioncanbeimprovedbyimplementing HRTF clustering when using machine learning. A new feature set of cross-correlation function, interaural level difference, and Gammatone cepstral coefficients is introduced and shown to outperform state-of-the-art methods in vertical localization in the mismatched HRTF condition by up to 5%. By examining the performance of Deep Neural Networks trained on single HRTF sets from the CIPIC database on other HRTFs, it is shown that HRTF sets can be clustered into groups of similar HRTFs. This results in the formulation of central HRTF sets representativeoftheirspecificcluster.BytrainingamachinelearningalgorithmonthesecentralHRTFs,itisshownthatamorerobustalgorithmcanbetrainedcapableofimprovingsound sourcelocalizationaccuracybyupto13%inthemismatchedHRTFcondition.Concurrently,localizationaccuracyisdecreasedbyapproximately6%inthematchedHRTFcondition,which accountsforlessthan9%ofalltestconditions.ResultsdemonstratethatHRTFclusteringcanvastlyimprovetherobustnessofbinauralsoundsourcelocalizationtounseenHRTFconditions.","PeriodicalId":50008,"journal":{"name":"Journal of the Audio Engineering Society","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49622444","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Annotation and Analysis of Recorded Piano Performances on the Web 网上钢琴演奏录音的注释与分析
IF 1.4 4区 工程技术 Q1 Arts and Humanities Pub Date : 2022-12-12 DOI: 10.17743/jaes.2022.0057
Lawrence Fyfe, D. Bedoya, E. Chew
Advancing knowledge and understanding about performed music is hampered by a lack of annotation data for music expressivity. To enable large-scale collection of annotations and explorations of performed music, the authors have created a workflow that is enabled by CosmoNote, a Web-based citizen science tool for annotating musical structures created by the performer and experienced by the listener during expressive piano performances. To enable annotation tasks with CosmoNote, annotators can listen to the recorded performances and view synchronized music visualization layers including the audio waveform, recorded notes, extracted audio features such as loudness and tempo, and score features such as harmonic tension. Annotators have the ability to zoom into specific parts of a performance and see visuals and listen to the audio from just that part. The annotation of performed musical structures is done by using boundaries of varying strengths, regions, comments, and note groups. By analyzing the annotations collected with CosmoNote, performance decisions will be able to be modeled and analyzed in order to aid in the understanding of expressive choices in musical performances and discover the vocabulary of performed musical structures.
缺乏音乐表现力的注释数据阻碍了对演奏音乐的认识和理解。为了实现对演奏音乐的注释和探索的大规模收集,作者创建了一个由CosmoNote支持的工作流,这是一个基于网络的公民科学工具,用于注释由表演者创建的音乐结构,并在富有表现力的钢琴表演中由听众体验。为了使用CosmoNote进行注释任务,注释者可以收听录制的表演并查看同步的音乐可视化层,包括音频波形,录制的音符,提取的音频特征(如响度和速度)以及和声张力等评分特征。注释者有能力放大到表演的特定部分,看到视觉效果,听那个部分的音频。演奏音乐结构的注释是通过使用不同强度、区域、注释和音符组的边界来完成的。通过分析用CosmoNote收集的注释,表演决策将能够建模和分析,以帮助理解音乐表演中的表达选择,并发现表演音乐结构的词汇。
{"title":"Annotation and Analysis of Recorded Piano Performances on the Web","authors":"Lawrence Fyfe, D. Bedoya, E. Chew","doi":"10.17743/jaes.2022.0057","DOIUrl":"https://doi.org/10.17743/jaes.2022.0057","url":null,"abstract":"Advancing knowledge and understanding about performed music is hampered by a lack of annotation data for music expressivity. To enable large-scale collection of annotations and explorations of performed music, the authors have created a workflow that is enabled by CosmoNote, a Web-based citizen science tool for annotating musical structures created by the performer and experienced by the listener during expressive piano performances. To enable annotation tasks with CosmoNote, annotators can listen to the recorded performances and view synchronized music visualization layers including the audio waveform, recorded notes, extracted audio features such as loudness and tempo, and score features such as harmonic tension. Annotators have the ability to zoom into specific parts of a performance and see visuals and listen to the audio from just that part. The annotation of performed musical structures is done by using boundaries of varying strengths, regions, comments, and note groups. By analyzing the annotations collected with CosmoNote, performance decisions will be able to be modeled and analyzed in order to aid in the understanding of expressive choices in musical performances and discover the vocabulary of performed musical structures.","PeriodicalId":50008,"journal":{"name":"Journal of the Audio Engineering Society","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45551012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Four-Directional Ambisonic Spatial Decomposition Method With Reduced Temporal Artifacts 减少时间伪影的四方向双声空间分解方法
IF 1.4 4区 工程技术 Q1 Arts and Humanities Pub Date : 2022-12-12 DOI: 10.17743/jaes.2022.0039
Elias Hoffbauer, M. Frank
For the creation of convincing virtual acoustics of existing rooms and spaces, it is useful to apply measured Ambisonic room impulse responses (ARIRs) as a convolution reverb. Typically, tetrahedral arrays offering only first-order resolution are the preferred practical choice for measurements, because they are easily available and processed. In contrast, higher order is preferred in playback because it is superior in terms of localization accuracy and spatial clarity. There are a number of algorithms that enhance the spatial resolution of first-order ARIRs. However, these algorithms may introduce coloration and artifacts. This paper presents an improvement of the Ambisonic Spatial Decomposition Method by using four directions simultaneously. The additional signals increase the echo density and thereby better preserve the diffuse sound field components during the process of enhancing measured first-order ARIRs to higher orders. An instrumental validation and a series of listening experiments compare the proposed Four-Directional Ambisonic Spatial Decomposition Method to other existing algorithms and prove its similarity to the best algorithm in terms of enhanced spatial clarity and coloration while producing the least artifacts.
为了在现有房间和空间中创建令人信服的虚拟声学,将测量的双声房间脉冲响应(ARIRs)应用于卷积混响是有用的。通常,四面体阵列只提供一阶分辨率是测量的首选实用选择,因为它们易于获得和处理。相比之下,在回放中首选高阶,因为它在定位精度和空间清晰度方面更胜一筹。有许多算法可以提高一阶arir的空间分辨率。然而,这些算法可能会引入颜色和伪影。本文提出了一种双声空间分解方法的改进方法,即同时使用四个方向。在将测量到的一阶arir提高到更高阶的过程中,附加信号增加了回波密度,从而更好地保留了漫射声场分量。通过仪器验证和一系列的听力实验,将提出的四方向双声空间分解方法与其他现有算法进行了比较,并证明了其在增强空间清晰度和色彩方面与最佳算法相似,同时产生的伪影最少。
{"title":"Four-Directional Ambisonic Spatial Decomposition Method With Reduced Temporal Artifacts","authors":"Elias Hoffbauer, M. Frank","doi":"10.17743/jaes.2022.0039","DOIUrl":"https://doi.org/10.17743/jaes.2022.0039","url":null,"abstract":"For the creation of convincing virtual acoustics of existing rooms and spaces, it is useful to apply measured Ambisonic room impulse responses (ARIRs) as a convolution reverb. Typically, tetrahedral arrays offering only first-order resolution are the preferred practical choice for measurements, because they are easily available and processed. In contrast, higher order is preferred in playback because it is superior in terms of localization accuracy and spatial clarity. There are a number of algorithms that enhance the spatial resolution of first-order ARIRs. However, these algorithms may introduce coloration and artifacts. This paper presents an improvement of the Ambisonic Spatial Decomposition Method by using four directions simultaneously. The additional signals increase the echo density and thereby better preserve the diffuse sound field components during the process of enhancing measured first-order ARIRs to higher orders. An instrumental validation and a series of listening experiments compare the proposed Four-Directional Ambisonic Spatial Decomposition Method to other existing algorithms and prove its similarity to the best algorithm in terms of enhanced spatial clarity and coloration while producing the least artifacts.","PeriodicalId":50008,"journal":{"name":"Journal of the Audio Engineering Society","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43340675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Web MIDI API: State of the Art and Future Perspectives Web MIDI API:现状与未来展望
IF 1.4 4区 工程技术 Q1 Arts and Humanities Pub Date : 2022-12-12 DOI: 10.17743/jaes.2022.0028
A. Baratè, L. A. Ludovico
The Web MIDI API is intended to connect a browser app with Musical Instrument Digital Interface (MIDI) devices and make them interact. Such an interface deals with exchanging MIDI messages between a browser app and an external MIDI system, either physical or virtual. The standardization by the World Wide Web (W3C) Consortium started about 10 years ago, with a first public draft published on October 2012, and the process is not over yet. Because this technology can pave the way for innovative applications in musical and extra-musical fields, the present paper aims to unveil the main features of the API, remarking its advantages and drawbacks and discussing several applications that could take benefit from its adoption.
Web MIDI API旨在将浏览器应用程序与乐器数字接口(MIDI)设备连接起来,并使它们进行交互。这样的接口处理浏览器应用程序和外部MIDI系统(物理或虚拟)之间的MIDI消息交换。万维网联盟的标准化始于大约10年前,2012年10月发布了第一份公开草案,但这一过程尚未结束。由于这项技术可以为音乐和音乐外领域的创新应用铺平道路,本文旨在揭示API的主要功能,指出其优点和缺点,并讨论可以从其采用中受益的几个应用。
{"title":"Web MIDI API: State of the Art and Future Perspectives","authors":"A. Baratè, L. A. Ludovico","doi":"10.17743/jaes.2022.0028","DOIUrl":"https://doi.org/10.17743/jaes.2022.0028","url":null,"abstract":"The Web MIDI API is intended to connect a browser app with Musical Instrument Digital Interface (MIDI) devices and make them interact. Such an interface deals with exchanging MIDI messages between a browser app and an external MIDI system, either physical or virtual. The standardization by the World Wide Web (W3C) Consortium started about 10 years ago, with a first public draft published on October 2012, and the process is not over yet. Because this technology can pave the way for innovative applications in musical and extra-musical fields, the present paper aims to unveil the main features of the API, remarking its advantages and drawbacks and discussing several applications that could take benefit from its adoption.","PeriodicalId":50008,"journal":{"name":"Journal of the Audio Engineering Society","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42887388","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
JSPatcher, a Visual Programming Environment for Building High-Performance Web Audio Applications JSPatcher,一个用于构建高性能Web音频应用程序的可视化编程环境
IF 1.4 4区 工程技术 Q1 Arts and Humanities Pub Date : 2022-12-12 DOI: 10.17743/jaes.2022.0056
Shihong Ren, L. Pottier, M. Buffa, Yang Yu
{"title":"JSPatcher, a Visual Programming Environment for Building High-Performance Web Audio Applications","authors":"Shihong Ren, L. Pottier, M. Buffa, Yang Yu","doi":"10.17743/jaes.2022.0056","DOIUrl":"https://doi.org/10.17743/jaes.2022.0056","url":null,"abstract":"","PeriodicalId":50008,"journal":{"name":"Journal of the Audio Engineering Society","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46732022","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Web-Based Networked Music Performances via WebRTC: A Low-Latency PCM Audio Solution 基于WebRTC的网络音乐表演:一种低延迟PCM音频解决方案
IF 1.4 4区 工程技术 Q1 Arts and Humanities Pub Date : 2022-12-12 DOI: 10.17743/jaes.2022.0021
Matteo Sacchetto, P. Gastaldi, C. Chafe, C. Rottondi, A. Servetti
Nowadays, widely used videoconferencing software has been diffused even further by the social distancing measures adopted during the SARS-CoV-2 pandemic. However, none of the Web-based solutions currently available support high-fidelity stereo audio streaming, which is a fundamental prerequisite for networked music applications. This is mainly because of the fact that the WebRTC RTCPeerConnection standard or Web-based audio streaming do not handle uncompressed audio formats. To overcome that limitation, an implementation of 16-bit pulse code modulation (PCM) stereo audio transmission on top of the WebRTC RTCDataChannel, leveraging Web Audio and AudioWorklets, is discussed. Results obtained with multiple configurations, browsers, and operating systems show that the proposed approach outperforms the WebRTC RTCPeerConnection standard in terms of audio quality and latency, which in the authors' best case to date has been reduced to only 40 ms between two MacBooks on a local area network. © 2022 Audio Engineering Society. All rights reserved.
如今,由于严重急性呼吸系统综合征冠状病毒2型疫情期间采取的社交距离措施,广泛使用的视频会议软件得到了进一步推广。然而,目前可用的基于Web的解决方案都不支持高保真立体声音频流,这是网络音乐应用程序的基本前提。这主要是因为WebRTC RTCPeerConnection标准或基于Web的音频流不处理未压缩的音频格式。为了克服这一限制,讨论了利用Web audio和AudioWorklets在WebRTC RTCDataChannel之上实现16位脉冲编码调制(PCM)立体声音频传输。使用多种配置、浏览器和操作系统获得的结果表明,所提出的方法在音频质量和延迟方面优于WebRTC RTCPeerConnection标准,在作者迄今为止的最佳情况下,局域网上两台MacBook之间的延迟已降至仅40毫秒。©2022音频工程学会。保留所有权利。
{"title":"Web-Based Networked Music Performances via WebRTC: A Low-Latency PCM Audio Solution","authors":"Matteo Sacchetto, P. Gastaldi, C. Chafe, C. Rottondi, A. Servetti","doi":"10.17743/jaes.2022.0021","DOIUrl":"https://doi.org/10.17743/jaes.2022.0021","url":null,"abstract":"Nowadays, widely used videoconferencing software has been diffused even further by the social distancing measures adopted during the SARS-CoV-2 pandemic. However, none of the Web-based solutions currently available support high-fidelity stereo audio streaming, which is a fundamental prerequisite for networked music applications. This is mainly because of the fact that the WebRTC RTCPeerConnection standard or Web-based audio streaming do not handle uncompressed audio formats. To overcome that limitation, an implementation of 16-bit pulse code modulation (PCM) stereo audio transmission on top of the WebRTC RTCDataChannel, leveraging Web Audio and AudioWorklets, is discussed. Results obtained with multiple configurations, browsers, and operating systems show that the proposed approach outperforms the WebRTC RTCPeerConnection standard in terms of audio quality and latency, which in the authors' best case to date has been reduced to only 40 ms between two MacBooks on a local area network. © 2022 Audio Engineering Society. All rights reserved.","PeriodicalId":50008,"journal":{"name":"Journal of the Audio Engineering Society","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47508527","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Expanding the Frontiers of Web Audio With Autoencoders and JavaScript 使用自动编码器和JavaScript扩展网络音频的前沿
IF 1.4 4区 工程技术 Q1 Arts and Humanities Pub Date : 2022-12-12 DOI: 10.17743/jaes.2022.0046
Mateo Cámara, José Luis Blanco
{"title":"Expanding the Frontiers of Web Audio With Autoencoders and JavaScript","authors":"Mateo Cámara, José Luis Blanco","doi":"10.17743/jaes.2022.0046","DOIUrl":"https://doi.org/10.17743/jaes.2022.0046","url":null,"abstract":"","PeriodicalId":50008,"journal":{"name":"Journal of the Audio Engineering Society","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48234855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Voltage-Controlled Amplifier Designs for a Field-Programmable Analog Array 现场可编程模拟阵列的压控放大器设计
IF 1.4 4区 工程技术 Q1 Arts and Humanities Pub Date : 2022-12-12 DOI: 10.17743/jaes.2022.0054
A. Lanterman, J. Hasler
{"title":"Voltage-Controlled Amplifier Designs for a Field-Programmable Analog Array","authors":"A. Lanterman, J. Hasler","doi":"10.17743/jaes.2022.0054","DOIUrl":"https://doi.org/10.17743/jaes.2022.0054","url":null,"abstract":"","PeriodicalId":50008,"journal":{"name":"Journal of the Audio Engineering Society","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43322857","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Phase-Aware Transformations in Variational Autoencoders for Audio Effects 音频效果变分自编码器的相位感知变换
IF 1.4 4区 工程技术 Q1 Arts and Humanities Pub Date : 2022-11-02 DOI: 10.17743/jaes.2022.0042
Mateo Cámara, José Luis Blanco
{"title":"Phase-Aware Transformations in Variational Autoencoders for Audio Effects","authors":"Mateo Cámara, José Luis Blanco","doi":"10.17743/jaes.2022.0042","DOIUrl":"https://doi.org/10.17743/jaes.2022.0042","url":null,"abstract":"","PeriodicalId":50008,"journal":{"name":"Journal of the Audio Engineering Society","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2022-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43402446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Audio Augmented Reality: A Systematic Review of Technologies, Applications, and Future Research Directions 音频增强现实:技术、应用和未来研究方向的系统回顾
IF 1.4 4区 工程技术 Q1 Arts and Humanities Pub Date : 2022-11-02 DOI: 10.17743/jaes.2022.0048
Jing Yang, Amit Barde, M. Billinghurst
Audio Augmented Reality (AAR) aims to augment people’s auditory perception of the real world by synthesizing virtual spatialized sounds. AAR has begun to attract more research interest in recent years, especially because Augmented Reality (AR) applications are becoming more commonly available on mobile and wearable devices. However, because audio augmentation is relatively under-studied in the wider AR community, AAR needs to be further investigated in order to be widely used in different applications. This paper systematically reports on the technologies used in past studies to realize AAR and provides an overview of AAR applications. A total of 563 publications indexed on Scopus and Google Scholar were reviewed, and from these, 117 of the most impactful papers were identified and summarized in more detail. As one of the first systematic reviews of AAR, this paper presents an overall landscape of AAR, discusses the development trends in techniques and applications, and indicates challenges and opportunities for future research. For researchers and practitioners in related fields, this review aims to provide inspirations and guidance for conducting AAR research in the future.
音频增强现实(AAR)旨在通过合成虚拟空间化的声音来增强人们对现实世界的听觉感知。近年来,AAR开始吸引更多的研究兴趣,特别是因为增强现实(AR)应用在移动和可穿戴设备上越来越普遍。然而,由于音频增强在更广泛的AR社区中研究相对较少,因此需要对AAR进行进一步研究,以便在不同的应用中广泛使用。本文系统地报告了过去研究中用于实现AAR的技术,并对AAR的应用进行了概述。共审查了563篇Scopus和Google Scholar上的索引出版物,从中确定并更详细地总结了117篇最具影响力的论文。作为对AAR的首次系统综述之一,本文介绍了AAR的总体概况,讨论了技术和应用的发展趋势,并指出了未来研究的挑战和机遇。本综述旨在为相关领域的研究人员和从业者提供启示和指导,以期在未来开展AAR研究。
{"title":"Audio Augmented Reality: A Systematic Review of Technologies, Applications, and Future Research Directions","authors":"Jing Yang, Amit Barde, M. Billinghurst","doi":"10.17743/jaes.2022.0048","DOIUrl":"https://doi.org/10.17743/jaes.2022.0048","url":null,"abstract":"Audio Augmented Reality (AAR) aims to augment people’s auditory perception of the real world by synthesizing virtual spatialized sounds. AAR has begun to attract more research interest in recent years, especially because Augmented Reality (AR) applications are becoming more commonly available on mobile and wearable devices. However, because audio augmentation is relatively under-studied in the wider AR community, AAR needs to be further investigated in order to be widely used in different applications. This paper systematically reports on the technologies used in past studies to realize AAR and provides an overview of AAR applications. A total of 563 publications indexed on Scopus and Google Scholar were reviewed, and from these, 117 of the most impactful papers were identified and summarized in more detail. As one of the first systematic reviews of AAR, this paper presents an overall landscape of AAR, discusses the development trends in techniques and applications, and indicates challenges and opportunities for future research. For researchers and practitioners in related fields, this review aims to provide inspirations and guidance for conducting AAR research in the future.","PeriodicalId":50008,"journal":{"name":"Journal of the Audio Engineering Society","volume":null,"pages":null},"PeriodicalIF":1.4,"publicationDate":"2022-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45338639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
期刊
Journal of the Audio Engineering Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1