首页 > 最新文献

Journal of the Acoustical Society of Korea最新文献

英文 中文
A quantitative analysis of synthetic aperture sonar image distortion according to sonar platform motion parameters 根据声纳平台运动参数定量分析合成孔径声纳图像畸变
IF 0.4 Q4 ACOUSTICS Pub Date : 2021-07-01 DOI: 10.7776/ASK.2021.40.4.382
Sea-Moon Kim and Sung-Hoon Byun
Synthetic aperture sonars as well as side scan sonars or multibeam echo sounders have been commercialized and are widely used for seafloor imaging. In Korea related research such as the development of a towed synthetic aperture sonar system is underway. In order to obtain high-resolution synthetic aperture sonar images, it is necessary to accurately estimate the platform motion on which it is installed, and a precise underwater navigation system is required. In this paper we are going to provide reference data for determining the required navigation accuracy and precision of navigation sensors by quantitatively analyzing how much distortion of the sonar images occurs according to motion characteristics of the platform equipped with the synthetic aperture sonar. Five types of motions are considered and normalized root mean square error is defined for quantitative analysis. Simulation for error analysis with parameter variation of motion characteristics results in that yaw and sway motion causes the largest image distortion whereas the effect of pitch and heave motion is not significant.
合成孔径声呐、侧扫声呐或多波束回声测深仪已经商品化,广泛用于海底成像。国内正在开发拖曳式合成孔径声呐系统等相关研究。为了获得高分辨率合成孔径声呐图像,需要准确估计其所安装的平台运动,并且需要精确的水下导航系统。本文将根据合成孔径声呐平台的运动特性,定量分析声呐图像的畸变程度,为确定所需的导航精度和导航传感器的精度提供参考数据。考虑了五种运动类型,并定义了归一化均方根误差进行定量分析。对运动特性参数变化的误差分析结果表明,偏航和摇摆运动对图像畸变的影响最大,俯仰和垂升运动对图像畸变的影响不显著。
{"title":"A quantitative analysis of synthetic aperture sonar image distortion according to sonar platform motion parameters","authors":"Sea-Moon Kim and Sung-Hoon Byun","doi":"10.7776/ASK.2021.40.4.382","DOIUrl":"https://doi.org/10.7776/ASK.2021.40.4.382","url":null,"abstract":"Synthetic aperture sonars as well as side scan sonars or multibeam echo sounders have been commercialized and are widely used for seafloor imaging. In Korea related research such as the development of a towed synthetic aperture sonar system is underway. In order to obtain high-resolution synthetic aperture sonar images, it is necessary to accurately estimate the platform motion on which it is installed, and a precise underwater navigation system is required. In this paper we are going to provide reference data for determining the required navigation accuracy and precision of navigation sensors by quantitatively analyzing how much distortion of the sonar images occurs according to motion characteristics of the platform equipped with the synthetic aperture sonar. Five types of motions are considered and normalized root mean square error is defined for quantitative analysis. Simulation for error analysis with parameter variation of motion characteristics results in that yaw and sway motion causes the largest image distortion whereas the effect of pitch and heave motion is not significant.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"40 1","pages":"382-390"},"PeriodicalIF":0.4,"publicationDate":"2021-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41804028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An explorative study on the perceived emotion of music: according to cognitive styles of music listening 音乐感知情感的探索性研究:基于音乐聆听的认知风格
IF 0.4 Q4 ACOUSTICS Pub Date : 2021-07-01 DOI: 10.7776/ASK.2021.40.4.290
Jin Hee Choi and Hyun Ju Chong
The purpose of this study was to examine the perceived emotion of music according to cognitive styles of music listening. A total of 91 music-related graduate students participated in this study. They were given a questionnaire about perceived emotions of music, musical elements, and Music Empathizing-Music Systemizing Inventory. To analyze statistically, Descriptive statistics, paired t-test, ANalysis Of VAriance (ANOVA), multivariate analysis, and Pearson correlation analysis were conducted. Results showed that participants had relatively universal experience in perceived emotions of both types of music, and also showed that musical elements contributed to the experience differed by cognitive styles of music listening.
本研究的目的是根据音乐听力的认知风格来检验音乐的感知情绪。共有91名音乐相关研究生参与了这项研究。他们接受了一份关于音乐情感感知、音乐元素和音乐移情音乐系统化清单的问卷调查。为了进行统计学分析,进行了描述性统计、配对t检验、方差分析(ANOVA)、多变量分析和Pearson相关分析。结果表明,参与者在两种类型的音乐的感知情绪方面都有相对普遍的体验,而且音乐元素对体验的贡献因音乐听力的认知风格而异。
{"title":"An explorative study on the perceived emotion of music: according to cognitive styles of music listening","authors":"Jin Hee Choi and Hyun Ju Chong","doi":"10.7776/ASK.2021.40.4.290","DOIUrl":"https://doi.org/10.7776/ASK.2021.40.4.290","url":null,"abstract":"The purpose of this study was to examine the perceived emotion of music according to cognitive styles of music listening. A total of 91 music-related graduate students participated in this study. They were given a questionnaire about perceived emotions of music, musical elements, and Music Empathizing-Music Systemizing Inventory. To analyze statistically, Descriptive statistics, paired t-test, ANalysis Of VAriance (ANOVA), multivariate analysis, and Pearson correlation analysis were conducted. Results showed that participants had relatively universal experience in perceived emotions of both types of music, and also showed that musical elements contributed to the experience differed by cognitive styles of music listening.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"40 1","pages":"290-296"},"PeriodicalIF":0.4,"publicationDate":"2021-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49035354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Measurements of mid-frequency transmission loss in shallow waters off the East Sea: Comparison with Rayleigh reflection model and high-frequency bottom loss model 东海浅海中频传输损耗的测量:与瑞利反射模型和高频底损耗模型的比较
IF 0.4 Q4 ACOUSTICS Pub Date : 2021-07-01 DOI: 10.7776/ASK.2021.40.4.297
D. Lee, Raegeun Oh, J. Choi, Seongil Kim, Hyuckjong Kwon
When sound waves propagate over long distances in shallow water, measured transmission loss is greater than predicted one using underwater acoustic model with the Rayleigh reflection model due to inhomogeneity of the bottom. Accordingly, the US Navy predicts sound wave propagation by applying the empirical formula-based High Frequency Bottom Loss (HFBL) model. In this study, the measurement and analysis of transmission loss was conducted using mid-frequency (2.3 kHz, 3 kHz) in the shallow water of the East Sea in summer. BELLHOP eigenray tracing output shows that only sound waves with lower grazing angle than the critical angle propagate long distances for several kilometers or more, and the difference between the predicted transmission loss based on the Rayleigh reflection model and the measured transmission loss tend to increase along the propagation range. By comparing the Rayleigh reflection model and the HFBL model at the high grazing angle region, the bottom province, the input value of the HFBL model, is estimated and BELLHOP transmission loss with HFBL model is compared to measured transmission loss. As a result, it agrees well with the measurements of transmission loss.
当声波在浅水中长距离传播时,由于底部的非均匀性,使用瑞利反射模型的水声模型测量的传播损失大于预测的传播损失。因此,美国海军采用基于经验公式的高频底损(HFBL)模型预测声波传播。本研究利用东海浅海夏季中频(2.3 kHz, 3 kHz)对传输损耗进行测量与分析。BELLHOP特征射线跟踪输出表明,只有小于临界角的声波才能传播数公里以上的距离,并且基于瑞利反射模型的预测传输损耗与实测传输损耗之间的差值沿传播范围呈增大趋势。通过比较高掠角区域的Rayleigh反射模型和HFBL模型,估计HFBL模型的底省输入值,并将HFBL模型下的BELLHOP传输损耗与实测传输损耗进行比较。结果表明,该方法与传输损耗测量结果吻合较好。
{"title":"Measurements of mid-frequency transmission loss in shallow waters off the East Sea: Comparison with Rayleigh reflection model and high-frequency bottom loss model","authors":"D. Lee, Raegeun Oh, J. Choi, Seongil Kim, Hyuckjong Kwon","doi":"10.7776/ASK.2021.40.4.297","DOIUrl":"https://doi.org/10.7776/ASK.2021.40.4.297","url":null,"abstract":"When sound waves propagate over long distances in shallow water, measured transmission loss is greater than predicted one using underwater acoustic model with the Rayleigh reflection model due to inhomogeneity of the bottom. Accordingly, the US Navy predicts sound wave propagation by applying the empirical formula-based High Frequency Bottom Loss (HFBL) model. In this study, the measurement and analysis of transmission loss was conducted using mid-frequency (2.3 kHz, 3 kHz) in the shallow water of the East Sea in summer. BELLHOP eigenray tracing output shows that only sound waves with lower grazing angle than the critical angle propagate long distances for several kilometers or more, and the difference between the predicted transmission loss based on the Rayleigh reflection model and the measured transmission loss tend to increase along the propagation range. By comparing the Rayleigh reflection model and the HFBL model at the high grazing angle region, the bottom province, the input value of the HFBL model, is estimated and BELLHOP transmission loss with HFBL model is compared to measured transmission loss. As a result, it agrees well with the measurements of transmission loss.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"40 1","pages":"297-303"},"PeriodicalIF":0.4,"publicationDate":"2021-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45128356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Segment unit shuffling layer in deep neural networks for text-independent speaker verification 深度神经网络中用于文本无关说话人验证的分段单元混洗层
IF 0.4 Q4 ACOUSTICS Pub Date : 2021-03-01 DOI: 10.7776/ASK.2021.40.2.148
Ju-Sung Heo, Hye-jin Shim, Ju-ho Kim, Ha-jin Yu
Text-Independent speaker verification needs to extract text-independent speaker embedding to improve generalization performance. However, deep neural networks that depend on training data have the potential to overfit text information instead of learning the speaker information when repeatedly learning from the identical time series. In this paper, to prevent the overfitting, we propose a segment unit shuffling layer that divides and rearranges the input layer or a hidden layer along the time axis, thus mixes the time series information. Since the segment unit shuffling layer can be applied not only to the input layer but also to the hidden layers, it can be used as generalization technique in the hidden layer, which is known to be effective compared to the generalization technique in the input layer, and can be applied simultaneously with data augmentation. In addition, the degree of distortion can be adjusted by adjusting the unit size of the segment. We observe that the performance of text-independent speaker verification is improved compared to the baseline when the proposed segment unit shuffling layer is applied.
文本无关说话人验证需要提取文本无关说话人嵌入,以提高泛化性能。然而,当从相同的时间序列中重复学习时,依赖于训练数据的深度神经网络有可能过度拟合文本信息,而不是学习说话者信息。在本文中,为了防止过拟合,我们提出了一个分段单元混洗层,它沿着时间轴划分和重新排列输入层或隐藏层,从而混合时间序列信息。由于分段单元混洗层不仅可以应用于输入层,还可以应用于隐藏层,因此它可以用作隐藏层中的泛化技术,与输入层中的推广技术相比,它是有效的,并且可以与数据扩充同时应用。此外,可以通过调整片段的单位大小来调整失真程度。我们观察到,当应用所提出的分段单元混洗层时,与基线相比,与文本无关的说话人验证的性能有所提高。
{"title":"Segment unit shuffling layer in deep neural networks for text-independent speaker verification","authors":"Ju-Sung Heo, Hye-jin Shim, Ju-ho Kim, Ha-jin Yu","doi":"10.7776/ASK.2021.40.2.148","DOIUrl":"https://doi.org/10.7776/ASK.2021.40.2.148","url":null,"abstract":"Text-Independent speaker verification needs to extract text-independent speaker embedding to improve generalization performance. However, deep neural networks that depend on training data have the potential to overfit text information instead of learning the speaker information when repeatedly learning from the identical time series. In this paper, to prevent the overfitting, we propose a segment unit shuffling layer that divides and rearranges the input layer or a hidden layer along the time axis, thus mixes the time series information. Since the segment unit shuffling layer can be applied not only to the input layer but also to the hidden layers, it can be used as generalization technique in the hidden layer, which is known to be effective compared to the generalization technique in the input layer, and can be applied simultaneously with data augmentation. In addition, the degree of distortion can be adjusted by adjusting the unit size of the segment. We observe that the performance of text-independent speaker verification is improved compared to the baseline when the proposed segment unit shuffling layer is applied.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"40 1","pages":"148-154"},"PeriodicalIF":0.4,"publicationDate":"2021-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47949865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Side scan sonar image super-resolution using an improved initialization structure 采用改进初始化结构的侧扫声纳图像超分辨率
IF 0.4 Q4 ACOUSTICS Pub Date : 2021-03-01 DOI: 10.7776/ASK.2021.40.2.121
Junyeop Lee, Bonhwa Ku, Wanjin Kim, Hanseok Ko
This paper deals with a super-resolution that improves the resolution of side scan sonar images using learning-based compressive sensing. Learning-based compressive sensing combined with deep learning and compressive sensing takes a structure of a feed-forward network and parameters are set automatically through learning. In particular, we propose a method that can effectively extract additional information required in the super-resolution process through various initialization methods. Representative experimental results show that the proposed method provides improved performance in terms of Peak Signal-to-Noise Ratio (PSNR) and Structure Similarity Index Measure (SSIM) than conventional methods.
本文提出了一种利用基于学习的压缩传感提高侧扫声纳图像分辨率的超分辨率方法。基于学习的压缩感知与深度学习和压缩感知相结合,采用前馈网络的结构,并通过学习自动设置参数。特别是,我们提出了一种方法,可以通过各种初始化方法有效地提取超分辨率过程中所需的附加信息。代表性的实验结果表明,与传统方法相比,该方法在峰值信噪比(PSNR)和结构相似性指数测量(SSIM)方面具有更好的性能。
{"title":"Side scan sonar image super-resolution using an improved initialization structure","authors":"Junyeop Lee, Bonhwa Ku, Wanjin Kim, Hanseok Ko","doi":"10.7776/ASK.2021.40.2.121","DOIUrl":"https://doi.org/10.7776/ASK.2021.40.2.121","url":null,"abstract":"This paper deals with a super-resolution that improves the resolution of side scan sonar images using learning-based compressive sensing. Learning-based compressive sensing combined with deep learning and compressive sensing takes a structure of a feed-forward network and parameters are set automatically through learning. In particular, we propose a method that can effectively extract additional information required in the super-resolution process through various initialization methods. Representative experimental results show that the proposed method provides improved performance in terms of Peak Signal-to-Noise Ratio (PSNR) and Structure Similarity Index Measure (SSIM) than conventional methods.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"40 1","pages":"121-129"},"PeriodicalIF":0.4,"publicationDate":"2021-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46567183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A robust data association gate method of non-linear target tracking in dense cluttered environment 密集杂波环境下非线性目标跟踪的鲁棒数据关联门方法
IF 0.4 Q4 ACOUSTICS Pub Date : 2021-03-01 DOI: 10.7776/ASK.2021.40.2.109
Seong-Weon Kim, Taek-ik Kwon, Hyeon‑Deok Cho
{"title":"A robust data association gate method of non-linear target tracking in dense cluttered environment","authors":"Seong-Weon Kim, Taek-ik Kwon, Hyeon‑Deok Cho","doi":"10.7776/ASK.2021.40.2.109","DOIUrl":"https://doi.org/10.7776/ASK.2021.40.2.109","url":null,"abstract":"","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"40 1","pages":"109-120"},"PeriodicalIF":0.4,"publicationDate":"2021-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41790298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of deep learning-based holographic ultrasound generation algorithm 基于深度学习的全息超声生成算法的开发
IF 0.4 Q4 ACOUSTICS Pub Date : 2021-03-01 DOI: 10.7776/ASK.2021.40.2.169
Moon Hwan Lee and Jae Youn Hwang
Recently, an ultrasound hologram and its applications have gained attention in the ultrasound research field. However, the determination technique of transmit signal phases, which generate a hologram, has not been significantly advanced from the previous algorithms which are time-consuming iterative methods. Thus, we applied the deep learning technique, which has been previously adopted to generate an optical hologram, to generate an ultrasound hologram. We further examined the Deep learning-based Holographic Ultrasound Generation algorithm (Deep-HUG). We implement the U-Net-based algorithm and examine its generalizability by training on a dataset, which consists of randomly distributed disks, and testing on the alphabets (A-Z). Furthermore, we compare the Deep-HUG with the previous algorithm in terms of computation time, accuracy, and uniformity. It was found that the accuracy and uniformity of the Deep-HUG are somewhat lower than those of the previous algorithm whereas the computation time is 190 times faster than that of the previous algorithm, demonstrating that Deep-HUG has potential as a useful technique to rapidly generate an ultrasound hologram for various applications.
近年来,超声全息图及其应用在超声研究领域引起了人们的关注。然而,产生全息图的发射信号相位的确定技术与以前的算法相比并没有显著进步,这些算法是耗时的迭代方法。因此,我们应用了以前用于生成光学全息图的深度学习技术来生成超声全息图。我们进一步研究了基于深度学习的全息超声生成算法(Deep-HUG)。我们实现了基于U-Net的算法,并通过在由随机分布的磁盘组成的数据集上进行训练和在字母表(a-Z)上进行测试来检验其可推广性。此外,我们还将Deep HUG与以前的算法在计算时间、精度和一致性方面进行了比较。研究发现,Deep HUG的精度和均匀性略低于先前算法的精度和一致性,而计算时间比先前算法快190倍,这表明Deep HUG作为一种用于各种应用的快速生成超声全息图的有用技术具有潜力。
{"title":"Development of deep learning-based holographic ultrasound generation algorithm","authors":"Moon Hwan Lee and Jae Youn Hwang","doi":"10.7776/ASK.2021.40.2.169","DOIUrl":"https://doi.org/10.7776/ASK.2021.40.2.169","url":null,"abstract":"Recently, an ultrasound hologram and its applications have gained attention in the ultrasound research field. However, the determination technique of transmit signal phases, which generate a hologram, has not been significantly advanced from the previous algorithms which are time-consuming iterative methods. Thus, we applied the deep learning technique, which has been previously adopted to generate an optical hologram, to generate an ultrasound hologram. We further examined the Deep learning-based Holographic Ultrasound Generation algorithm (Deep-HUG). We implement the U-Net-based algorithm and examine its generalizability by training on a dataset, which consists of randomly distributed disks, and testing on the alphabets (A-Z). Furthermore, we compare the Deep-HUG with the previous algorithm in terms of computation time, accuracy, and uniformity. It was found that the accuracy and uniformity of the Deep-HUG are somewhat lower than those of the previous algorithm whereas the computation time is 190 times faster than that of the previous algorithm, demonstrating that Deep-HUG has potential as a useful technique to rapidly generate an ultrasound hologram for various applications.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"40 1","pages":"169-175"},"PeriodicalIF":0.4,"publicationDate":"2021-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48661920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Performance analysis of weakly-supervised sound event detection system based on the mean-teacher convolutional recurrent neural network model 基于均值-教师卷积递归神经网络模型的弱监督声音事件检测系统性能分析
IF 0.4 Q4 ACOUSTICS Pub Date : 2021-03-01 DOI: 10.7776/ASK.2021.40.2.139
Seokjin Lee
This paper introduces and implements a Sound Event Detection (SED) system based on weaklysupervised learning where only part of the data is labeled, and analyzes the effect of parameters. The SED system estimates the classes and onset/offset times of events in the acoustic signal. In order to train the model, all information on the event class and onset/offset times must be provided. Unfortunately, the onset/offset times are hard to be labeled exactly. Therefore, in the weakly-supervised task, the SED model is trained by “strongly labeled data” including the event class and activations, “weakly labeled data” including the event class, and “unlabeled data” without any label. Recently, the SED systems using the mean-teacher model are widely used for the task with several parameters. These parameters should be chosen carefully because they may affect the performance. In this paper, performance analysis was performed on parameters, such as the feature, moving average parameter, weight of the consistency cost function, ramp-up length, and maximum learning rate, using the data of DCASE 2020 Task 4. Effects and the optimal values of the parameters were discussed.
本文介绍并实现了一种基于弱监督学习的声音事件检测系统,该系统只对部分数据进行标记,并分析了参数的影响。SED系统估计声信号中事件的类别和开始/偏移时间。为了训练模型,必须提供关于事件类和开始/偏移时间的所有信息。不幸的是,开始/偏移时间很难准确标记。因此,在弱监督任务中,SED模型由“强标记数据”(包括事件类和激活)、“弱标记数据”(包括事件类)和没有任何标记的“未标记数据”来训练。近年来,使用均值-教师模型的SED系统被广泛用于多参数任务。这些参数应该谨慎选择,因为它们可能会影响性能。本文利用DCASE 2020 Task 4的数据,对特征、移动平均参数、一致性代价函数权重、爬坡长度、最大学习率等参数进行性能分析。讨论了各参数的影响及最优值。
{"title":"Performance analysis of weakly-supervised sound event detection system based on the mean-teacher convolutional recurrent neural network model","authors":"Seokjin Lee","doi":"10.7776/ASK.2021.40.2.139","DOIUrl":"https://doi.org/10.7776/ASK.2021.40.2.139","url":null,"abstract":"This paper introduces and implements a Sound Event Detection (SED) system based on weaklysupervised learning where only part of the data is labeled, and analyzes the effect of parameters. The SED system estimates the classes and onset/offset times of events in the acoustic signal. In order to train the model, all information on the event class and onset/offset times must be provided. Unfortunately, the onset/offset times are hard to be labeled exactly. Therefore, in the weakly-supervised task, the SED model is trained by “strongly labeled data” including the event class and activations, “weakly labeled data” including the event class, and “unlabeled data” without any label. Recently, the SED systems using the mean-teacher model are widely used for the task with several parameters. These parameters should be chosen carefully because they may affect the performance. In this paper, performance analysis was performed on parameters, such as the feature, moving average parameter, weight of the consistency cost function, ramp-up length, and maximum learning rate, using the data of DCASE 2020 Task 4. Effects and the optimal values of the parameters were discussed.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"40 1","pages":"139-147"},"PeriodicalIF":0.4,"publicationDate":"2021-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42050357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analysis of false alarm possibility using simulation of back-scattering signals from water masses 用模拟水团后向散射信号分析虚警可能性
IF 0.4 Q4 ACOUSTICS Pub Date : 2021-03-01 DOI: 10.7776/ASK.2021.40.2.099
Yonghoon Ha
In this paper numerical wave propagation experiments have been performed to visually confirm whether the signals scattered by water masses can be a false alarm in active sonar. The numerical environments consist of exaggerated water masses as targets in free space. Using a pseudospectral time-domain model for irregular boundary, the back-scattered signals have been calculated and compared with analytic solutions. Also, the sound propagation was simulated. Consequently, it was verified that water masses themselves could not be detected as a false target.
本文通过数值波传播实验直观地验证了水团散射信号在主动声纳中是否存在虚警现象。在自由空间中,数值环境由夸张的水团作为目标组成。利用不规则边界的伪谱时域模型,计算了后向散射信号,并与解析解进行了比较。并对声传播进行了模拟。因此,证实水团本身不能作为假目标被探测到。
{"title":"Analysis of false alarm possibility using simulation of back-scattering signals from water masses","authors":"Yonghoon Ha","doi":"10.7776/ASK.2021.40.2.099","DOIUrl":"https://doi.org/10.7776/ASK.2021.40.2.099","url":null,"abstract":"In this paper numerical wave propagation experiments have been performed to visually confirm whether the signals scattered by water masses can be a false alarm in active sonar. The numerical environments consist of exaggerated water masses as targets in free space. Using a pseudospectral time-domain model for irregular boundary, the back-scattered signals have been calculated and compared with analytic solutions. Also, the sound propagation was simulated. Consequently, it was verified that water masses themselves could not be detected as a false target.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"40 1","pages":"99-108"},"PeriodicalIF":0.4,"publicationDate":"2021-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44425605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Snoring sound detection method using attention-based convolutional bidirectional gated recurrent unit 基于注意的卷积双向门控循环单元的鼾声检测方法
IF 0.4 Q4 ACOUSTICS Pub Date : 2021-03-01 DOI: 10.7776/ASK.2021.40.2.155
Min-soo Kim, Gi Yong Lee, Hyoung‐Gook Kim
This paper proposes an automatic method for detecting snore sound, one of the important symptoms of sleep apnea patients. In the proposed method, sound signals generated during sleep are input to detect a sound generation section, and a spectrogram transformed from the detected sound section is applied to a classifier based on a convolutional bidirectional gated recurrent unit (CBGRU) with attention mechanism. The applied attention mechanism improved the snoring sound detection performance by extending the CBGRU model to learn discriminative feature representation for the snoring detection. The experimental results show that the proposed snoring detection method improves the accuracy by approximately 3.1 % ~ 5.5 % than existing method.
打鼾声是睡眠呼吸暂停患者的重要症状之一,本文提出了一种自动检测打鼾声的方法。在该方法中,输入睡眠过程中产生的声音信号来检测声音产生部分,并将检测到的声音部分变换后的频谱图应用到基于具有注意机制的卷积双向门控循环单元(CBGRU)的分类器中。所应用的注意机制通过扩展CBGRU模型学习判别特征表示来提高打鼾声音检测的性能。实验结果表明,本文提出的鼾声检测方法比现有方法准确率提高了3.1% ~ 5.5%。
{"title":"Snoring sound detection method using attention-based convolutional bidirectional gated recurrent unit","authors":"Min-soo Kim, Gi Yong Lee, Hyoung‐Gook Kim","doi":"10.7776/ASK.2021.40.2.155","DOIUrl":"https://doi.org/10.7776/ASK.2021.40.2.155","url":null,"abstract":"This paper proposes an automatic method for detecting snore sound, one of the important symptoms of sleep apnea patients. In the proposed method, sound signals generated during sleep are input to detect a sound generation section, and a spectrogram transformed from the detected sound section is applied to a classifier based on a convolutional bidirectional gated recurrent unit (CBGRU) with attention mechanism. The applied attention mechanism improved the snoring sound detection performance by extending the CBGRU model to learn discriminative feature representation for the snoring detection. The experimental results show that the proposed snoring detection method improves the accuracy by approximately 3.1 % ~ 5.5 % than existing method.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"40 1","pages":"155-160"},"PeriodicalIF":0.4,"publicationDate":"2021-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42589155","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of the Acoustical Society of Korea
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1